Logic, Sets and Functions
Daniel Bonevac Nicholas M. Asher Robert C. Koons
University of Texasat Austin
1[-1 •
KENDALLIHUNT PUBLlSH'NG COMPANY 4050
Westmark
Drive
Dubuque,
Iowa
52002
,.
To Beverly, Sheila and Debbie.
Copyright c 1999 by KendalllHunt Publishing Company ISBN 0-7872-6355.-9
£\:J O? ,..
0
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior written permission of the copyright owner. Printed in the United States of America 10 9 8 7 6 5 4 3 2 1
~
Contents 1
2
3
4
Basic Concepts of Logic 1.1 Arguments...... 1.2 Validity . . . . . . . . 1.3 Implication and Equivalence . 1.4 Logical Properties of Sentences . 1.5 Satisfiability...........
1
2 8 14 18 23
Sentences 2.1 The Language of Sentential Logic 2.2 Truth Functions . . . . 2.3 A Sentential Language 2.4 Translation. 2.5 Validity . . . . . . . . . 2.6 Truth Tables . . . . . . 2.7 Truth Tables for Formulas 2.8 Examples.......... 2.9 Truth Tables for Argument Forms 2.10 Implication, Equivalence and Satisfiability
27 27 30 33 36 38 41 43 45
Natural Deduction 3.1 Natural Deduction Systems . 3.2 Rules for Negation and Conjunction . 3.3 Rules for the Conditional and Biconditional '3.4 Rules for Disjunction 3.5 Derivable Rules
55 55
Quantifiers 4.1 Constants and Quantifiers 4.2 Categorical Sentence Forms. 4.3 Polyadic Predicates . 4.4 The Language Q . . . 4.5 Translation...... 4.5.1 Noun Phrases. 4.5.2 Verb Phrases 4.5.3 Connectives 4.6 Interpretations . . .
48 51
57 64
67 69 77
78 82
84 87 90 91 95 97 100
iii
"...
CONTENTS
iv
5
6
109
Identity and Function Symbols 6.1 Identity . . . . . . . . . . . . 6.2 Deduction Rules for Identity 6.3 Function Symbols . . . . . .
123
109 114 119
123 127 129
7 Sets 7.1 Extensionality . . 7.2 Abstraction 7.3 Pair Sets, Unit Sets and Enumeration 7.4 The Null Set . 7.5 Binary Unions, Intersections and Complements 7.6 Unions and Intersections of Single Sets. 7.7 Power Sets . 7.8 Complex Abstraction.
137
8
Relations 8.1 Sequences and Ordered Pairs 8.2 Cartesian Products . . . 8.3 Relations . 8.4 Properties of Relations . 8.5 Ordering Relations . . . 8.6 Relations between Relations . 8.7 Restrictions and Images . . . 8.8 Equivalence Relations and Partitions
163
Functions 9.1 Functions and Relations 9.2 Properties of Functions
199
9
~
Quantified Natural Deduction 5.1 Deduction Rules for Quantifiers . 5.2 Universal Proof 5.3 Derived Rules for Quantifiers .
137 142 145 147 149 156 159 161
163 165 170 175 180 185 191 195
199 203
10 Induction 10.1 The Natural Numbers and Definition by Recursion 10.2 Weak Induction on the Natural Numbers . 10.3 Strong Induction on Natural Numbers . 10.4 Induction on Sets other than the Natural Numbers 10.5 Graphs, Trees and Lists 10.6 Formal Languages
209
A Plato's Users' Guide A.1 Introduction . A.2 Getting Started . A.3 Starting Plato on the Macintosh A.3.1 The Menus: An Overview A.3.2 The File Menu . . . . A.3.3 The Edit Menu A.3.4 The Annotations Menu
251
209 217 221 224 233 241
251 252 252 252 253 253 254
CONTENTS A.3.5 The Font and Size Menus . . . . AA Homework Mode the Windows Version . A.5 The Windows Menu . . . . . . . . . . . A.6 Languages of Logic . A.6.l The Language of Sentential Logic A.6.2 The Language of Quantificational Logic . A.7 Systems of Proof. . . . . A.7.l Proof Formats .. A.7.2 Derivation Rules. A.7.3 The Proofs Menu A.7A Imported Rules . A.7.5 The Quantification Rule Set A.8 Macintosh Keyboard Shortcuts . . . A.8.l Mackintosh Keyboard Equivalents for Logical Symbols.
v
255 255 256 256 256 257 257 257 259 260 264 265 266 267
B Answers to' Selected Problems
269
Index
284
" vi
Preface This book is intended as an introduction to symbolic logic and elementary set theory, including mathematical induction. It is designed to meet the needs of a semester-long course, with the needs and interests of computer science majors and others requiring a foundation in the construction of proofs and in the use of discrete mathematics. The text is accompanied by a software application, Plato, available in both Windows and Macintosh versions. Plato can be used by students in producing derivations that employ the proof styles and rules introduced in this text. The use of Plato provides the student with immediate feedback, since it will not allow any logical errors to be committed. Plato is quite flexible, capable of being used to generate a derivation of any valid argument in first-order logic. In addition, Plato greatly eases the burden of grading homework. We would like to express our appreciation to the College of Liberal Arts at the University of Texas-Austin, and to the Multimedia Task Force of the UT System, for their financial support for the development of Plato. Austin, Texas May, 1999
....
Chapter 1
Basic Concepts of Logic Clerk: Mr. McClory moves to postpone for ten days further consideration of whether sufficient grounds exist for the House of Representatives to exercise constitutional power of impeachment unless by 12 noon, eastern daylight time, on Saturday, July 27, 1974, the President fails to give his unequivocal assurance to produce forthwith all taped conversations subpoenaed by the committee which are to be made available to the district court pursuant to court order in United States v. Mitchell... Mr. Latta: ... I just want to call [Mr. McClory's] attention before we vote to the wording of his motion. You move to postpone for ten days unless the President fails to give his assurance to produce the tapes. So, if he fails tomorrow, we get ten days. If he complies, we do not. The way you have it drafted I would suggest that you correct your motion to say that you get ten days providing the President gives his unequivocal assurance to produce the tapes by tomorrow noon. Mr. McClory: I think the motion is correctly worded, it has been thoughtfully drafted. Mr. Latta: I would suggest you rethink it ... Mr. Mann: Mr. Chairman, I think it is important that the committee vote on a resolution that properly expresses the intent of the gentleman from Illinois [Mr. McClory] and if he will examine his motion he will find that the words 'fail to' need to be stricken... Mr. McClory: If the gentleman will yield, the motion is correctly worded It provides for a postponement for ten days unless the President fails tomorrow to give his assurance, so there is no postponement for ten days if the President fails to give his assurance, just one day. I think it is correctly drafted. I have had it drafted by counsel and I was misled originally, too, but it is correctly drafted. There is a ten day postponement unless the President fails to give assurance. If he fails to give it, there is only a twenty-four-hour or there is only a twenty-three-and-a-half hour day. House Judiciary Committee, July 26, 1974 Logic is the study of correct reasoning. Logic pertains to all subjects, since people can reason about anything they can think about. Politics, the arts, literature, business, the sciences, everyday problems are all subjects open to reasoning. Sometimes the reasoning is good; sometimes, not so good. People use logic to tell the difference. Using logic, we can evaluate bits of reasoning as proper or improper, good or bad. Logic is not the study of how people do reason, but how they should reason. Logic does not describe real reasoning, with its errors, omissions and oversights; it prescribes methods for justifying reasoning, 1
,... Logic, Sets and Functions
2
that is, for showing that a given bit of reasoning is proper. Logic thus describes an ideal that actual reasoning strives for but often fails to reach. Logic begins with the study of language. To develop a system of logic, it is necessary to understand how people actually reason. To eliminate the errors that creep into people's performance, we need to examine people's considered judgments about the correctness or incorrectness of inferences. No matter what mental processes people go through to achieve the right result, they try to follow rules for putting sentences together to form proper bits of reasoning. Logic is not merely a description of reasoning; logicians examine people's evaluations of bits of reasoning to say what the rules of correct reasoning are. Logic describes not the process of reasoning but the rules for correct reasoning. The logical systems discussed in this book are designed both to explicate the meanings of certain expressions of natural languages such as English, Chinese, Swahili, and German and also to analyze reasoning in formal systems such as those used in mathematics and computer science.
1.1
Arguments
Arguments represent reasoning in language. Frequently, we think of arguments as heated debates, disagreements or disputes. Sometimes, however, we speak of a politician arguing for the passage of a bill, a lawyer arguing a case, or a moviegoer arguing that North By Northwest is better than The 39 Steps. An argument in this sense starts with some assertions called premises and tries to justify a conclusion. Many arguments in natural language are complicated. A lawyer arguing for the innocence of a client, for instance, offers many more specific arguments in presenting the case. The lawyer may argue that a piece of evidence is inadmissible, that results from a lab test are ambiguous, that the client could not have reached the scene of the crime by the time it was committed, and so on. All these smaller arguments form part of the larger argument for the client's innocence. We can divide arguments, then, into two groups: extended arguments, which contain other arguments, and simple arguments, which do not. Extended arguments may have several conclusions. Such arguments may consist of several simple arguments in sequence. They may contain other extended arguments. And they may consist of a list of premises, followed by several conclusions stated at once. Mathematical proofs are extended arguments. A mathematician may begin a proof by stating some assumptions. The mathematician then draws out consequences of the assumptions, perhaps making other assumptions along the way. Finally, the proof ends with a conclusion, the theorem it proves. A mathematical proof is thus a series of simple arguments. A simple argument, like an extended argument, starts with premises justifying a conclusion. We will be so often concerned with simple arguments that we will drop the adjective simple and speak of arguments. (Later, when we examine proofs, we will just call them proofs.) An argument consists of a finite sequence of sentences, called premises, together with another sentence, the conclusion, which the premises are taken to support. An argument in ordinary language or in mathematics is a string or sequence of sentences. The sentences making up the argument are in a particular order, whether the argument is spoken, written, or encoded in a computer language. For our purposes in this text, the order of the premises makes no difference. So we will not worry about order of presentation. But we will require that the string of premises be finite. No one has the patience to listen to an argument that runs on forever. Arguments consist of sentences. In this text, we will be interested only in sentences that can be true or false. Many ordinary sentences, including almost all in this book, fall into this category. They say something about the way the world is, and might be correct or incorrect in so describing it. But commands, for example, are different. Shut the door can be appropriate or inappropriate,
....
Basic Concepts of Logic
3
irritating or conciliatory, friendly or hostile, but it cannot be true or false. Questions, such as 'What is the capital of Zaire?', and interjections, such as 'Ouch!', are likewise neither true nor false. A sentence is true or false in a particular context: as used on a particular occasion by a particular speaker to a particular audience, in a given circumstance and as part of a discourse. Without contextual information, we cannot say whether a sentence such as I love you is true or false. Sentences have truth values only relative to a context of use. Nevertheless, very little in the following pages will involve context directly. So, we will generally speak of sentences as having truth values, trusting ourselves to remember that these values are relative to context. An argument, according to our definition, contains one sentence that is its conclusion. This is an idealization: In natural language, a conclusion may be a clause in a sentence; it may be spread across several sentences; or it may be left unstated. The same is true of premises. The definition does not specify how to pick out the conclusion of an argument. In English, certain words or phrases typically signal the conclusion of an argument, while others signal premises: Conclusion Indicators therefore, thus, hence, so, consequently, it follows that, in conclusion, as a result, then, must Premise Indicators because, for, SInce All these words and phrases have other uses; they are not always premise or conclusion indicators. But these words and phrases can, and often do, serve as indicators because they can attest to relations of support among the sentences of an argument. 'Since Fred forgot to go to the interview, he won't get the job' presents a simple argument within a single English sentence. The word 'since' indicates that we should take 'Fred forgot to go to the interview' as a premise, supporting the conclusion 'he won't get the job.' Similarly, 'Jane's business must be doing well; she drives a Mercedes' constitutes an argument. The auxiliary verb 'must' marks 'Jane's business is doing well' as the conclusion, supported by the evidence in 'she drives a Mercedes.' Premise indicators often signal not only that one or more sentences are premises,but that a certain sentence is a conclusion. Since, for example, exhibits a relation of support between the sentences it links. The occurrence of 'since' in 'Since Fred forgot to go to the interview, he won't get the job' points out, not only that the sentence immediately following it is a premise but also that the sentence 'he won't get the job' is a conclusion. Similarly, the occurrence of 'for' in 'Northern Indiana Public Service will not pay its usual dividend this quarter, for the court refused to allow expenditures on its now-cancelled nuclear project into the rate base' indicates both that 'the court refused to allow expenditures on its now-cancelled nuclear project into the rate base' is a premise and that 'Northern Indiana Public Service will not pay its usual dividend this quarter' is a conclusion. Indicators provide important clues to the structure of arguments. Often, however, no explicit indicators appear. Sometimes the conclusion is not even stated. In such cases, we must consider the point of the argument. What is the author trying to establish? Consider some examples: Suppose we argued that what was true was true for us, that two assertions met on no common ground, so that neither was "really true" or "really false". This position went further than skepticism and declared the belief in error itself to be erroneous. Royce called this view that of the total relativity of truth, and he had an argument against
,.. Logic, Sets and Functions
4
it. If the statement "There is error" is true, there is error; if it is false, then there is, ipso facto, error. He could only conclude that error existed; to deny its existence was contradictory. Bruce Kuklick, The Rise of American Philosophy This is an extended argument. The conclusion of Royce's smaller argument is plainly 'error exists'; the words conclude that make this obvious. Royce then uses this conclusion to argue that the view of the total relativity of truth is false. The sentence 'error exists' thus functions as the conclusion of one argument and as a premise of another, all within the same extended argument. If it were permitted to reason consistently in religious matters, it is clear that we all ought to become Jews, because Jesus Christ was born a Jew, lived a Jew, and died a Jew, and because he said that he was accomplishing and fulfilling the Jewish religion. Voltaire Voltaire seems to be arguing for the conclusion 'we all ought to become Jews.' Here the key word is because, which indicates that the rest of the argument is a list of premises. Voltaire, a satirist, is really aiming not at this conclusion but at another. Everything he says is supposed to follow from the hypothetical 'if it were permitted to reason consistently in religious matters.' Like Royce, he is offering an argument within an extended argument. The conclusion of the larger argument is not stated. Nevertheless, it is easy to see that Voltaire is trying to establish that it is not permitted to reason consistently in religious matters. The conclusion of the smaller argument- 'we all ought to become Jews' - is an observation that few Christians would be willing to accept, even though, according to Voltaire, their own doctrine commits them to it. The final example is a mathematical proof: the traditional proof that the square root of 2 is irrational. [Suppose] for the sake of argument that J2 is rational, i.e. that there are two integers, say m and n, which are mutually prime and which are such that ~ = .j2 or m 2 = 2n 2 . From this it follows that m 2 must be even and with it m, since a square number cannot have any prime factor which is not also a factor of the number of which it is the square. But if m is even, n must be odd according to our initial supposition that they are mutually prime. Assuming that m = 2k, we can infer that 2n 2 = 4k 2 , or n 2 = 2k 2 ; and from this it can be shown by a repetition of the reasoning used above that n must be even. Our hypothesis, therefore, entails incompatible consequences, and so it must be false.
w.
Kneale and M. Kneale, The Development of Logic
Like almost any proof, this one is an extended argument; in fact, it is a series of simple arguments. The proof begins with the assumption that J2 is rational. The first simple argument concludes that m 2 must be even; very quickly follows another simple argument concluding that m must also be even. The third simple argument concludes that n is odd. The fourth concludes that 2n2 = 4k 2 ; the fifth, that n 2 = 2k 2 ; the sixth, that n is even. Finally, the proof ends with a seventh simple argument that the hypothesis that the square root of 2 is rational is false. When we write an argument "officially", in standard form, we will list the premises in the order in which they are given, and then list the conclusion. So, in our official representations, conclusions will always come last. This is not true in natural language, as Voltaire's argument shows; conclusions may appear at the beginning, in the middle, or at the end of arguments, if they are stated at all. In addition, we'll preface the conclusion with the symbol :., which means "therefore". To see how these representations work, let's write Royce's smaller argument in standard form:
...
Basic Concepts of Logic
5
If the statement "There is error" is true, there is error. If the statement "There is error" is false, there is error. :. There is error.
Royce's larger argument, then, is: Error exists. The view of the total relativity of truth holds that the belief in error is erroneous. :. The view of the total relativity of truth is false. We can similarly express Voltaire's two arguments in standard form: Jesus Christ was born a Jew, lived a Jew, and died a Jew. Jesus Christ said he was accomplishing and fulfilling the Jewish religion. :. If it were permitted to reason consistently in religious matters, it is clear that we all ought to become Jews. If it were permitted to reason consistently in religious matters, it is clear that we all ought to become Jews. (It is not clear to religious Christians that we all ought to become Jews.) :. It is not premitted to reason consistently in religious matters.
Finally, we can express the proof of the irrationality of
v'2 as a series of simple arguments:
J2 is rational, i.e., there are two integers, say m and n, that are mutually prime and that are such that ~ = v'2 or m 2 = 2n 2 . :. m 2 is even. m 2 is even. A square number cannot have any prime factor that is not also a factor of the number of which it is the square. :. m is even. m is even. m and n are mutually prime. :. n is odd. m=2k :. 2n 2 = 4k 2
2n 2 = 4k 2 :. n 2 = 2k 2 n 2 = 2k 2
(A repetition of the reasoning used above.) :. n is even. The hypothesis that v'2 is rational entails incompatible consequences. :. The hypothesis that v'2 is rational is false.
Problems Write each of the following arguments in standard form. If there are several arguments in a passage, write each separately.
~'
Logic, Sets and Functions
6
1. The Bears did well this year, so they'll probably do well again next year.
2. John must have left already; his books are gone. 3. Few contemporary novels deal explicitly with political themes. The study of contemporary literature is therefore largely independent of the study of political culture. 4. Mary dislikes Pat. Consequently, it's unlikely that they'll work on the project together. 5. Most criminals believe that their chances of being caught and punished are small; thus, the perceived costs of a life of crime are low. 6. The building will generate large tax write-offs. As a result, it will be a good investment even if it yields little direct profit. 7. No one has ever constructed a convincing case that Bacon or someone else wrote the plays we generally attribute to Shakespeare. Shakespeare, then, almost certainly wrote the plays we attribute to him. 8. There are no centaurs, for centaurs are mythical creatures, and no mythical creatures really exist. 9. Nobody will ever find an easy way to get rich, because people have been looking for centuries, and nobody's ever found one yet. 10. Swedish is an Indo-European language, but Finnish isn't. Hence Finnish is more difficult for English-speakers to learn than Swedish. 11. Many people are easily shocked by unusual or threatening events. No one who is thunderstruck can think clearly. It follows that the emotions can obstruct reason. 12. Since happiness consists in peace of mind, and since durable peace of mind depends on the confidence we have in the future, and since that confidence is based on the science we should have of the nature of God and the soul, it follows that science is necessary for true happiness. (Leibniz) 13. In Europe pupils devote time during each school day to calisthenics. American schools rarely offer a daily calisthenic program. Tests prove that our children are weaker, slower, and shorterwinded than European children. We must conclude that our children can be made fit only if they participate in school calisthenics on a daily basis. (LSAT test, 1980)
...
14.
* First, the personality and character- which are really synonymous- take their form during the first six or eight years of life. During this period of infancy and childhood, we select and develop the techniques which gain us satisfaction, defend us against threats, and become the tools in coping with the endless variety of problems situations that will be encountered later in life. It is during this time that we develop our methods of relating ourselves to other people and undergo the experiences which determine the strengths and weaknesses within our personalities. As adults we are not able to remember the details of these formative years. Therefore, we cannot understand our own behavior fully. (William Menninger)
15.
* The
earth receives radiant heat from the sun and loses heat to outer space by its own radiative emissions. The energy received undergoes many transformations. But in the long run no appreciable fraction of this energy is stored on the earth, and there is no persistent trend toward higher or lower temperatures. (LSAT test, 1980)
Basic Concepts of Logic
7
16. *To describe an equilateral triangle on a given finite straight line. Let AB be the given straight line: it is required to describe an equilateral triangle on AB. From the center A, at the distance AB, describe the circle BCD. From the center B, at the distance BA, describe the circle ACE. From the point C, at which the circles cut one another, draw the straight lines CA and CB to the points A and B. ABC shall be an equilateral triangle. Because the point A is the center of the circle BCD, AC is equal to AB. And because the point B is the center of the circle ACE, BC is equal to BA. But it has been shown that CA is equal to AB; therefore CA and CB are each of them equal to AB. But things which are equal to the same thing are equal to one another. Therefore CA is equal to CB. Therefore, CA, AB, BC are equal to one another. Wherefore the triangle ABC is equilateral, and it is described on a given straight line AB. (Euclid) 17. * One may well ask, "How can you advocate breaking some laws and obeying others?" The answer is found in the fact that there are two types of laws: There are just laws and there are unjust laws. I would be the first to advocate obeying just laws. One has not only a legal but a moral responsibility to obey just laws. Conversely, one has a moral responsibility to disobey unjust laws. I would agree with St. Augustine that "An unjust law is no law at all." Now what is the difference between the two? How does one determine whether a law is just or unjust? A just law is a man-made code that squares with the moral law or the law ofGod. An unjust law is a code that is out of harmony with the moral law. To put it in the terms of St. Thomas Aquinas, an unjust law is a human law that is not rooted in eternal and natural law. Any law that uplifts human personality is just. Any law that degrades human personality is unjust. All segregation statutes are unjust because segregation distorts the soul and damages the personality. . . . (Martin Luther King Jr.) 18. * Theorem (Mean value theorem). Let a, bE 'R, a < b, and let f be a continuous real-valued function on [a, b] that is differentiable on (a, b). Then there exists a number c E (a, b) such that feb) - f(a) = (b - a)f'(c). Proof: Define a new function F : [a, b]
---+
'R by
F(x) = f(x) - f(a) -
feb) - f(a) b_ a . (x - a)
for all x E [a, b]. (Geometrically F(x) is the vertical distance between the graph of f over [a, b] and the line segment through the end points of this graph.) Then F is continuous on [a, b], differentiable on (a, b), andF(a) = F(b) = O. By Rolle's theorem, there exists acE (a, b) such that F'(e) = O. Thus,
F'(e)
= f'(e)
- feb) - f(a) b-a
=0
proving the result. (M. Rosenlicht) 19. * If the premise, that independence means complete absence of subconscious bias, were carried to its logical conclusion, no one could be found independent in any absolute sense. One is bound to be influenced by his environment, and everyone has some subconscious bias. On this premise it could be argued that the accountant who frequently lunches with a client, or plays golf or bridge with him, or serves with him on a board of trustees of a church or school, should not be considered independent in certifying the financial statements of that client. . . . [This
IJ!I""'-t
8
Logic, Sets and Functions would lead] to the conclusion that certified public accountants who act as independent auditors for a company must avoid all other relationships with that company, leaving the accounting work to be performed by other CPAs, and accepting accounting work only from clients who retain other independent auditors. This conclusion seems almost fantastically inconsistent with settled practice and the temper of the community generally. There would be serious economic loss in depriving the business man who needs accounting service of the knowledge which the independent auditor gains about the business, particularly its accounting aspects, in the course of his audit. (William Carey) 20.
1.2
* I should like to suggest that neither composition nor literature is an intellectual field in its own right. Literary study obviously connects with a number of genuine intellectual fields like history and philosophy. Composition, too, has disciplinary connections with linguistics and psychology. But neither literary study nor composition is an intellectual discipline. Both are primarily cultural subjects with cultural missions of unparalleled importance. To the extent that we evade those missions under the banner of some neutral formalism or disciplinary pretense, we are neglecting our primary educational responsibilities and are also making an empirical mistake. This clearly implies that we should return to an integrated conception of "English" based on the pattern of the old literature-and-composition course originated by Blair and followed traditionally in the schools and colleges. (E. D. Hirsch, Jr.)
Validity
Some arguments are good; others are not. According to the definition of argument given above, any collection of sentences counts as an argument if it is possible to single out one sentence in the collection as the conclusion, purported supported by the others. What distinguishes good from bad arguments? What makes a good argument succeed? What makes a bad argument fail? People typically demand many things of an argument. Throughout this book, we will focus on one of them. A good argument should link its premises to its conclusion in the right way. There should be some special connection between the premises and the conclusion. To see what this special sort of connection is, consider an argument that has true premises and a true conclusion, but is nevertheless bad: Harrisburg is the capital of Pennsylvania. Richmond is the capital of Virginia. :. Austin is the capital of Texas. What is wrong with this argument? The facts cited in the premises have nothing to do with the truth or falsehood of the conclusion. Texas could move its capital- to, say, Del Rio- while Harrisburg and Richmond remained the capitals of their respective states. That is, the conclusion of this argument could turn out to be false, even when the premises were true. The truth of the premises does nothing to guarantee the truth of the conclusion. This is the mark of a deductively invalid argument: its premises could all be true in a circumstance in which its conclusion is false. In a deductively valid argument, the truth of the premises guarantees the truth of the conclusion. If the premises are all true, then the conclusion has to be true. Consider, for example, this argument: Paris is, and has always been, the capital of France. Ed has never visited Paris. :. Ed has never visited the capital of France.
Basic Concepts of Logic
9
In any circumstance in which the premises of this argument are true, the conclusion must be true as well. It is impossible to conceive of a state of affairs in which Ed has visited the French capital without visiting Paris, if Paris is, and has always been, the capital of France. To put this differently, the only way to imagine Ed visiting the French capital without visiting Paris is to imagine a case where Paris is not the capital of France. In a deductively valid argument, the truth of the premises guarantees the truth of the conclusion. Or, to say the same thing, if the conclusion of a deductively valid argument is false, at least one premise must also be false.
Definition 1.1 An argument is deductively valid if and only if it is impossible for its premises all to be true while its conclusion is false. It is possible, then, for a deductively valid argument to have true premises and a true conclusion; (at least some) false premises and a false conclusion; and false premises and a true conclusion. But no deductively valid argument has true premises and a false conclusion. Some Deductively Valid Arguments True Premises False Premises False Premises True Conclusion False Conclusion True Conclusion Daniel is human. Daniel is a dog. Daniel is a dog. All humans are mortal. All dogs eat mice. All dogs sleep. :. Daniel is mortal. :. Daniel eats mice. :. Daniel sleeps. Each of these arguments is deductively valid: In each case, there is no possible circumstance in which the premises are all true but the conclusion is false. How could it be true that Daniel is a dog, and true that all dogs eat mice, but false that Daniel eats mice? Whether the premises and conclusion are actually true or false makes little difference to the validity of the argument. What matters is that if the premises are true the conclusion cannot be false. Thus, not every argument with true premises and a true conclusion is deductively valid, as the argument concerning state capitals shows. Similarly, many arguments with false premises and a true conclusion are deductively invalid. The same is true for arguments with false premises and a false conclusion. So, although valid arguments can have any of these three combinations of truth and falsity, not every argument with those combinations is valid. An argument is deductively invalid if it is possible for the premises to be true while the conclusion is false. Similarly, an argument is deductively valid just in case its conclusion has to be true if its premises are all true.! Some deductively invalid arguments nevertheless have some legitimate force in reasoning. Although the truth of the premises of such an argument does not guarantee the truth of its conclusion, it does make the truth of the conclusion probable. Consider for example, this argument: Most cats like salmon better than beef. Gwen is a cat. :. Gwen likes salmon better than beef. It is possible for the premises to be true while the conclusion is false. Gwen may be atypical; she may
prefer beef to salmon. So the argument is deductively invalid. Nevertheless, the premises lend some support to the conclusion. Given just the information in the argument, the conclusion is more likely to be true than false. Arguments such as this are called inductively reliable. They are important in both scientific and everyday reasoning. Evaluating them, however, requires developing theories of IThese links between deductive validity and truth or falsehood were first recognized explicitly by the Greek philosopher Aristotle (384-322 B. C.), the father of logic.
",..-
'"
10
Logic, Sets and Functions
probability and statistics. In this text, therefore, we will restrict our attention to deductive validity and invalidity. When we imagine a circumstance in which some sentences would be true, and others would be false, we normally imagine a situation that settles the matters that those sentences involve, but that leave lots of other things unsettled. Above, for example, we imagined a case in which Texas moved its capital, but Pennsylvania and Virginia didn't. That was all we said, or, apparently, needed to say to convince ourselves that the argument was invalid. But that was not even close to a complete description of an entire world. We said nothing about what happened to Montana, or Alaska, or Afghanistan, or the pennant hopes of the Mets, or the price of pork bellies on the Chicago Board of Trade. The case we've described, therefore, isn't very determinate. There are many different ways the world might be that all agree in fitting our description. So, it might be more correct to say that we imagined, not a single case, but a kind of case in which the premises are all true and the conclusion is false. Many circumstances might fit the description we gave. The logic we study in this book assumes that some circumstances are so comprehensive that they determine whether each declarative sentence of a language is true or false. Every sentence that can be true or false at all must, in such a complete situation, be either true or false. The logic studied throughout most of this book is a bivalent logic because it says that, given any sentence capable of truth or falsehood, the question, "Is this sentence true, or false, or whatever?" always has only two possible answers: "True" and "False". In other words, classical logic allows only two truth values: truth and falsehood. The truth value of a sentence is truth, if the sentence is true, and falsehood or falsity, if it is false. So far in this section we've examined only the second criterion for success in arguments: we want arguments to be valid. Deductively valid arguments always preserve truth; if they begin with true premises, they carry us to true conclusions. It's nice, of course, to have an argument that not only preserves truth but has some truth to preserve. Thus, our first criterion was that a successful argument should have true premises and a true conclusion. A sound argument meets both criteria for success. It has true premises; it is valid. Furthermore, since, in any valid argument, the truth of the premises guarantees the truth of the conclusion, it also has a true conclusion.
Definition 1.2 An argument is sound if and only if (1) it is valid and (2) all its premises are true. Sound arguments, then, are often paradigms of successful arguments. They derive the truth of their conclusions by arguing validly from true premises. Nevertheless, most of this book will focus, not on soundness, but on validity. Logicians have always concentrated on validity. 2 This focus is easy to understand. First, validity is obviously a crucial component of soundness. We can't evaluate whether an argument is sound without first determining whether it's valid. Second, evaluating soundness requires judging the actual truth or falsehood of premises. This, however, is the job, not of logical theory, but of those who know enough physics, history, business, or whatever facts are relevant to the argument at hand. Third, although we usually want to argue from true premises, many useful arguments start from false ones. Some arguments try to show that a certain sentence is false by using it as a premise to reach an outrageous or absurd conclusion. Others adopt a premise purely as a hypothesis, to see what would follow if it were true. Aristotle first realized how important such arguments are; he characterized them as having dialectical, rather than demonstrative, premises. As we shall see later, these forms of argument are much more common and useful than most people would imagine. 2Indeed, although Aristotle and some earlier thinkers talked of valid arguments, a term for soundness was only introduced later, by logicians of the Stoic school, which thrived in Greece from the third century B. C. to the second century A. D.
Basic Concepts of Logic
11
A simple example occurs at the beginning of this chapter in the proof of the irrationality of the square root of two. The proof starts with the assumption that J2 is rational, and deduces from it a contradiction. The point of this argument is precisely to show that the premise that J2 is rational is false. The argument's success, therefore, depends solely on validity, not on soundness. Our study of reasoning will therefore focus on validity. It's worth noting that the logical concept of soundness approximates, but is not the same as our ordinary, intuitive concept of a successful argument. Successful arguments generally lead us from premises for which we have good evidence to a conclusion that follows from those premises and to which the premises are relevant. The logical concept of soundness, however, mentions neither evidence nor relevance. Our technical definition thus calls sound a variety of arguments that seem, from an intuitive point of view, peculiar. Some violate our ordinary notion of evidence. So, suppose that the earth will be invaded by little green men in 2025, but that we possess no evidence now to support this. Then the argument The earth will be invaded by little green men in 2025 . .. The earth will be invaded by little green men in 2025. is sound; given our supposition, the premise is true, and the conclusion is surely true whenever the premise is, since they are the same sentence. But this argument won't convince anyone that we ought to be building defenses; it doesn't establish its conclusion in the usual, evidence-related sense of establish. Some arguments also count as sound even though they violate our usual notion of relevance. The argument Coffee ice cream is more popular than chocolate in Rhode Island. :. Cats are cats. is sound, since the premise is true, and the conclusion can never be false while the premise is true, simply because the conclusion can never be false. Yet this argument, too, seems bizarre. The premise is irrelevant to the conclusion. Thus, just as validity is a part, but only a part of soundness, so soundness is a part, but only a part, of our intuitive notion of success in argumentation.
Problems Evaluate these arguments as valid or invalid. If the argument is invalid, describe a circumstance in which the premises would be true but the conclusion would be false. 1. John and Mary came to the party. Hence, Mary came to the party. 2. Larry got angry and stormed out of the room. Consequently, Larry stormed out of the room. 3. If Susan's lawyer objects, she will not sign the contract. Susan's lawyer will object. Therefore she will not sign the contract. 4. If Frank takes the job in Cleveland, he'll make a lot of money on the sale of his house. Frank won't take the job in Cleveland. It follows that Frank won't make a lot of money on the sale of his house. 5. If Strawberry hits 30 home runs, the Mets will be contenders. The Mets will be contenders. So Strawberry will hit 30 home runs. 6. If Lynn testifies against the mobsters, she'll endanger her life. So, she won't testify against them, since she won't put her own life in danger. 7. Max is mayor of either Abilene or Anarene. Max isn't mayor of Abilene. Hence Max must be mayor of Anarene.
,., ,.
Logic, Sets and Functions
12
8. Pamela played Shelley for the tournament trophy. Consequently, Pamela played either Shelley or Tracy for the trophy. 9. Henry doesn't know anyone. So Henry doesn't know Kim. 10. Rocky has beaten everyone he's faced. Thus, Rocky has beaten Mad Moe, if he's faced him. 11. Since all who have been accepted have scores over 1300, either Jim has been accepted, or his scores weren't over 1300. 12. Since all who have been accepted have scores over 1300, either Jim hasn't been accepted, or his scores were over 1300. 13. Everyone who has thought about the political tensions of the Middle East realizes that they're complicated. Deborah doesn't realize that these political tensions are complicated, so she mustn't have thought about them. 14. Everyone who admires Frost also admires Dickinson. Some people who normally hate poetry admire Frost. Therefore some people who normally hate poetry admire Dickinson. 15. Some politicians are demagogues, but no demagogues are good leaders. Hence, some politicians are not good leaders. 16. All scientists have a deep interest in the workings of nature. All who devote their lives to the study of the physical world have a deep interest in the workings of nature. Consequently, all scientists devote their lives to the study of the physical world. 17. Some modern art shows the strong influence of primitivism. No modern art is primarily representational. Thus, some art that exhibits the influence of primitivism is not primarily representational. 18. Most Americans like baseball. Anyone who likes baseball likes sports. So most Americans like sports. 19. Most medieval theories of motion were, in essence, Aristotelian. No theory of motion that uses a concept corresponding to inertia is essentially Aristotelian. It follows that most medieval theories of motion used no concept corresponding to inertia. 20. The patient will surely die unless we operate. We will operate. Therefore the patient will not die. 21. Jerry will take the job unless we match the salary offer. Since we won't match the offer, Jerry will take the job. 22. The launch will be delayed unless the weather clears. So, if the weather clears, the launch won't be delayed. 23. The meeting will take place only if both parties agree on the agenda. So, if the parties don't agree on the agenda, the meeting will not take place. 24. Marilyn will finish the brief on time only if she gets an extension on the Morley case. Therefore, if Marilyn gets an extension on the Morley case, she will finish the brief on time. 25. If Jack understands how important this sale is, he'll devote most of the next two weeks to securing it. It follows that Jack won't devote most of the next two weeks to securing this sale unless he understands how important it is.
Basic Concepts of Logic
13
26. The boss won't understand what you're trying to say unless you put it in the bluntest possible terms. Consequently, if you don't put what you're trying to say in the bluntest terms possible, the boss won't understand it. 27. Either the city will raise electric rates, or it will raise taxes. Thus, if the city does not raise electric rates, it will raise taxes. 28. Nancy will not marry Alex unless he signs a prenuptial agreement. So, if Alex signs a prenuptial agreement, Nancy will marry him. 29. This album will sell only if it contains at least one hit song. Hence, unless it contains a hit song, this album will not sell. 30. John is watching Mary run through the park. So Mary must be running through the park. 31.
* Few students fully appreciate the value of an education while they are in school. Only those who fully appreciate the value of their education while they are in school devote themselves to their studies as much as they ought to. Therefore, most students don't devote themselves to their studies as much as they ought to.
32.
* Corporate taxes result in higher prices for consumer goods, increases in interest rates, reduced employment at lower wages, and reduced levels of savings and investment, depending on whether corporations pass along the cost of taxation to the consumer, borrow to replace these funds, take steps to reduce labor costs, or reduce the return they offer to shareholders. Consequently, corporate taxes should be repealed.
33.
* Most Americans who travel in Europe know no language other than English. All Americans who travel in Europe are affluent. Thus, most affluent Americans know no language other than English.
34.
* The President didn't know that several of his subordinates had started "the company within the company", a small, highly secret group within the CIA. All of the President's subordinates belong to the President's political party. It follows that the President didn't know that several people of his own political party started a secret group within the CIA.
35.
* Few mathematics students take courses in logic. All accounting majors take courses in logic. So few accounting majors are students of mathematics.
36.
* Almost all Asian nations have socialist or statist or otherwise centralized economies.
All our allies in Eastern Asia are, of course, Asian nations. So most of our Eastern Asian allies have centralized economies.
37.
** By 1988, the capital of Israel will be either Tel Aviv or Jerusalem. Thus, if in 1988 the Israeli capital is not Jerusalem, it will be Tel Aviv.
38.
**
39.
** Terry's mother gave her permission to go to the movies or to the park. Thus, Terry's mother gave her permission to go to the park.
40.
** The collapse of the Austro-Hungarian Empire at the end of the First World War caused the fragmentation and political divisions that led, ultimately, to an easy Soviet takeover of most of Eastern Europe at the end of World War II. Thus, if the end of World War I had not witnessed the collapse of Austria-Hungary, the Soviets would have found it more difficult to take over most of Eastern Europe after the Second World War.
Dogs are animals. Dogs bark. John owns a dog. So John owns an animal that barks.
".
. Logic, Sets and Functions
14
1.3
Implication and Equivalence
A concept closely related to validity is implication. We might express the idea that an argument is valid by saying that its conclusion follows from its premises. Equivalently, we might say that its premises imply or entail its conclusion. At least part of what we mean, in either case, is that the truth of the premises guarantees the conclusion's truth. If the premises are true, the conclusion has to be true too. Implication, then, is very similar to validity. But validity is a property of arguments; implication is a relation between sentences and sets of sentences. A set of sentences implies a given sentence just in case the truth of that sentence is guaranteed by the truth of all the members of the set. 3
Definition 1.3 A set of sentences S implies a sentence A if and only if it's impossible for every member of S to be true while A is false. It should be clear from this definition that, if an argument is valid, the set consisting of its premises implies its conclusion. We can also speak of a single sentence implying another sentence.
Definition 1.4 A sentence A implies a sentence B if and only if it's impossible for A to be true while B is false. One sentence implies another, that is, just in case the truth of the former guarantees the truth of the latter. In every circumstance in which the first is true, the second must be true as well. Consider these two pairs of sentences. 1.
(a) Mary likes Chinese food, but Bill hates it. (b) Mary likes Chinese food.
2.
(a) Susan is going to spend her summer in either Palo Alto or Pittsburgh. (b) Susan is going to spend her summer in Pittsburgh.
Sentence (l)a. implies (l)b. It's impossible to conceive of a situation in which it's true that Mary likes Chinese food, but Bill hates it, and false that Mary likes Chinese food. In such a circumstance, Mary would have to like and not like Chinese food; the sentence 'Mary likes Chinese food' would have to be both true and false at the same time. There are no such circumstances. No sentence can be both true and false at the same time. So the truth of (l)a. guarantees the truth of (l)b. Does the truth of (2)a. similarly guarantee the truth of (2)b.? Obviously, the answer is no. Imagine a world in which Susan is going to spend her summer in Palo Alto, never setting foot outside California. In this situation, (2)a. is true, but (2)b. is false. So (2)a. does not imply (2)b. A sentence A implies a sentence B just in case B is true in all those possible circumstances in which A is true. B implies A, of course, just in case A is true in all those cases in which B is true. If A implies Band B implies A, then A and B must be true in exactly the same circumstances. In such a case, we say that A and B are equivalent.
Definition 1.5 A sentence A is equivalent to a sentence B if and only if it's impossible for A and B to disagree in truth value. If A and B are equivalent, then they must be true in the same circumstances, and false in the same circumstances. There could be no situation in which one would be true while the other would be false. Thus, equivalence amounts to implication in both directions. A is equivalent to B just in case A implies Band B implies A. To make this more concrete, consider four more pairs of sentences: 3Throughout this book, 'just in case' will be used as a synonym for 'if and only if.'
Basic Concepts of Logic 3.
15
(a) No apples are oranges. (b) No oranges are apples.
4.
(a) All apples are fruits. (b) All fruits are apples.
5.
(a) The Senator is neither worried nor angry about the investigation. (b) The Senator is not worried about the investigation; he is not angry about the investigation.
6.
(a) Professor Pinsk saw that no one left. (b) Professor Pinsk saw no one leave.
The sentences in (3) are equivalent. Any circumstance in which no apples are oranges is one in which no oranges are apples, and vice versa. Both sentences say that nothing is both an orange and an apple. In (4), however, the sentences are obviously not equivalent. All apples are fruits, so (4)a. is true. But not all fruits are apples, so (4)b. is false. The real world is thus a case in which these sentences disagree in truth value. Similarly, the sentences in (5) are equivalent; they are true in exactly the same circumstances. If the Senator is neither worried nor angry, then he is not worried, and he is not angry. Conversely, if he is not worried, and is not angry, then he is neither worried nor angry. The sentences in (6), however, are not equivalent. Imagine a case where a student left without being observed by the professor. In such a case, it could well be true that Professor Pinsk saw no one leave; it would nonetheless be false that the professor saw that no one left, since, in fact, someone did leave.
Problems Consider the sentences in each pair: are they equivalent? If not, does either sentence imply the other? 1. (a) Both Alan and Bob took their vacations in California.
(b) Alan took his vacation in
California. 2. (a) Vivian and Beth both majored in English in college. (b) Beth majored in English in college, and so did Vivian. 3. (a) Pittsburgh will face Dallas or New York in the championship game. (b) Either Pittsburgh will face Dallas in the championship game, or Pittsburgh will face New York. 4. (a) Hance will run for the governorship or the Senate. (b) Hance will run for the Senate. 5. (a) Mali will continue to experience severe food shortages throughout most of the countryside unless more roads are built. (b) More roads will be built in Mali. 6. (a) Neon and Xenon are inert. (b) Xenon and Neon are inert. 7. (a) Pluto or Uranus is now directly aligned with Neptune. (b) Pluto and Uranus are now directly aligned with Neptune. 8. (a) Not both whales and dolphins are fish. (b) Whales are not fish, and dolphins aren't either. 9. (a) Either Sam or Peter failed to give the play an appropriate sense of place. (b) Peter and Sam did not both give the play an appropriate sense of place. 10. (a) Columbia and Universal cannot both be the year's most successful studio. (b) Neither Columbia nor Universal is the year's most successful studio.
~
Logic, Sets and Functions
16
11. (a) Aunt Alice will not come to the wedding, and neither will Uncle Harry. (b) Not both Uncle Harry and Aunt Alice will come to the wedding. 12. (a) Either the Babylonians or the Assyrians employed the lex talionis. (b) If the Assyrians employed the lex talionis, the Babylonians didn't. 13. (a) Either the physical world really exists, independently of our minds, or our senses systematically deceiv~ us. (b) If our senses systematically deceive us, then the physical world doesn't really exist independently of our minds. 14. (a) If pay-per-view television catches on, cable companies will make huge profits. (b) If payper-view TV doesn't catch on, cable companies will not make huge profits. 15. (a) Universities will continue to grow only if they find new markets for their services. (b) If universities do not find new markets for their services, they will not continue to grow. 16. (a) If Elizabeth did not sign this letter, then her assistant did. (b) If Elizabeth had not signed this letter, her assistant would have. 17. (a) If Caesar had not crossed the Rubicon, he would never have become Consul. (b) If Caesar had become Consul, he would have crossed the Rubicon. 18. (a) No high-paying job is easy; (b) No easy job is high-paying. 19. (a) Some small law firms have specialists in municipal bonds. (b) Some law firms who have specialists in municipal bonds are small. 20. (a) All corporations primarily in the metals business are looking to diversify. (b) All corporations looking to diversify are primarily in the metals business. 21. (a) Most foods that are high in carbohydrates are high in calories. (b) Most foods that are high in calories are high in carbohydrates. 22. (a) At least three of my friends own cats. (b) At least three people who own cats are my friends. 23. (a) At most five doctors in the United States can perform that operation. (b) At most five doctors who can perform that operation are in the United States. 24. (a) The President is a Republican. (b) The Republican is a President. 25. (a) Several cities with populations over 700,000 have no baseball franchises. (b) Several cities without baseball franchises have populations over 700,000. 26. (a) A European country bordering on the Adriatic has close ties with China. (b) A European country having close ties with China borders on the Adriatic. 27. (a) Anybody who can speak effectively can find a job in sales. (b) Anybody who can find a job in sales can speak effectively. 28. (a) Either the Federal Reserve Board or foreign investments will increase the supply of capital. (b) If the Federal Reserve Board doesn't increase the supply of capital, foreign investments will. 29. (a) Fred knows that Jupiter is closer to the sun than Saturn. (b) Jupiter is closer to the sun that Saturn, and Fred knows it.
Basic Concepts of Logic
17
30. (a) Meg thinks that Oswald did not shoot Kennedy. (b) Meg realizes that Oswald did not shoot Kennedy. 31. (a) It wasn't necessary for things to turn out as they did. (b) Things could have turned out differently. 32.
** (a) Many films that make a lot of money are tailored to the teenage audience. (b) Many films tailored to the teenage audience make a lot of money.
33.
** (a) Our soil tests show that we will strike either oil or gas. (b) Our soil tests show that we will strike oil unless we strike gas.
34.
** (a) The tachyon is a particle that either travels backwards in time or travels faster than the speed of light. (b) The tachyon is a particle that travels backwards in time unless it travels faster than the speed of light.
35.
** (a) It's not true that Donna will come to the party but won't enjoy herself. (b) If Donna comes to the party, she'll enjoy herself.
36.
** (a) Few who read Hemingway write like Melville. (b) Few who write like Melville read Hemingway.
37.
** (a) Good wine isn't inexpensive. (b) Inexpensive wine isn't good. ** (a) Even Ralph found your comments offensive. (b) Ralph found your comments offensive. ** (a) We have some excellent redfish today, if you would like some. (b) We have some excellent
38. 39.
redfish today. 40.
** (a) If historians revise their analysis of the impact of refugees from Weimar Germany on American intellectual history, they will revise their entire conception of that history. (b) Historians will revise their analysis of the impact of refugees from Weimar Germany on American intellectual history only if they revise their entire conception of that history.
Consider the statement: If a fetus is a person, it has a right to life. Which of the following sentences follow from this? Which imply it? 41. A fetus is a person. 42. If a fetus has a right to life, then it's a person. 43. A fetus has a right to life only if it's a person. 44. A fetus is a person only if it has a right to life. 45. If a fetus isn't a person, it doesn't have a right to life. 46. If a fetus doesn't have a right to life, it isn't a person. 47. A fetus has a right to life. 48. A fetus isn't a person only if it doesn't have a right to life. 49. A fetus doesn't have a right to life only if it isn't a person. 50. A fetus doesn't have a right to life unless it's a person.
,..,
18
Logic, Sets and Functions
51. A fetus isn't a person unless it has a right to life. 52. A fetus is a person unless it doesn't have a right to life. 53. A fetus has a right to life unless it isn't a person. Consider the statement: The patient will die unless we operate immediately. What follows from this, together with the information listed? 54. The patient will die. 55. The patient will not die. 56. We will operate immediately. 57. We won't operate immediately. In a quotation at the beginning of this chapter, Mr. McClory introduces a motion saying: There will be a ten day postponement unless the President fails to give his assurance to produce the White House tapes. Throughout the subsequent debate, congressmen offer paraphrases of this motion which they believe better express Mr. McClory's intentions, and Mr. McClory also expresses his motion in other terms which he takes to be equivalent to his original. Which of these paraphrases are in fact equivalent to the original motion? 58. If the President fails to give assurance, there is a ten day postponement. (Mr. Latta proposes this as equivalent.) 59. If the President gives his assurance, there is no postponement. (Mr. Latta proposes this as equivalent. ) 60. There is a ten day postponement provided that the President gives his assurance. (Mr. Latta suggests this as a nonequivalent revision.) 61. There is a ten day postponement unless the President gives his assurance. (Mr. Mann suggests this as a nonequivalent revision.) 62. There is no postponement if the President fails to give his assurance. (Mr. McClory proposes this as equivalent.)
1.4
Logical Properties of Sentences
Logic deals primarily with the logical connections between sentences. Nevertheless, it also classifies individual sentences. The overwhelming majority of sentences we use could, depending on what the facts are, be either true or false. It's possible to conceive of cases in which they would be true, and other cases in which they would be false. For instance, each of the following sentences would be true in some circumstances and false in others: . 7. The snow is falling all over Ireland. 8. The King recognized that the some of the nobles would oppose him.
Basic Concepts of Logic
19
9. The earth is the third planet from the sun. 10. Francis Bacon, not Shakespeare, wrote The Merchant of Venice. Such sentences are contingent:
Definition 1.6 A sentence is contingent if and only if it's possible for it to be true and possible for it to be false. Contingent sentences could be true, given the right set of circumstances. Of course, they could also be false, depending on the facts of the situation. They are immensely useful precisely because they assert, in effect, that the real circumstance is among those in which they are true. Some sentences, in contrast, cannot help being true. It's simply impossible for them to be false. They are true in every possible circumstance. Such sentences are valid, or logically true:
Definition 1.7 A sentence is valid (or logically true) if and only if it's impossible for it to be false. If you doubt that there are any sentences that cannot be false, no matter what the facts may be, then try to imagine circumstances in which these sentences are false.
11. Either Lima is in Ecuador or it's not. 12. A rose is a rose. 13. Wherever you go, there you are. 14. It ain't over 'til it's over. 15. When you're hot, you're hot. 16. Either some of my friends are crazy, or none of them are. These sentences are true in every possible world. They also seem to say very little. But not all valid sentences are so straightforward and unsurprising. (17), for example, is logically true: 17. If everyone loves a lover, and Sam doesn't love Jeanne, then Jeanne doesn't love Greg. But it doesn't seem as trivial as (12)-(16). Notice, furthermore, that even those sentences can be useful. Sometimes they set up the structure of an argument, as when a mathematician begins a proof by saying, "the number n is either prime or not prime. If it is prime. . . ." At other times, they serve a function in discourse by forcing the listener to interpret certain terms as ambiguous. We normally assume that a speaker is making a good faith effort to communicate information. So, when Yogi Berra said "It ain't over 'til it's over," he presumably meant something like "it ain't over 'til it's really over,'" that is, "the outcome isn't fully determined until the game ends." So interpreted, the sentence isn't valid at all, but contingent. Some sentences, furthermore, could never be true. They are false, regardless of the facts. These sentences are contradictory (or contradictions).
Definition 1.8 A sentence is contradictory if and only if it's impossible for it to be true. Here are some examples of contradictions: 18. Fred is both bald and not bald.
J"'"
Logic, Sets and Functions
20
19. Sheila is irritated, and she's not. 20. This set belongs to itself if and only if it doesn't belong to itself.
21. Nobody's seen the trouble I've seen. In no conceivable circumstance could any of these sentences be literally true. Try, for example, to imagine a situation in which Fred is both bald and not bald at the same time. Whatever the state of Fred's scalp, he's either bald, or not bald, but not both. No matter what Sheila's state of mind may be, she is either irritated or not. (Of course, (19) might be used to suggest that she is irritated at one thing but not at another. On that interpretation, (19) is contingent.) Similarly, the set in question must belong to itself or not. And, since I've seen the trouble I've seen, somebody (namely, me) has indeed seen the trouble I've seen. Like logical truths, contradictions tend to signal that we should interpret some terms generously, since we assume that our colleagues in communication are trying to say something that could be true. Hearing (21), then, we tend to read the Nobody as Nobody else, reading the sentence as a whole as if it were Nobody else has seen the trouble I've seen. Contradictions too may ulfill important functions in arguments. (20), for example, might be a crucial step in showing that the set under consideration can't exist. Nevertheless, contradictions are disruptive enough that it's worth having a term for sentences that, whether they are valid or contingent, at least are not contradictory. Such noncontradictory sentences are satisfiable.
Definition 1.9 A sentence is satisfiable if and only if it's not contradictory. Obviously, a sentence is satisfiable just in case it's either contingent or valid. That is, it must be possible for the sentence to be true. Since every sentence is either valid, contingent or contradictory, the terms introduced in this section divide sentences into three groups, as shown in this diagram. Sentences Valid Contingent True in every True in some circumstances, false in others circumstance Satisfiable (true in some circumstances)
Contradictory False in every circumstance
We've assumed, throughout this chapter, that all sentences are either true or false, but not both. We've thus been treating sentences as the bearers of truth value. This is plausible only if we think of a sentence as used on a particular occasion, by a particular speaker, to a particular audience, in a given context. Without all this contextual information, we can't begin to say whether a sentence such as 'I love you' is true or false. The truth value of the sentence clearly depends on who 'I' and 'you' refer to, when the sentence is uttered, etc. It would be better, therefore, to speak of utterances of sentences as true or false, rather than sentences themselves, or to speak of sentences as having truth values only relative to a certain context of use. Some sentences- such as 'I am here now' - are true whenever they are uttered, but are nevertheless not valid. No matter who is speaking, or when or where the utterance takes place, 'I am here no'w is true. But that does not make it a necessary truth. Suppose, for example, that I utter the sentence as I write this line. Supplying information from the context, that utterance has the force of saying that Daniel Bonevac is in Austin on May 7, 1999. But this is not necessarily true; it's easy to imagine possible circumstances in which I am in Istanbul or, less fancifully, in Dripping Springs. The role of context in language is critically important, and should not be forgotten. Nevertheless, very little in the following pages will involve context directly. We'll continue, then, to speak of sentences as having truth values, trusting ourselves to remember that these values are relative to context.
...
Basic Concepts of Logic
21
Problems Classify these sentences as valid, contradictory or contingent. 1. I am who I am. 2. All dogs are dogs. 3. Some dogs are not dogs.
4. Some cars are red.
5. All red automobiles are automobiles. 6. All red automobiles are red. 7. Every German car is a car. 8. Every German car is a German.
9. I know what I know. 10. Some people are friendly, and otOren't. 11. Some people are friendly and not friendly. 12. Some people aren't friendly, but everybody is friendly. 13. There are many trees in Yosemite National Park. 14. Every student studies. 15. Nobody loves everybody. 16. Everyone who loves everyone loves every loser. 17. Everyone who loves every loser loves everyone. 18. Everyone who drives a Mercedes drives a car. 19. Not everyone who drives a car drives a Mercedes. 20.
* If what you
21.
* Nobody can defeat everyone without
22.
* Today is the
23.
* No
24.
* You
25.
* Everything is what it is, and not another thing. (Plato) * The business of America is business. (Calvin Coolidge) * There are two kinds of people in the world: those who
26. 27.
say is true, then it's false. being defeated at least once.
first day of the rest of your life.
batter ever made a hit with the bat on his shoulder. (John McGraw) are what you eat. (Ludwig Feuerbach)
people, and those who don't. (H. L. Mencken)
divide the world into two kinds of
"
Logic, Sets and Functions
22 28.
*I
29.
* There comes a time to put principle aside and
30.
* I don't
31.
**
32.
* I exist.
33.
* Some dogs
34.
* No dogs
35.
* Most dogs are dogs.
36.
* Many dogs are dogs.
37.
* All former
Congressmen are Congressmen.
38.
* Some fake
diamonds are diamonds.
39.
* John
40.
* Alice taught
41.
* Say that
42.
* Say that
am never less alone than when I am alone, nor less at leisure than when I am at leisure. (Scipio Africanus) do what's right. (Michigan legislator)
know what the previous speaker said, but I agree with him. (Texas legislator)
this poem is the reader and the reader this poem (Ishmael Reed)
are dogs.
are dogs.
r", \ .)
..
likes baseball, but hates all sports. Sarah some chemistry, but Sarah learned no chemistry from Alice.
a sentence A implies another sentence B. What can we conclude about B, if A is (a) valid? (b) contingent? (c) satisfiable? (d) contradictory? a sentence A implies another sentence B. What can we conclude about A, if B is (a) valid? (b) contingent? (c) satisfiable? (d) contradictory?
The fourteenth century logician Pseudo-Scot (so-called because his writings, for many years, were attributed to John Duns Scotus) raised several objections to definitions of validity such as that of this section. The following two arguments, he thought, showed that there was a problem with saying that an argument is valid if and only if it's impossible for its premises to be true while its conclusion is false. Do these arguments really pose a problem for such definitions of validity? Explain. 43.
**
Every sentence is affirmative. :. No sentence is negative.
(Note: assume that no sentence can be both affirmative and negative. The argument seems valid. But, Pseudo-Scot argued, even though the premise could be true, the conclusion can't be; it refutes itself, since it, itself, is negative.) 44.
**
God exists. :. This argument is not valid.
(Note: Pseudo-Scot assumes that the premise is necessarily true. Any necessary truth would serve here in place of God exists. So, if the argument is valid, the conclusion must be true; but then the argument isn't valid, contradicting the hypothesis that it is valid. So suppose the argument is not valid. Then the conclusion is true; in fact, it must be necessarily true. In that case, we have an invalid argument in which it cannot happen that the premise is true while the conclusion is false, because the conclusion can never be false.)
Basic Concepts of Logic
1.5
23
Satisfiability
A sentence is satisfiable just in case it is not contradictory; that is, just in case it can be true. Any true sentence, obviously, is satisfiable. But false sentences can also be satisfiable, so long as they are true in some other possible circumstance. We can speak of sets of sentences, too, as satisfiable or contradictory. It's easy to think of sets of sentences that, in some sense, contain contradictions, even though each sentence in the set is itself satisfiable: 22.
(a) Beer and sauerkraut are very good together. (b) Beer and sauerkraut aren't very good together.
23.
(a) Many of my friends belong to the Flat Earth Society. (b) Nobody in the Flat Earth Society believes in modern science. (c) All my friends believe in modern science.
The sentences in (22), like those in (23), are not themselves contradictions. Taken individually, each could be true. Taken together, however, they describe an impossible situation. Though each could be true, the sentences in (22) or (23) couldn't be true together. In such cases, the set of sentences is contradictory, whether or not any individual sentence in the set is itself contradictory.
Definition 1.10 A set of sentences is contradictory if and only if it's impossible for all its members to be true. A set is satisfiable otherwise. If a set is contradictory, we can also say that its members are mutually inconsistent, and that any member contradicts, or is inconsistent with, the set containing all the rest. If the set is satisfiable, then its members are mutually consistent, and each member is consistent or compatible with the set containing all the rest. Two sentences contradict each other just in case the set containing just the two of them is contradictory. If a set is satisfiable, then all its subsets are satisfiable: each member is consistent or compatible with each other member of the set. From a logical point of view, contradictory sets of sentences can be described in two ways. First, the sentences in the set can't all be true at the same time. Second, the set implies a contradiction. Although a contradictory set of sentences might not contain a contradiction, it must imply one. The sentences in (21), for example, together imply 'Although many of my friends don't believe in modern science, all my friends do believe in modern science.' This is an outright contradiction. To see that these two characterizations come to the same thing, recall that a set S of sentences implies a sentence A just in case it's impossible for every sentence in S to be true while A is false. Contradictions, of course, are always false. When A is a contradiction, then, this amounts to the following: S implies A if and only if it's impossible for every sentence in S to be true. Therefore, a set of sentences implies a contradiction just in case it is itself contradictory. Or, to put it another way, satisfiability is freedom from contradiction. SatisfiabiIity is important: sets of sentences that are not satisfiable don't have a fighting chance at truth. They must contain at least one false sentence, no matter what the facts might be. A satisfiable set may also contain false sentences, but at least there is a possibility that all the sentences it contains are true. This explains the significance of satisfiabiIity in legal contexts. A lawyer may try to trap an opposing witness in a contradiction. The lawyer, in most cases, cannot alone provide any direct testimony relevant to the case. He or she may introduce witnesses of his or her own to dispute what the opposing witness says. If the opposing witness falls into a contradiction, however, then the witness must be saying something false, regardless of the facts of the case.
r';
Logic, Sets and Functions
24
Even more fundamentally, people often use arguments to disprove someone else's contention. To refute an assertion, we have to recognize when we have shown something that contradicts that assertion. So the notion of refutation depends on the notion of contradiction. Finally, the concept of satisfiability has been very important in modern mathematics. Around the turn of the century, several mathematicians and logicians deduced contradictions from mathematical theories in use at the time. Ever since, mathematicians have been extremely cautious about the satisfiability of their theories, and have sought, whenever possible, proofs that theories are satisfiable. This concern has led to some of the most important developments in twentieth-century logic, mathematics and computer science.
Problems Evaluate these sets of sentences as satisfiable or contradictory. 1. The yard isn't white unless it's snowing. It's not snowing. But the yard is white. 2. If a student's GPA is very high, he or she will get into a good graduate school. Frank's GPA is not very high. Nevertheless, he'll get into a good graduate school. 3. John is a good guitarist. John is also an accountant, but not a good one. 4. Everyone who can cook a good chicken kung pao knows the value of hot peppers. Some who know the value of hot peppers don't themselves like hot food. Anybody who can cook a good chicken kung poo likes hot food. 5. If Marsha takes a job with a state commission, she'll gain much experience in new areas, although she won't get to travel. If she takes a job with a private company, she'll get to travel, and she'll be paid well, although she won't gain much experience outside her area. Marsha won't be paid well, but she will get to travel. 6. If the Court's decision here is consistent with the decision in Yick Wo v. California, it will hold that statistical arguments alone can suffice to establish discrimination. If, however, it is compatible with the decision in several recent cases, it will hold that establishing discrimination requires something beyond purely statisical argumentation. 7. I like this painting, even though I don't think it's very good. I like everything that Elmer likes, and Elmer likes everying painting that's good. 8. Many Indo-European languages are descended from Latin. All languages descended from Latin developed a word for yes from the Latin sic (meaning thus). But few Indo-European languages have a word for yes developed from sic. 9. No drugs are approved for use without careful screening. Careful screening takes years. A few drugs in great demand, however, are approved for use in less time. 10. Few communist parties in Europe seek to identify themselves with the Soviet party. Parties seeking to identify themselves with the Soviets have a difficult time becoming part of coalition governments. Almost all European communist parties find it difficult, however, to become part of coalition governments. 11.
* Stocks of companies with
high debt-equity ratios are fairly risky. If a stock is fairly risky, it must reward investors with better-than-average returns, or they will eschew the risk. But many stocks that fail to reward investors with better-than-average returns are those of companies with high debt-equity ratios.
Basic Concepts of Logic
25
12.
* People have a right to life. Fetuses are not people. If something has a right to life, it wrong to kill it. Abortion is the killing of a fetus. Abortion is wrong.
13.
* Few contemporary composers write anything that could reasonably be called twelve-tone compositions. If so, then atonal music is defunct. But atonal principles of composition still exert some influence on contemporary composers. And nothing that still exerts influence is defunct.
14.
* Many football stars never graduate from the colleges where they first become famous. Most of these colleges insist that almost all their football players receive degrees. These schools are telling the truth.
15.
* Most actresses begin their careers as successful models. Every woman who begins her career as a successful model is very glamorous. Nevertheless, few actresses are very glamorous.
16.
* Many well-known American novels deal with the character of a specific region of the country. Every well-known American novel, of course, portrays a certain conception of America itself. Nonetheless, many novels that portray a conception of America do not deal with any specific region of the country.
17.
* My barber, who lives and works in town, shaves every man in town who doesn't shave himself. Furthermore, my barber doesn't shave anyone in town who does shave himself.
18.
** If 3 were an even number, then 4 would be odd.
But, if 3 were even, then 6 would be even,
and if 6 were even, then 4 would be even too. 19.
** If God exists, it's surely true that He exists necessarily. It's possible that God exists. It's also possible that He doesn't exist.
20.
** Because you promised, you ought to take your brother to the zoo. But, since you also have duties as club treasurer, you have an obligation to go to the club meeting. And you can't do both.
True or false? Explain. 21. If a set of sentences is satisfiable, no member of that set implies a contradictory sentence. 22. If no member of a set implies a contradictory sentence, that set is satisfiable. 23. Some satisfiable sets of sentences imply contradictions. 24. Some satisfiable sets of sentences imply no contingent sentences. 25. Every contradictory set of sentences implies every contradiction. 26. No satisfiable sets of formulas imply every sentence. 27. Some contradictory sets of sentences imply every sentence. 28. If A implies B, then the set consisting of A and B together is satisfiable. 29. If the set consisting of just A together with B is contradictory, then A implies that B is false. 30. Any argument with a contradictory set of premises is valid. 31. Arguments with satisfiable sets of premises have satisfiable conclusions.
r
26
Logic, Sets and Functions
32. Every satisfiable set of sentences contains at least one true sentence. 33. Every contradictory set of sentences contains at least one false sentence. 34. Any set consisting of all valid sentences is satisfiable. 35. Any set consisting of all contingent sentences is satisfiable. The Englishman William of Ockham (1285-1349), perhaps the most influential philosopher and logician of the fourteenth century, recorded eleven rules of logic in a chapter of his Summa Totius Logicae. Ten of these use concepts we've already developed. Say whether each is true, given the definitions of this chapter, and explain why. 36. The false never follows from the true. 37. The true may follow from the false. 38. Whatever follows from the conclusion of a valid argument follows from its premises. 39. The conclusion of a valid argument follows from anything that implies the argument's premises. 40. Whatever is consistent with the premises of a valid argument is also consistent with the argument's conclusion. 41. Whatever is inconsistent with the conclusion of a valid argument is also inconsistent with the argument's premises. 42. The contingent does not follow from the valid. 43. The contradictory does not follow from the satisfiable. 44. Anything whatsoever follows from the contradictory. 45. The valid follows from anything whatsoever.
Chapter 2
Sentences Sentential logic examines the relationships between sentences that pertain to reasoning. This chapter will focus on only a portion of that logic. First, attention will be restricted to sentences that are either true or false, such as 1. Nome is in Alaska.
2. No other university in the United States is as old as Harvard. 3. If Michael Jackson had been born a woman, he'd have wanted to marry himself. We will ignore sentences such as 4. Ouch! 5. Shut the door. 6. Does anybody really know what time it is? which cannot be true or false, though they may be classified as appropriate or inappropriate, wise or unwise, etc. Second, we'll develop a theory only of the connections between sentences that are truthfunctional. These connections have a rather neat logical character. The truth values of compound sentences formed by way of them depend entirely on the truth values of the smaller sentences they connect. The result of forming these connections is thus completely predictable on the basis of the truth values of the smaller sentences alone.
2.1
The Language of Sentential Logic
Sentences, we've seen, can be parts of other sentences. Sentences that appear within other sentences are often called embedded or subordinate clauses. In this section we'll discuss a symbolic language designed to clarify the structure of some of these embedding mechanisms. The language of sentential logic does not treat complemented verbs- indeed, their logic is extremely complex- but it can represent many of the expressions that operate on one or more sentences to produce a new sentence.
Definition 2.1 An n-ary sentence connective is a word or phrase that forms a single, compound sentence from n component sentences. 27
,...
28
Logic, Sets and Functions
Sentence connectives constitute the chief object of study for sentential logic. Some examples of singulary connectives are not, maybe, of course, possibly, necessarily, and- somewhat controversiallyauxiliary verbs such as may, can, could, might, must, should, etc. Examples of binary connectives are and, but, however, although, if, or, unless, though,bejore and because. Some of these connectives behave very predictably, in a certain sense, while others are more complex. Valid arguments never lead us from truths to falsehoods; in every circumstance in which the premises are true, the conclusion is true as well. In trying to construct a theory of validity, therefore, it seems natural to focus on the concepts of truth and falsehood. When I say that some connectives are predictable, I mean that their effects on truth values are predictable on the basis of truth values. That is, predictable connectives form compound sentences whose truth values are a function of the truth values of their components. Predictable connectives are called truth-junctional.
Definition 2.2 An n-ary sentence connective is truth-functional if and only if the truth values of the n component sentences always completely determine the truth value of the compound sentence formed by means of the connective.
To put this another way, if a connective is truth functional, then compounds formed from it match in truth value whenever the truth values of their components match in truth value. Knowing whether the component sentences are true or false is all one needs to know the truth value of the compound. Consider, for instance, the sentence 'It's not snowing.' This contains the component 'It's snowing.' Suppose that this component is true. Then the compound is false; if it's true that it's snowing, it's false that it's not snowing. Suppose that the component is false. If it's false that it's snowing, then it's true that it's not snowing, so the compound is true. The truth value of 'It's snowing,' in other words, completely determines the truth value of 'It's not snowing.' Nothing about this depends on any special feature of the sentence 'It's snowing.' Any sentence in this role should produce much the same result. 'Not' is thus a truth-functional sentence connective. In contrast, consider 'after.' This connective is not truth-functional. To see why, consider a sentence such as 'George resigned after the commissioner stopped the trade.' This contains the components 'George resigned' and 'The commissioner stopped the trade.' Suppose that both these sentences are true. That does not determine whether the compound sentence is true. George, after all, may have resigned before the commissioner nixed the trade. Note that, if either or both of the component sentences are false, the compound must be false; if George didn't resign, then he didn't resign after the commissioner stopped the trade. If the commissioner didn't stop the trade, then did George resign after he stopped the trade? Here, matters are less clear, but the compound sentence, if not false, at least is not true, and is highly misleading. So the fact that the truth values of the components sometimes suffice to determine the compound's truth value isn't enough; they must suffice in every case if the connective is to be truth-functional. We can summarize these examples readily with the help of a small table. Consider the possible truth values of the sentences involved, and see if, in every case, then truth value of the compound is fully determined. This results in the tables: It's snowing It's not snowing T F
F
...
T
Sentences
29
George resigned
The commissioner stopped the trade
T
T
T F
F T
George resigned after the commissioner stopped the trade ? F, or misleading F
F
F
F
The letters T and F here abbreviate true and false. A connective is truth-functional just in case it's possible to fill in every entry under the compound sentence with a determinate truth value. If, as with 'after,' there is any row that can't be filled in with either T or F , then the connective is not truth-functional.
Problems Are these connectives truth-functional? Why, or why not? 1. because 2. or 3. before
4. nevertheless 5. may
6. in order that 7. in spite of 8. regardless whether 9. implies 10. should 11. it's improbable that
12. it's a logical truth that 13. can 14. could 15. maybe 16. it's obvious that 17. it's surprising that 18. when 19. provided that 20. if
,..
30
2.2
Logic, Sets and FUnctions
Truth Functions
We can think of sentence connectives as functions that take sentences as inputs and yield compound sentences as outputs. Truth-functional connectives yield outputs whose truth values depend solely on the truth values of the inputs. For each truth-functional connective, then, there is a corresponding function from truth values into truth values. They take as inputs the truth values of the component sentences and yield the truth value of the compound sentence. Such functions are called truth functions.
Definition 2.3 An n-ary truth function is a function taking n truth values as inputs and producing a truth value as output. There are four singulary truth functions. Such a function takes a single truth value as input, and produces a truth value as output. There are only two candidates for inputs- truth and falsehoodand, similarly, only two candidates for outputs. One of the four singulary truth functions takes both truth and falsehood into truth; another takes both into falsehood. One takes each value into itself; another takes them into each other. That's all the possible singulary functions. There are 16 binary 2n functions, and, in general, for each n,2 truth functions. Since we may make n as large as we like, there are infinitely many truth functions all together. How can we formulate a theory to describe this infinite array of functions? Luckily, it's not very difficult. I'll present a few commonly used truth functions. Any truth function at all can be defined in terms of them alone. In fact, as I'll explain, a single binary truth function suffices to define every truth function in this infinite collection. The first function I'll define is singulary. It is called negation, and I'll use the symbol -, to represent it. (This symbol has no name in the literature; I'll call it the hoe.) In this definition, A is any sentence or formula. (Formulas will correspond to sentences in our symbolic language.)
A -,A T F
F T
Negation transforms the truth value of the component sentence into its opposite. We've already seen an English connective that has this effect: the logical particle not. Other English expressions having much the same impact are it is not the case that, it's false that, no, never, and, often, the prefixes un-, dis-, a-, im-, etc. The second function is binary. Called conjunction, I'll represent it with the ampersand, & .
A B (A&B) T T T T F
F T
F F
F F F
A conjunction is true just in case both its components- called conjuncts- are true. English expressions functioning in this way include most grammatical conjunctions: and, both. . . and, but, though and although. The third function is also binary. Represented by v, the wedge, it's called disjunction.
Sentences
31
A T T F F
B T F T F
(A
V
B)
T T T F
A disjunction is true just in case either of its components- called disjuncts - is true. English expressions corresponding to this function are or, either. . . or and unless. The fourth function, again binary, is represented by the arrow, -., and is called the conditional:
A B T T T F
(A
-->
B)
T F T T
F T
F F
Here, for the first time, the order of the components makes a difference. The first component of a conditional is its antecedent; the second is its consequent. A conditional is true just in case it doesn't have a true antecedent and false consequent. English expressions having, roughly, the force of the conditional truth function are B if A, if A then B, A only if B, B so long as A, B provided that A, B assuming that A, B on the condition that A, A is a sufficient condition for B, and B is a necessary condition for A. Finally, the biconditional is a binary truth function, symbolized by ........
A B T T T F
F T
F F
(A ....... B) T F F T
Biconditionals are true just in case their components agree in truth value. English expressions such as if and only if, when and only when, is a necessary and sufficient condition for, and just in case correspond to the biconditional. The symbols -, , & , V, -. and ....... are in common use. Unfortunately, however, there is no standard logical notation. This table shows other symbols that have been used as logical connectives: Truth Function Negation Conjunction Disjunction Conditional Biconditional
Our Symbol -,p P&Q PvQ P-+Q
P ....... Q
Other Symbols _P,""" P, pI, P P A Q, PQ, p. Q, K PQ PY-Q,APQ P::l Q, CPQ P-=Q,P,....,Q,EPQ
These five truth functions have enough expressive power to allow us to construct any truth function from them. Indeed, not even all five are necessary. Negation and conjunction alone suffice; so do negation and disjunction, or negation and the conditional. These sets of truth functions{-,,&}, {-', V}, and {-,,-+}- are functionally complete. Two binary truth functions are functionally complete all by themselves. They are called Sheffer functions, after Henry Sheffer (1883-1964), a Harvard logician. (Actually, they were discovered by
".
32
Logic, Sets and Functions
the American philosopher Charles Sanders Peirce around 1880, but went unnoticed until Sheffer reproduced the discovery in 1913.) These functions are Sheffer's stroke (j) and nondisjunction ( 1 )
A B T T F
T F T
F F
(A/B) F T T T
(A 1B) F
F F T
English correlates of Sheffer's stroke are not both. . . and, is incompatible with, and is inconsistent with. The chief correlate of nondisjunction is neither. . . nor. The existence of these functions is quite intriguing; computer logic gates equivalent to them (called NAND and NOR gates) have been extremely important in the development of computer logic design. Nevertheless, we will generally restrict our sentential language to the five connectives first presented. The English correlates to truth functions are often only approximately equivalent in meaning to them. The correlates to negation express that function quite faithfully, except that English seems to admit some intermediate, "fuzzy" ranges, while the negation truth function does not. If Uncle Fred has only a small amount of hair on his head, for example, it may not be quite correct to say that he's bald, but it's not exactly false either. Cousin Fran may not really be happy these days, but she's not really unhappy either. And so on. Another problem arises with sentences involving attitudes, which sometimes seem to have presuppositions. If Roger Staubach never had to serve time in Huntsville prison, then neither 'Roger Staubach regrets that he had to serve time in Huntsville prison' nor 'Roger Staubach doesn't regret that he had to serve time in Huntsville prison' seem to be true. Similarly, if Yogi Berra has never thought much about Spinoza, then neither 'Yogi Berra believes that Spinoza was a great philosopher' nor 'Yogi Berra doesn't believe that Spinoza was a great philosopher' seem true. The correlates of conjunction are very faithful to it, except that and sometimes seems to express a temporal ordering. There is a difference in meaning between 'Heidi got married and got pregnant' and 'Heidi got pregnant and got married', even though the definition of conjunction indicates that these should be true in exactly the same circumstances. Also, English treats conjunction as a multigrade connective, that is, one that can take two, three, or more sentences and combine them into a single compound sentence. The logical & always links two sentences, but in English we can say' John brought the mustard, Sally brought the pickles, and I brought the hot dogs.' The correlates to disjunction are also quite close to the logical definition. Or, too, appears to be multigrade in English. The biggest surprise is that unless has the same meaning, logically, as aT. Consider the sentence: 'The patient will die unless we operate.' This is equivalent to 'Either we operate, or the patient will die.' According to the definition, this should be true if we operate, and true if the patient dies. Yet the English sentence seems to assert that there is a connection, usually causal, between the operation and the patient's chances for survival. We wouldn't normally count the sentence true, just because the operation was performed, or just because the patient died. In fact, a surgeon who said this when he knew that the patient was going to die in any case seems to be unethical. This becomes even more acute with the conditional. At first it seems surprising that A only if B and If A, then B are both correlates of the conditional (A ---> B). But both these sentences seem equivalent to the earlier examples and to each other: 'If we don't operate, the patient will die' and 'The patient will survive only if we operate.' These, of course, are not quite the same: one is (-,A ---> B) while the other is (-,B ---> A). Switching the negations and the English connectives, furthermore, yields sentences that ought to be equivalent to these, logically speaking, but sound bizarre: 'If the patient survives, we operate' and 'We don't operate only if the patient dies.' The
Sentences
33
first makes it sound as if we operate only when the patient is out of danger, and the second makes it sound as if we operate as long as we have a live body to operate on. The conditional doesn't capture these differences, and for the same reason that disjunction deviates from unless. Whatever follows the English words if and unless states a condition, usually causal, on the other component. So there is a strong feeling that whatever happens to make the condition true should happen before what happens to make the other component true. The then of if. . . then, in other words, has some temporal meaning, and our logical rendering of the conditional can't capture it. The same problem, of course, occurs doubly with the biconditional. These problems with the truth-functional rendering of the conditional have been recognized since the third century B.C., when Philo of Megara (a classmate of Zeno, the founder of Stoicism) first proposed such an analysis. Diodorus, Philo's teacher, held that conditionals involve necessity, and Chrysippus, the third head of the Stoic school and widely acclaimed as the greatest logician of his time, held that the sort of necessity involved is specifically logical, as opposed, for instance, to causal necessity. The controversy among their followers became so intense that Callimachus wrote that "even the crows on the roofs caw about the nature of conditionals." Philo's treatment has the advantage of being truth-functional. It is also the weakest analysis that preserves the inference patterns that everyone agrees should characterize the conditional. Consequently, many other analyses of the conditional can be defined in terms of the Philonian conditional together with another connective, such as necessity. Finally, Philo's truth-functional construal of the conditional has proved valuable for expressing sentences such as 'All humans are mortal' in quantificational logic.
Problems Explain why a connective in each of these sentences deviates somewhat from the truth function with which it is generally correlated. 1. Mary fell down and got up. 2. My heart rate approaches 200 if and only if I do heavy exercise. 3. Only if you come to me on your knees will I let you out of the contract. 4. Give me a place to stand, and I will move the earth. (Archimedes) 5. If the moon smiled, she would resemble you. (Sylvia Plath) 6. Why, if 'tis dancing you would be, there's brisker pipes than poetry. (A. E. Housman) 7. It is not the case that if God is dead, life is meaningless. 8. I'll leave only if you have somebody to take my place. 9. If Mike clears his throat once more, I'll strangle him. 10. Fame and rest are utter opposites. (Richard Steele)
2.3
A Sentential Language
Aristotle, the first logician, was also the first logician to use symbols. The Stoics made the practice common, using numbers to stand for sentences. Nevertheless, no one developed a fully symbolic logical language until the nineteenth century, when George Boole (1815-1864) saw that logic could profit from mathematical analysis. (His greatest work, in fact, he entitled The Mathematical Analysis of Logic.) In this section we'll present a symbolic language that permits ready evaluation of
'"
34
Logic, Sets and Functions
arguments depending on sentence connectives. The syntax of this language is fairly simple. The vocabulary falls into three basic categories: sentence letters (or variables), connectives, and grouping indicators.
Vocabulary Sentence letters: A, B, C, ... , Z with or without numerical subscripts Connectives: ...." &, V, --., and ~ Grouping Indicators: ( and) The syntactic rules that allow us to combine these elements to form meaningful strings of the language can be presented in several ways. First, we'll define the notion of a meaningful string- a formula - as logicians usually do. Then we'll present some phrase structure rules that allow us to see the similarity between the syntax of this artificial language and the syntax of a natural language such as English.
Formation Rules Any sentence letter is a formula. If A is a formula, then ..... A is a formula. If A and 13 are formulas, then (A&13), (A V 13), (A --.13) and (A ~ 13) are too.
Every formula can be constructed by a finite number of applications of these rules. In presenting the above definitions and rules we've assumed that sentence connectives and other vocabulary items of our symbolic language are names of themselves. (A --. 13) thus refers to the formula resulting from concatenating a left parenthesis, the formula A, an occurrence of the conditional connnective, the formula 13, and a right parenthesis. Notice that we can't refer to this formula by using ordinary quotation: '(A --. 13)' both signifies the string of symbols quoted, i.e., a left parenthesis, followed by the calligraphic letter A, followed by an occurrence of the conditional, followed by the calligraphic letter 13, followed by a right parenthesis. This is not a formula of our sentential language; calligraphic letters never appear as sentence letters. It is, instead, a part of the language we use to talk about the sentential language. We'll call calligraphic letters such as A and 13 schematic letters, and strings such as (A --. 13) schemata. We will adopt three conventions for using our symbolic language. These conventions allow us to work with the language more easily. First, we'll count brackets- [ and I - and braces- { and }- as poorly drawn parentheses, letting them act as grouping indicators when that makes a formula more readable. Second, we'll allow ourselves to drop the outside parentheses of a formula, since they do no further work. One can imagine them drawn in so lightly that we can't see them. Thus, we'll count these pairs of symbol strings as equally acceptable as formulas, even though only those on the left are formulas, strictly speaking. (P --. Q) P --. Q (((P --. Q) --. P) --. P) {(P --. Q) --. P} --. P We'll close this section with two definitions that will prove extremely useful throughout the entire text. An occurrence of a connective forms a compound sentence from one or more component sentences. The connective occurrence itself, together with these components, constitute the scope of the connective occurrence in a given formula. The connective occurrence with the largest scope is the main connective of the formula. Definition 2.4 The scope of a connective occurrence in a formula is the connective occurrence itself, together with the components (and any grouping indicators) it links together in the formula. Thus the scope of these connective occurrences, in these formulas, is:
Sentences --, --,
V
&
35 (P --> --,Q) : --,(P --> Q) : (P V (Q&R)) : (P V (Q&R)) :
in in in in
--,Q --,(P --> Q) (P V (Q&R)) (Q&R)
Definition 2.5 The main connective of a formula is the connective occurrence in the formula with the largest scope. The main connective is always the connective occurrence having the entire formula as its scope.
Problems Classify each of these as (a) an official formula, (b) a conventional abbreviation of a formula, or (c) neither. If you answer (a) or (b), identify the main connective. 1. P&--,Q
2. --,PV--,Q
3. --,(P&--,Q) 4. (--,P V --,Q) 5. (&P V &Q) 6. (P V Q) 7. P V Q
R
-->
-->
R
8. (PV Q
-->
R)
9. P V (Q
-->
R)
10. (PVQ)
R)
-->
11. «P V Q)
-->
A)
12. --,«P V (Q V R))&Q 13. -.[P 14. M
-->
-->
15. -.«P
(R
-->
--,Q) ...... -.Pj
[R ...... {KVL}] -->
(Q
-->
R)) V P
16. The rules for forming formulas of sentential logic admit no syntactic ambiguity. How could these rules be altered to admit ambiguity? Give an example of a set of rules allowing syntactic ambiguity, and specify an ambiguous formula.
""
36
Logic, Sets and Functions
2.4
Translation
We can use the symbolic language we've constructed to help evaluate arguments in natural language only if we can translate from English into the symbolic language. Luckily, this is not usually very difficult. To translate an English discourse into symbolic notation, • Identify sentence connectives; replace them with symbolic connectives. • Identify atomic sentential components of the English sentences, and replace each distinct component with a distinct sentence letter. • Use the structure of the English sentence to determine grouping. Several factors complicate these steps. The first step- identify and replace sentence connectivesrelies on the correlation we have outlined between truth functions and certain English sentence connectives. As we've noted, these correlations are not always exact. So some linguistic sensitivity needs to accompany this process. Step two- identify and replace atomic sentences- is complicated chiefly by the fact that English arguments try not to be repetitive or dull. Rarely will an author use even a component sentence twice in the same passage without varying the wording of it. This means that deciding when two atomic sentence components are "distinct" or "the same" is no easy matter. Here, again, linguistic sensitivity is needed. Components should be judged "the same" only if they have basically the same meaning. The third step- determining the grouping- is, in a sense, the trickiest of all. Our symbolic language avoids ambiguity by using parentheses as grouping indicators. English doesn't use parentheses in this way, and so natural language sentences are sometimes ambiguous. But English offers a number of devices for making grouping clear. One is the use of commas. The English sentence 'John will come and Fred will leave only if you sign' has no very clear grouping. But, by placing a comma judiciously, we can make it clear that we intend and to be the main connective: 'John will come, and Fred will leave only if you sign.' Basically, a comma emphasizes a break in the sentence, stressing the presence of two different phrases being combined in the sentence. Commas therefore tend to suggest that the nearest connective has some priority. Second, English offers cooordinate phrases such as either. . . or, both. . . and, and- if . . . then. These can make clear the grouping by identifying the components clearly. 'Either Bill brought Mary and Susan brought Sam, or Susan brought Bob' makes the intended grouping clear. The coordinated connective takes priority over any connective appearing within the coordinated sections. Note how much clearer than 'Bill brought Mary and Susan brought Sam or Susan brought Bob' the above sentence is. Third, English allows a device that logicians have called "telescoping" and that linguists call "conjunction reduction". 'Susan brought Sam or Susan brought Bob', for example, can be. "reduced" to the shorter sentence 'Susan brought Sam or Bob'. Similarly, 'Fred likes Wanda and Kim likes Wanda' can be reduced to 'Fred and Kim like Wanda'. This too can clarify grouping: we could group the last paragraph's sentence in another way by saying 'Bill brought Mary and Susan brought Sam or Bob'. Problems Translate each of these sentences into our symbolic language. If any translate only with difficulty, explain why. 1. Kindness is in our power, bur fondness is not. (Samuel Johnson)
2. My father taught me to work, but not to love it. (Abraham Lincoln)
Sentences
37
3. And April's in the west wind, and daffodils. (John Masefield)
4. But if you wisely invest in beauty, it will remain with you all the days of your life. (Frank Lloyd Wright) 5. If you get simple beauty and naught else, you get about the best thing God invents. (Robert Browning) 6. A man can hardly be said to have made a fortune if he does not know how to enjoy it. (Vauvenargues) 7. It is not poetry, if it make no appeal to our passions or our imagination. (Samuel Taylor Coleridge) 8. The forces of a capitalist society, if left unchecked, tend to make the rich richer and the poor poorer. (J. Nehru) 9. Honesty pays, but it don't seem to pay enough to suit a lot of people. (K. Hubbard) 10. If you don't get what you want, it is a sign either that you did not seriously want it, or that you tried to bargain over the price. (R. Kipling) 11. Honesty is praised, but it starves. (Latin proverb) 12. It does not do to leave a live dragon out of your calculations, if you live near him. (J. R. R. Tolkein) 13. I am an idealist. I don't know where I'm going but I'm on my way. (Carl Sandburg) 14. Size is not grandeur, and territory does not make a nation. (Thomas H. Huxley) 15. And the light shineth in the darkness; and the darkness comprehendeth it not. (John 1:5) 16. If you are patient in one moment of anger, you will escape 100 days of sorrow. (Chinese proverb) 17. It is only by painful effort, by grim energy and resolute courage, that we move on to better things. (Theodore Roosevelt) 18. If we are to preserve civilization, we must first remain civilized. (Louis St. Laurent) 19. Miracles sometimes occur, but one has to work terribly hard for them. (Chaim Weizmann) 20. Happiness is not the end of life, character is. (Henry Ward Beecher) 21. A sense of duty is useful in work but offensive in personal relations. (Bertrand Russell) 22. Loafing needs no explanation and is its own excuse. (Christopher Morley) 23. So then neither is he that planteth any thing, neither he that watereth; but God that giveth the increase. (I Corinthians 3:7) 24. In cases of difficulty and when hopes are small, the boldest counsels are the safest. (Livy) 25. The mind itself, like other things, sometimes be unbent; or else it will be either weakened or broken. (Philip Sidney) 26. I could not love thee, dear so much/ Loved I not honor more. (Richard Lovelace)
~
Logic, Sets and Functions
38
27. Nothing will ever be attempted if all possible objections must first be overcome. (Jules W. Lederer) 28. A man doesn't need brilliance or genius; all he needs is energy. (Albert M. Greenfield) 29. It's not good enough that we do our best; sometimes we have to do what's required. (Winston Churchill) 30. Action does not always bring happiness, but there is no happiness without it. (Benjamin Disraeli)
2.5
Validity
Our tactic for analyzing English arguments and sentences is twofold. First, we will translate the sentences or arguments into our symbolic language. Second, we will evaluate the results for validity, satisfiability, etc. The clarity and convenience of the symbolic language pays big dividends whenever logical problems become at all complicated. So far, we have set up our language and looked at ways of translating English into it. This solves only part of the problem, however. We want to be able to assess the validity, or satisfiability, or whatever, of the logical formulas we obtain through translation. But the notions of validity, etc., as we've defined them, make no sense when applied to formulas. To see why, consider a simple formula, P. Is this formula valid? That is, is it true in every possible world? It doesn't even make sense to speak of P as true or false in our world, since it is not a sentence, with a determinate meaning, but simply a sentence letter. The same thing happens to arguments. Sequences of sentences, when translated into a symbolic language, become sequences of formulas, or argument forms. Definition 2.6 An argument form is a finite sequence of formulas. The last is the conclusion formula; the others are premise formulas.
Neither premise nor conclusion formulas can be true directly in any possible world. Consequently, we need to reformulate our basic semantic concepts for formulas. The first person to do this successfully was Bernard Bolzano (1781-1848), a professor at the University of Prague. Formulas are not directly true or false, in this or any other possible world. They are, however, true or false under some interpretation. If we decide that P represents the sentence 'Frank Langella plays a good Dracula', then we can assess the truth value of this sentence, and, so, speak of P as either true or false under that interpretation. Clearly, P could represent either a true or a false sentence, so it should be a contingent formula. The formula P - t P, on the other hand, will represent a truth, whether P stands for 'John is smiling', 'Bruce Wayne is Batman' or 'Ice cream and strawberries are a tasty combination. This formula, then, should be valid- it comes out true no matter how it is interpreted. In speaking of "interpretations" here, we mean two things. First, we can interpret a sentence letter by specifying that it represents a certain English sentence. Second, we can interpret it more directly, from the perspective of sentential logic, by simply assigning it truth or falsehood. All the analysis we are currently prepared to do uses nothing about a sentence beyond its truth value. So we can interpret a sentence letter, from the point of view of the logic of sentences, by saying whether the sentence letter is true or false. We can interpret a formula, in general, by saying whether the sentence letters it contains are true or false. Definition 2.7 A n interpretation of a sentence letter is an assignment of a truth value to it. An interpretation of a formula of sentential logic is an assignment of truth values to its sentence
Sentences
39
letters. An interpretation of an argument form or set of formulas in sentential logic is an assignment of truth values to all the sentence letters in the argument form or set. Now we can easily formulate definitions of validity and other semantic notions for formulas and argument forms. They correspond very closely to the definitions for sentences and arguments. Definition 2.8 A formula is valid if and only if it's true on every interpretation of it. A formula is contradictory if and only if it's false on every interpretation of it, satisfiable if and only if it's true on at least one interpretation of it, and contingent if and only if it's neither valid nor contradictory. Definition 2.9 An argument form is valid if and only if there is no interpretation of it making its premise formulas all true and its conclusion formula false. Definition 2.10 A set of formulas S implies a formula A if and only if there is no interpretation of S together with A making every member of S true but A false. Definition 2.11 Two formulas are equivalent if and only if they agree in truth value on every interpretation of them. Definition 2.12 A set of formulas is satisfiable if and only if, on some interpretation of it, every member of the set comes out true; the set is contradictory otherwise. These definitions are almost exactly the same as those for sentences and arguments, but with "possible worlds" replaced by "interpretations". Though our method of translating sentences and arguments into a symbolic language is fundamental to modern logic, one must be aware of its limitations. If a translation is a good translation, and if the argument form, say, that results is valid, then the original argument must have been valid. But, even if the translation is good, an invalid argument form doesn't show that the original argument was invalid. Many English arguments depend on logical relationships that sentential logic doesn't capture. An appropriate form for that argument in sentential logic, then, may be invalid, even though the argument is valid. Because of the nature of our method, therefore, "valid" verdicts merit more trust that "invalid" verdicts. It's always possible that, to establish the validity of an argument, we would have to bring out bigger guns than are available in the logical theory we're working in.
Problems Classify these sentences as true or false, and explain. 1. Valid argument forms having valid conclusion formulas are valid.
2. Some argument forms with contradictory premise formulas aren't valid. 3. There is a formula that implies every other formula. 4. There is a formula that is equivalent to every other formula. 5. No satisfiable sets of formulas imply every formula. 6. Any formula that follows from a satisfiable formula is satisfiable. 7. Any formula that implies a contingent formula is not valid. 8. Any formula that follows from a contingent formula is contingent.
r
40
Logic, Sets and Functions 9. Any formula that follows from a valid formula is valid.
10. Any formula that implies a valid formula is valid. 11. All contradictory formulas imply one another. 12. All contingent formulas imply one another. 13. All valid formulas imply one another. 14. If a set of formulas is satisfiable, no member of that set implies a contradictory formula. 15. If no member of a set of formulas implies a contradictory formula, that set is satisfiable. 16. If a set of formulas implies no contradictory formula, that set is satisfiable. 17. Some argument forms with contradictory formulas as conclusion formulas are valid. 18. No formula implies its own negation. 19. Any formula that implies its own negation is contradictory. 20. Any formula implied by its own negation is valid. Bolzano calls an argument form exact or adequate if and only if (1) it is valid and (2) the conclusion formula does not follow from any proper su bset of the premise formulas. (That is, omitting any premise formula or formulas makes the argument form invalid.) In such an argument form, the premise formulas exactly imply the conclusion formula. Are these assertions about Bolzano's concept true or false? Explain. 21. No exact argument form has a valid formula as conclusion. 22. Some exact argument forms have valid formulas as premises. 23. Some exact argument forms have contradictory conclusion formulas. 24. Some exact argument forms have contradictory premise formulas. 25. Some exact argument forms have contingent conclusions. 26. Every formula exactly implies itself. 27. Some formula exactly implies every formula. 28. Some formula is exactly implied by every formula. 29. Sets of contingent formulas exactly imply only contingent formulas. 30. No formula exactly implies its own negation.
Sentenc~
2.6
41
Truth Tables
The language of sentential logic has the power to express forms of very many English arguments. Merely expressing these arguments, however, is only a part of the task. To evaluate arguments in natural language, we must also find ways of evaluating argument forms as valid or invalid. The same point holds, of course, for sentences and set of sentences. The strategy underlying the use of symbolic logic has two components: translation into a symbolic language, and evaluation of formulas, argument forms, etc., in that symbolic language. So far in this chapter we have discussed only the first component. In the remaining three sections of this chapter, we will present a method for testing sentential argument forms for validity. The same method can also test for satisfiability, equivalence, implication, and validity of formulas. The method we're about to present, that of truth tables, is simple but powerful. Although it becomes cumbersome when dealing with complex problems, it can in principle answer any question we might raise about validity, implication, satisfiability, etc., within the bounds of sentential logic. To make this more precise: truth tables constitute a decision procedure for all the logical properties and relations we've discussed. Definition 2.13 A decision procedure for a property P is a mechanical, infallible method for determining whether any given object has the property P. (Similarly, a decision procedure for a relation R is a mechanical, infallible method for determining whether any given objects stand in the relation R.) The truth table technique is mechanical in the sense that no particular ingenuity is required to implement it. One needs only to follow the rules. In this respect, constructing a truth table is like dividing two numbers, or performing some other elementary arithmetical operation, but very unlike painting a portrait, writing a short story or getting a job. The truth table technique is also infallible: it always yields a correct answer after a finite period of time (at least, if it is constructed correctly). One can think of mechanical and infallible procedures in another way: th~e are the procedures that a computer can implement. Theoretically, a computer can perform any mechanical procedure. If the procedure is not infallible, however, the computer may run on forever without producing an answer. Practically speaking, therefore, mechanical and infallible procedures- often called effective procedures- are uniquely suited to computerization. Because there is a decision procedure for validity and the other basic semantical notions of sentential logic, the set of valid formulas, the set of satisfiable formulas, the set of valid argument forms, etc., are all said to be decidable. Definition 2.14 A set is decidable if and only if there is a decision procedure for membership in the set. To put this another way, a set is decidable just in case there is a mechanical, infallible test for determining whether any given object belongs to the set or not. Any finite set is decidable, because it's possible to list its members. To test a given object for membership, one can simply search through the list. This tactic isn't any help when the set is infinite, however. So only when we deal with infinite sets do we need to worry about decidability. Having a decision procedure for a set- or a theory, since a theory is a set of sentences, together with all their logical consequencestells us a great deal about that set or theory. The question whether a decision procedure exists for various theories and sets has thus taken on substantial importance in 20th-century logic. In fact, this question, when applied to a particular theory or set, has been given a special name: the decision problem for that theory or set.
~
Logic, Sets and Functions
42
Earlier in this chapter, we defined five connectives that form the heart of our symbolic language. These definitions themselves have the form of small truth tables:
A
-,A A
T F
F T
B
(A V B)
(A&B)
T F T
T T T
F F
F
T F F F
T T F
(A
~
B)
T
F T T
(A
f->
B)
T F F T
These definitions specify the values of complex formulas on the basis of the values of their components. It is fairly easy to see that this allows us to compute the truth value of a formula of any length, given the values of the atomic formulas- i.e., sentence letters- appearing in the formula. (Actually, in the history of logic, this was not so easy to see: Boole and Gottlob Frege, the German mathematician and logician who first invented quantification theory, both defined connectives using tables like those above in the nineteenth century. But not until around 1920 did Ludwig Wittgenstein, J. Lukasiewicz and Emil L. Post recognize, independently, that the tabular definitions suffice for a computation of the truth values of complex formulas.) To take a simple example, suppose that P and Q are true, but R is false. What is the truth value of the formula ((P ~ Q) ~ R? The main connective of this formula is the second conditional. The antecedent of this conditional is another conditional, both of whose components are true; the definition of the conditional indicates that this smaller conditional is true. The larger conditional, then, has a true antecedent, but its consequent, R, is false. The definition of the conditional tells us that the larger conditional is therefore false. Approaching this problem more systematically, we might list the values of the sentence letters first, and then proceed to generate the values of subformulas until we reach the formula as a whole. The structure of the formula on the left, then, would lead us to the table on the right, and, perhaps, to the more compressed table below it: ((P ~ Q) ~ R) (P ~ Q) R P Q
P
T
Q T
R
F
(P~
T
Q)
T T
((P ~ Q) ~ R) T T T F F
P Q R ((P ~ Q) ~ R) TTF TTTFF A boldface letter in the tables represents the truth value of the entire formula above it. Notice that, in constructing these tables, we begin with sentence letters. We then assign values to items one level up, and work our way, gradually, to the top level, where the formula itself appears. This has the effect that we work from inside parentheses out. Whenever two subformulas are equally "deep" inside parentheses, it makes no difference which we attack first. Our definitions thus allow us to compute the value of a formula, given some interpretation assigning values to its sentence letters. This is the central idea behind truth tables.
Problems Oalculate the truth values of these formulas on the interpretations listed. ~.,
.1.
(}P V Q) V (P V -,R)
(P and Q false; R true)
....'Wlt.,r;'"
2. (P ~ (Q 3. ((R&Q)
f->
~
R)) ~ (Q V -,R) (P and R true; Q false)
-,P)&(-,Q
f->
P) (P true; Q and R false)
4. -,((P ~ -,Q) V (Q V -,R)) (P false; Q and R true)
Sentences
43
5. (P ....... ---,R)
6. ---,(P
V
->
(R
Q)&---,(P
7. (P V (Q&R))
->
->
->
Q) (P, Q and R true)
R) (P, Q and R false)
(P&Q) (P and Q true; R false)
/'8. (P&Q) ....... ---,R (P and R true; Q false)
<·~9·.·'~( ---,( ---,P&---,Q)&---,R) 10. ---,«Q V ---,P)
->
12. (P V ---,Q)
14. ---,«P 15. (P
2.7
->
->
---,R) (P false; Q and R true)
R)) ....... ---,p (P and Q false; R true)
11. (P ....... ---,(Q
13. (---,(P&R)
(P true; Q and R false)
(P V ---,R) (P and R false; Q true) ->
---'Q)
->
---'Q (P, Q and R false)
---'Q) ....... ---,R) (P, Q and R true)
---,R)
->
---,(P&Q) (P true; Q and R false)
Truth Tables for Formulas
Given an interpretation, we can now compute the truth value of a formula on that interpretation. This in itself is interesting only if we have some reason for singling out a particular interpretationsay, because we think it corresponds to reality, given some translation between English and our symbolic language. Nevertheless, it becomes far more interesting when we realize that the validity of a formula amounts to its truth on every interpretation of it. Though no individual interpretation may seem very intriguing, therefore, the set of all possible interpretations merits some attention. A truth table is, in essence, a computation of the truth value of a formula under each of its possible interpretations. If the formula is valid, then the outcome of the truth table technique will be a string of T s. If it's contradictory, the result will be a string of F s. If the formula is contingent, finally, the result will be a string containing both T sand F s. Truth tables amount to nothing more than several simple tables, of the kind we saw in the last section, done at once. But they allow us to evaluate any formula of sentential logic as valid, contradictory or contingent. A truth table for a formula consists of four elements: (1) a listing of the sentence letters of the formula; (2) the formula itself; (3) a list of all possible interpretations of the formula; and (4) a computation of the truth value of the formula on each interpretation. We will write truth tables in this configuration, which is very similar to that we developed for evaluating a formula on a single interpretation: Sentence letters List of Interpretations
Formula Computation
The top of a truth table lists the formula to be evaluated, preceded by the sentence letters it contains. Under the sentence letters, the table lists all possible interpretations of them. If there is one sentence letter, then there are obviously only two interpretations. One assigns truth to the letter, and the other assigns falsehood. If there are two letters, then there are four interpretations; three letters have eight possible interpretations; and so on. In general, n sentence letters can be given 2n interpretations. There is a simple method for listing all these possibilities, and in a standard order: count backwards in base 2 numbers from 2 n - 1 to 0, replacing 0 and 1 with F and T, respectively. Alternatively, imagine an array of the following kind, which is based on the notion that the truth
r
44
Logic, Sets and Functions
value of any sentence letter is independent, logically speaking, of the truth value of any other. The second letter, for instance, could be either true or false, whether the first letter happened to be true or happened to be false. T T F
T T F F
T T F
F T
F F
T F T F T F T F T F T F T F T F
This table suggests that the listings of possible interpretations, for cases involving one, two, three or four sentence letters, should look like this: T F
T T F F
T F T F
T T T F F F F
T F F T T F F
T T F T F T F
T T T T T T T T F F F F F F F F
T T T T F F F
T T F F T T F
T F T F T F T
F
F
F
T T T T F F F F
T T F F T T F F
T F T F T F T F
If the formula contains more than four sentence letters, the table will have 32 or more rowsone for each possible interpretation of the formula. In such cases, the method works, but is so cumbersome that other techniques are probably better suited to the task. To perform the computation, do precisely what we did to compute the value of a formula on a single interpretation. It's often easiest to think vertically rather than horizontally in doing this. That is, we could simply do what we did in the last section 2n times, or we could think things through as we did in the last section, but, where we filled in a value for a subformula there, fill in a column of values for that subformula on all interpretations. The latter strategy is generally more efficient.
Sentences
45
To break this down step-by-step: first, copy the interpretations for each letter under its occurrences in the formula. This gives us values for the bottom nodes of the formula's phrase structure tree. Then, begin searching for subformulas one level up from the bottom of the tree; look, that is, for subformulas as far inside parentheses as possible. Negations of single sentence letters, although they might not be at this level of the tree, are always safe as well, since the value of the negated letter depends on nothing but the value of the letter itself. To systematize this process, we can decide to compute values for negations of single sentence letters first. Then, compute values for subformulas, working from inside parentheses out. This brings us up the phrase structure tree until, finally, we reach the formula as a whole. The last computation should be for the main connective of the entire formula. Truth tables thus follow a "bottom-up" strategy; they begin with the letters at the bottom of the phrase structure tree and work up the tree to compute values for more complex units. Under each sentence letter and connective of the formula, a completed table will have a column of T sand F s. These represent the truth values the formula or subformula has under each interpretation of it. The column for the entire formula itself- the last to be filled in- is the table's final column. It specifies the truth value of the formula on each interpretation.
Definition 2.15 The final column of a truth table is the column under the main connective of the formula at the top of the table. By now, it may be obvious how to use truth tables to evaluate formulas for validity, contradictoriness, etc. A valid formula is true on every interpretation of it; the final column of a table for it, therefore, should contain all T s. A contradictory formula is false on every interpretation of it; the final column of a table for it, therefore, should contain all F s. Satisfiable formulas, which are true on at least one interpretation of them, give rise to tables whose final columns contain at least one T. The final columns of tables for contingent formulas, finally, contain both T sand F s. To summarize: Final Column Formula All Ts Valid
All Fs Contradictory
Ts and Fs Contingent
At least one T Satisfiable
It should now be clear how truth tables function as a decision procedure for the various logical properties of formulas. Constructing a table is mechanical, and tables always have a finite size.
2.8
Examples
To see how this works in practice, let's take an example. Is the formula (P V (Q -+ P)) valid, contradictory, or contingent? It says something like "if Q is true, so is P, unless of course P is true anyway". This doesn't sound like either a logical truth or a contradiction, so a plausible guess would be that it's contingent. We can set up the table by listing the sentence letters in the formula, followed by the formula, and then listing all possible interpretations: P Q (P V (Q -+ P))
T T T F
F T
F F At this point, we copy the truth value assignments under the relevant letters in the formula: P Q (P V (Q -+ P)) T T T T T T F T F T F T F T F F F F F F
,.
46
Logic, Sets and Functions
Now we begin moving up the "grammatical tree" (which linguists call the phrase structure tree, working from negations of single sentence letters (in this case, there aren't any) to other subformulas, from inside parentheses out. In this case, then, we begin with the conditional: P
Q
(P
T T F F
T F T F
T T F F
V
(Q
->
P))
T F T F
T T F T
T T F F
Finally, we conclude by computing the value for the larger disjunction. Note that this gives us the final column; the V is the main connective of the formula, so the truth values we are computing are those of the formula as a whole. P
Q
(P
V
(Q
->
P))
T T F F
T F T F
T T F F
T T F T
T F T F
T T F T
T T F F
The final column contains both T's and F's, so the formula is indeed contingent. Let's look at at another, more complex example. The formula -,((P -> Q) V (Q -> R)) says, roughly, that the conditionals "if P, then Q" and "if Q, then R" are both false. This may sound as if it ought to be contingent also. To find out, we begin a truth table, writing the three sentence letters in the formula, followed by the formula itself, and then listing all possible interpretations, copying these down under the occurrences of the sentence letters in the formula: P
Q
R
T T T T F F F F
T T F F T T F F
T F T F T F T F
-,
((P
->
T T T T F F F F
Q)
V
T T F F T T F F
(Q
->
T T F F T T F F
R))
T F T F T F T F
Now, we begin to compute the values of subformulas, working our way up the phrase structure tree. There are no negations of single sentence letters- the negation here is, in fact, the main connective- so begin as far inside parentheses as possible, with the two conditionals. We can do these in either order, ending up with the table: P
Q R
T T T T F F F F
T T F F T T F F
T F T F T F T F
-,
((P
->
Q)
T T T T F F F F
T T F F T T T T
T T F F T T F F
V
(Q
->
T T F F T T F F
T F T T T F
T T
R)) T F T F T F T F
Computing the truth values of the two conditionals puts us in a position to calculuate the values of the disjunction, which would be immediately above them on a phrase structure tree:
Sentences
P
Q
T T T T F F F F
T T F F T T F F
47
....,
((P
-
-
V R)) (Q T T T T T T F F T F T T T F T F T T T T T T F F T F T T T F T F Notice that the disjunction, surprisingly, is valid; it is true on every interpretation. Finally, then, we use the definition of negation to calculate the values of the entire formula: , ((P P Q R R)) Q) V (Q F T T T T T T T T T T F T T F F T T T T T F F T F T T T F T F T F F F T F F T F T F T F F T F T T T T T F T T F F T F F T T T T F F F T T F F T F F T F T F F T T F F T F F F F Since the formula is false on every row- that is, on every interpretation of it- it is contradictory.
R T F T F T F T F
T T T T F F F F
T T F F T T T T
-
Q) T T F F T T F F
-
Problems Determine whether these formulas are valid, contradictory or contingent. 1. P 2. P&,P
3. PV,P 4. ,(,P - P) 5. P-P 6. p - , p 7. P - (Q -
P)
8. (....,p -
P) ...... P
9. ,p -
(P -- Q)
10. P -
(Q&P)
11. (P--Q)V(Q-P)
12. (P - Q)&,(Q - P) 13. (P - Q) 14. (P V Q) -
(Q - P) (P ...... Q)
15. ((P&,Q) - Q) - ....,p
p"
Logic, Sets and Functions
48
16. (P&((P ~ Q)&(P ~ -,Q))) 17. (P
V
Q)
18. (P&Q) 19. (P
~
<-> -,( -,P&-,Q)
~
~
Q)
20. (P&Q)
(P
~
~
~
Q)(Q Vr)
~
(P&Q)(Q V R)
~
((R ((P
(P V Q)(P&-,(Q V R)) <->
((P&Q) V (P&R))
22. (P
V
<->
((P
23. (P
-->
(Q&R)) (Q
24. (PV (Q 25. (P
2.9
<->
V
~
R))
<->
((P
~
R))
<->
((PVQ)
~
R))
<->
Q) V (P
-,(P
~
~ ~
P)) R))
Q)
Q)&(P&R))
V
(Q
~
<->
21. (P&(Q V R))
P) V (Q
((P
Q)
<->
V
~
Q)
(P
-->
R))
(PV R)) ~
(P
<->
R))
Truth Tables for Argument Forms
Truth tables thus serve readily as decision procedures for the validity, contingency, and contradictoriness of formulas. They also function as decision procedures for argument forms. Recall that an argument form is valid just in case every interpretation making the premise formulas true makes the conclusion formula true as well. Truth tables let us calculate the truth values of formulas under each of their interpretations, so it is easy to set up a truth table to give us the information needed to evaluate an argument form. Actually, there are two ways of going about this. One might set up a table on which one computed the values of each of the premise formulas, and the conclusion formula, separately, and then learn to interpret these tables as evaluating argument forms. An argument form would be valid, of course, just in case no interpretation made the premise formulas all true but the conclusion formula false. Alternatively, one could find a single formula that "expressed" the content of the argument form, and evaluate it for validity. If the formula were set up correctly, then its validity would guarantee the validity of the argument form. In this section we'll present a method that uses the latter strategy. We'll first translate an argument form into a formula- the argument form's associated formula - and then use the technique of the last section to determine whether that formula is valid. If the associated formula is valid, the argument form itself will be valid. If the associated formula isn't, then neither is the argument form. The associated formula of an argument form is a conditional, with the conclusion formula as its consequent and the conjunction of the premise formulas as its antecedent. That is: Definition 2.16 The associated formula of an argument form is the formula that results from (1) conjoining the premise formulas of the argument form, in order, and (2) linking that conjunction to the conclusion formula by means of a conditional. Below are several argument forms, together with their associated formulas: Argument forms (1) (2) (3) P P-->Q (P V Q) ~ -.(Q&R) P-->Q Q<->R Q .. Q :. (R&Q) --> P :. -,R Associated Formulas
49
Sentences 1. (P&(P ---. Q)) ---. Q
2. «P ---. Q)&(Q
<-+
R)) ---. «R&Q) ---. P)
3. [«P V Q) ---. -.(Q&R))&QJ ---. -.R
At this point it's easy to evaluate the argument forms for validity. An argument form is valid if and only if its associated formula is valid. So the associated formulas can be tested for validity, just· as any other formula can be. This results in the tables:
P Q T T
(P
&(P
---.
Q))
---.
T
T
T F
T
F
F F
F F
T T F F
T F T T
T F T F
F T
F F
Q T T
T F T F
T T
(Valid)
P Q R T T T T F F F F
T T F F T T F F
T F T F T F T F
«P T T T T F F
---.
Q)
<-+
R))
F
T
F
T
T T F F T T F F
& T F F F T F F T
(Q
T T F F T
T T F F T T F F
T F F T T F F T
T F T F T F T F
Q) T T F F T T F F
---. F T T T F T T T
-. F T T T F T T T
(Q
T
---. T T T T F T T
T
(R T
& T F F F T F F F
F T F
T F T F
Q) T T F F T T F F
---. T T T T F T T T
P) T T T T F F F F
(Invalid)
P
Q R
T T T T F F F
T T F F T T F
T F T F T F T
F F F
[(P
V
T T T T F F F F
T T T T T T F F
T T F F T T F F
& T F F
R)) T
F T
F T
F T
F
F
F
T
F
F
& F T F F F T F F
QJ ---. -.
R
T T T T T T T T
T F T F T F T F
T T F F T T F F
F T F T F T F T
(Valid) The first and third tables have all T's in their final columns, so they indicate that the associated formulas they are testing are valid. The validity of these formulas, furthermore, demonstrates the validity of the argument forms with which they are associated. The second table has both T's and F's in its final column. The associated formula and its corresponding argument form, therefore, are invalid. When, as in this case, the truth table shows an argument form to be invalid, it also specifies an interpretation making the premise formulas true and the conclusion formula false. The table, in other words, not only indicates that there is such an interpretation, but tells us what it is. If there are several, it specifies them all. To see under what circumstances the premise formulas would be true while the conclusion formula would be false, look at those rows of the table which have an F in the final column. In the example above, there are two such rows. The rows detail interpretations. On the first row having an F in the final column, P is false and Q and R are true; on the second, the last row of the table, all three sentence letters are assigned falsehood. The table thus spells out two interpretations making the premise formulas true and the conclusion formula false.
r
50
Logic, Sets and Functions
On rows having an F in the final column, the associated formula comes out false. But a conditional comes out false only when the antecedent is true and the consequent false. That means that the associated formula is false only when the conclusion formula is false, and the antecedent formula- the conjunction of the premise formulas- is true. But a conjunction is true just in case all the conjuncts are true. So, on any interpretation making the associated formula false, the premise formulas must all be true, and the conclusion formula must be false. That is why the associated formula method works: the formula is valid when and only when the argument form itself is valid. In fact, Aristotle stated all the arguments and argument forms he discussed as conditionals; historically, then, the associated formulas preceded the argument forms with which they are associated. The Stoics were the first to connect the validity of inferences and the necessary truth of the associated conditionals explicitly.
Problems Translate these arguments into our symbolic language, and evaluate the corresponding argument forms as valid or invalid. 1. Sondra mailed the grant proposal I placed on her desk. So she either mailed the grant application or threw it away. 2. You are either a knave or a fool. You are a knave, so you're no fool. 3. If I'm right, then I'm a fool. But if I'm a fool, I'm not right. Therefore, I'm no fool. 4. If I'm right, then I'm a fool. But if I'm a fool, I'm not right. Therefore, I'm not right. 5. Unless I'm mistaken, I'm a fool. But if I am a fool, I must be mistaken. So I'm mistaken. 6. If Einstein's theory of relativity is correct, light bends in the vicinity of the sun. Light does indeed bend in the vicinity of the sun. It follows that Einstein's theory is correct. 7. The thief is in either Tewkesbury or Bristol. He is not in Tewkesbury; therefore he is in Bristol. 8. Aristotle was a brilliant thinker. If his theory of remembering is right, then modern psychological accounts are wrong. So either modern accounts of memory are wrong, or Aristotle wasn't so brilliant after all. 9. Jessica meows just in case she is hungry. She is meowing, but she isn't hungry. Therefore the end of the earth is at hand. 10. If Socrates died, he died either while he was living or while he was dead. But he did not die while living; moreover, he surely did not die while he was already dead. Hence, Socrates did not die. 11. Nothing can be conceived as greater than God. If God existed in our imaginations, but not in reality, then something would be conceivable as greater than God (namely, the same thing, except conceived as existing in reality). Therefore, if God exists in our imaginations, He exists in reality. 12. You believe this only if you're a turkey. You don't believe this, so you are no turkey. 13. If you have a cake, just looking at it will make you hungry; if looking at it makes you hungry, you will eat it. So you can't both have your cake and fail to eat it. 14. A man cannot serve both God and Mammon. But if a man does not serve Mammon, he starves; if he starves, he can't serve God. Therefore a man cannot serve God.
Sentences
51
15. Either we ought to philosophize or we ought not. If we ought, then we ought. If we ought not, then also we ought (to justify this view). Hence in any case we ought to philosophize. (Aristotle) The Stoic logician Chrysippus (280-207 B.C.) regarded these five inference schemata as basic to sentential logic. Which of these are valid in our logic?
16. A
-4
8;A; .. 8.
17. A
-4
8; ---.8;:.
---.A.
18. ---.(A&8); A;:. ---.8.
19. A
V
8;A; .. ---.8.
20. A
V 8; ---.8; ..
A.
21. Cicero, the famous Roman orator, summarized Stoic logic by citing seven principles: the five of Chrysippus above, a repeat of the third, and ---.(A&8); ---.A;:. 8. Is this valid in modern sentential logic? Sextus Empiricus, a sceptic who wrote in the third century A.D., preserves some theorems that the Stoics deduced from their basic principles. Which of these are valid?
22. A
-4
(A - 8); A; :. 13.
f./'+!
/23. "·(A&8)
\.;~~
)4. fA ·••.•,,1
-4
-4
C; ---.C; A;:. ---.8.
8; A - ---.13; :. ---.A.
25. A - A; ---.A
2.10
-4
A; :. A.
Implication, Equivalence and Satisfiability
As we've seen, argument form validity is tantamount to implication. A set of formulas implies a given formula just in case every interpretation making every member of the set true makes the given formula true as well. This is exactly the relation that holds between the premise formulas of an argument form and the conclusion formula. Consequently, the method of associated formulas also works for implication problems. To find out whether a set offormulas {AI, ... ,An } implies a formula 8, simply form the conditional statement ((AI& ...&An ) - 4 8), and test it for validity. The formula will be valid just in case the implication holds. If only two formulas are involved, this test is even simpler: A implies 8 just in case the conditional (A -4 8) is valid, This fact can be summarized in a slogan: implication is the validity of the conditional. Equivalence, of course, is just implication in both directions. It would be easy, then, to use truth tables as a test for equivalence: form two associated formulas, and test both for validity. A formula A is equivalent to a formula 8 just in case both A -4 8 and 8 - A are valid. But there is an easier way. A and A are equivalent just in case they have the same truth value on every interpretation of them. There is a truth function that comes out true just in case the components agree in truth value: the biconditional. So, A is equivalent to A if and only if the biconditional statement A ...... 8 is valid. We can summarize this by saying that equivalence is the validity of the biconditional. To show that P is equivalent to ---.---.P, for example, we can test the biconditional P ...... ---.---.P for validity:
r'
52
Logic, Sets and Functions
P +-7 P P T T T F T T F T F T F F The final column of this table contains all T's, so the formula is valid. That means that P and " p agree in truth value on every interpretation of them, so they are equivalent. Finally, truth tables also serve as a decision procedure for satisfiability. We have already seen this with respect to individual formulas. Even when we are concerned with the satisfiability of sets of formulas, however, truth tables offer a simple test. A set of formulas is satisfiable if and only if its members are all true on some interpretation of them. Of course, if every member of a set is true on a given interpretation, the conjunction of the members must be true on that interpretation as well. So a set is satisfiable just in case the conjunction of the members of the set has a true interpretation. If the conjunction is contradictory, the set must be contradictory; if the conjunction is satisfiable, so is the set. This, too, may be put into a slogan: satisfiability (of a set) is the satisfiability of the conjunction. Consider, for instance, the set of formulas {,Q, P V Q, P --. Q}. Is this set satisfiable? To find out, we can form the conjunction of the members of the set, and test it for satisfiability: P Q (-. Q & (P V Q)) & (P --. Q) TT FTFTT T FTTT TF TFTTT FFTFF FT FTFFT T FFTT FF TFFFF F FFT F The final column here consists of all F's, so the formula is contradictory- false on every interpretation of it. But the conjunction would be true if every member of the set of formulas were true. So this indicates that there is no interpretation making every member of the set true. The contradictoriness of the conjunction implies the contradictoriness of the set of formulas. The truth table technique, therefore, is a simple but remarkably powerful tool for solving problems in sentential logic. It serves as a decision procedure for argument form validity, formula validity, implication, equivalence, formula satisfiability and set satisfiability. It does all this, furthermore, in a clear and easily understandable way. A truth table, in effect, surveys all possible interpretations of a formula, argument form, etc., and computes a truth value for it on each interpretation. This clarity, of course, comes at a price. When the formula or argument form becomes very complex, the number of possible interpretations may be very large. In the next chapter, therefore, we'll turn toward a method that verifies the validity of formulas and argument forms, without running through every possible way of assigning truth values to sentence letters. Problems Determine whether the formulas in each pair are equivalent. If they are not, say whether either formula implies the other. 1. P --. Q and P
+-7
(P&Q)
2. -.(P V Q) and -.P V -.Q 3. -.(P&Q) and -.P&-.Q
4. -.(P --. Q) and -.P --. --.Q 5. -.(P
+-7
Q) and -.P +-7 -.Q
6. -.(P
+-7
Q) and -.P
+-7
7. -.(P --. Q) and P&--.Q
Q
Sentences
53
8. -.(P&Q) and -.P V -.Q 9. -.(P
Q) and -.P&-.Q
V
10. P&Q and (P V Q)&(P
<-+
Q)
11. P&Q and (P V Q)&(P
->
Q)
12. P&Q and (P
->
P)
Q)&(Q
V
13. P V Q and -.P
->
Q
14. P V Q and -.Q
->
P
15. P
->
Q and P
<-+
(P
16. P
->
Q and Q
<-+
(P V Q)
17. P
<-+
Q and (P
18. P
<-+
Q and (P&Q)&(-,P&-.Q)
19. P
<-+
Q and -,p
20. P
<-+
Q and (P
Q)&(Q
->
->
P)
-.Q
<-+ <-+
Q)
V
Q)
<-+
P
Evaluate these sets of formulas as satisfiable or unsatisfiable. For each satisfiable set, specify an interpretation making every member of the set true. 21. {P, (P V Q)
->
Q, -.(P&Q)}
22. {P
->
(Q
->
R), -.((P&-.R)
23. {P
->
(Q
V
24. {P
->
Q, P
25. {Q
<-+
(P
R), P
(Q V -.R)), -.(P
26. {P, Q
->
27. {P
-.Q, (R&P)
->
->
-.R}
-.Q}
->
->
-'Q, P
->
-.Q)}
->
->
(Q V R)), R}
-.R, R} Q, (R&-.P)
->
P}
->
28. {-,(P-> (Q&-.R)),PV(Q&-.R),R-> (PVQ)} 29. {(P
-.Q)
V
30. {-.(P
<-+
<-+
(P
-.Q), P
<-+
31. {P, (Q
V
32. {Q
(PV R),P
->
33. {(R&Q)
R)
->
R), P
->
(P
(r
V ->
<-+
<-+
(Q&(R
Q), Q
-,Q), P
<-+
R)}
(P
<-+
(Q
<-+
-.R)}
->
(Q&P))
<-+
(Q
V
r), (P
34. {P V (Q V (-.P&R)), -.((P&-.Q)
->
(r
V
35. {((P
->
->
Q)
->
P)
->
V
Q)
-.P)), Q}
->
(R&-.Q),R
-.(Q&P), (P
->
R, ((R
->
P)
->
R)
<-+
R)
->
(Q
<-+
P)), -.(P
<-+
-.R)}
Q, ((Q
->
R)
->
->
-.P)}
Q)
->
P}
~
Logic, Sets and Functions
54
Consider the following formulas:P V Q, P&Q, P ---; Q, p ....... Q, -,P, and -'Q. Where A and Bare different formulas from among this collection, there are 30 possible statements of the form A implies B. Of these, only five are true. Show that these are the five.
36. P&Q implies P
V
Q
37. P&Q implies P ---; Q 38. P&Q impliesP
f-+
Q
39. p ....... Q implies P ---; Q
40. -,p implies P ---; Q Chrysippus held as a basic valid argument form PVQ; P;:. -'Q, which has the associated formula ((P V Q)&P) ---; -'Q. Subsequent logicians advanced a number of principles which we would today consider invalid. Do any of these follow from the formula associated with Chrysippus's argument form? 41. ((P ---; -,Q)&P) ---; -,(P ---; Q) (Stoics)
42. ((-,P ---; Q)&-,Q) ---; -,(P ---; Q) (Stoics) 43. (-,(P&Q)&-,P) ---; Q (Cicero)
44. (P ---; -'Q)
f-+
-,(P ---; Q) (Boethius)
45. ((P ---; (Q ---; R))&(Q ---; -,R)) ---; -,p (Boethius)
Chapter 3
Natural Deduction The method of truth tables can evaluate arguments in a theoretically clear and practically efficient way. But it provides little insight into how people construct arguments and, especially, extended arguments and proofs. In this chapter we'll develop a system designed to simulate people's construction of arguments. Although it forces arguments into rather rigid structures, it is natural in the sense that it comes close, in certain respects, to the way people argue, particularly in legal, scientific and philosophical contexts. It comes even closer to the way mathematicians prove theorems.
3.1
Natural Deduction Systems
A natural deduction system is a collection of rules of inference. The central notion of a natural deduction system is that of proof. Some proofs, called hypothetical proofs, begin with assumptions, or hypotheses. The assumptions serve as premises for the argument found in the proof; the conclusion depends on these assumptions. Such proofs show that the conclusion is true, not outright, but if the assumptions are true. Other proofs, however, use no assumptions; they show that their conclusions are true outright. A proof in a natural deduction system is a sequence of lines; on each line is a formula. Each formula in a proof must either be an assumption or derive from formulas on previously established lines by a rule of inference. In our formulation, the formula on the topmost Show line of a proof is its conclusion; the proof is a proof of that formula from the assumptions. Formulas proved from no assumptions are theorems. A theorem of a natural deduction system, then, is any formula that can be proved from no hypotheses in the system. Rules of inference are either simple or complex. Simple rules allow us to write down formulas having certain shapes in a proof if other formulas of certain kinds are on already-established lines in that proof. For example, we'll have a rule that lets us write A or B if we've already established A&B. Complex rules, in contrast, allow us to write down a formula of a certain shape in a proof if some other proof is completed. We'll have a rule, for instance, that allows us to write -.A if we've proved a contradiction from the assumption A. Another complex rule allows us to assert a conditional formula A --> B if we can, in a subordinate proof, assume A and derive B. Because our system has some complex rules, proofs will sometimes appear within other proofs. A proof appearing within another is subordinate to it. The larger, superordinate proof will make use of the information in the subordinate proof by way of a complex rule. Sometimes the larger proof will simply take over the conclusion of its smaller partner. In other
55
~
Logic, Sets and Functions
56
circumstances, the larger proof will use the fact that the subordinate was able to prove a conclusion from a given assumption to state an assertion that doesn't itself depend on this assumption. An assertion within a proof established by a subordinate proof is called a lemma. As our talk of "shapes" suggests, natural deduction systems are purely syntactic. It's possible to verify that a proof is successful without any reference to the meanings of the symbols in the proof. Nevertheless, we use the rules of inference we do because, taken together, they allow us to prove from a set of assumptions only those formulas that follow from the set. The rules themselves are syntactic, based on the shapes of formulas, but their justification is semantic. For the most part, the deduction system of this book has two rules for each connective. One rule tells us how to prove a formula with that connective as main connective. In short, it tells us how to introduce formulas of that form into proofs. The other rule tells us how to use the information encoded in a formula having that connective as main connective. That is, it tells us how to exploit formulas of that kind in proofs. For this reason, the basic rules of the system will largely fall into two groups: introduction rules and exploitation (or elimination) rules. Most connectives will have one rule of each sort. Before discussing these rules in detail, we need to know more about the kinds of proofs that the system allows. Proofs will be able to appear inside other proofs. The subordinate proofs will fulfill various functions. In this chapter, there will be three such functions. Accordingly, there will be three methods of proof. Two of them are really complex connective-introduction rules. The remaining, "pure" method is that of direct proof. All proofs are sequences of lines which are structured in certain ways. As we'll write them, proofs will have three columns. The middle column will consist of a sequence of formulas; some may be preceded by the word Show. The left column will number these formulas. The right column will provide justifications for the formulas. Thus, if a formula on a given line derives from previously established formulas by a rule of inference, the right column will say what rule of inference and what earlier lines were used. If the formula is an assumption, the right column will say so. Only when the formula derives from an entire subordinate proof will the right column be empty. A direct proof begins with premises- or, if it appears within another proof, formulas deduced from earlier lines- and proceeds to its conclusion. We'll begin a direct proof, once any premises or earlier lines are recorded, by stating what we want to show: n. Show
A
Lines that we hope to establish by doing proofs- that is, theorems and lemmas- will always have this form. The left column contains a line number; the right column is empty. The formula we hope to prove is prefaced by the word Show to indicate that we haven't proved it yet; so far, the information recorded on the line is just wishful thinking. How can we make our wishes reality? Clearly, by proving A. If we can prove A from what we are given, then we can go back to line n and cross out the Show that signalled that the following formula was, at that point, only fantasy. The proof allowing us to do this will provide our explanation for cancelling the Show. It will follow line n immediately. To show graphically what lines constitute the proof, we'll draw a bracket encompassing those lines to the left of the formulas in the proof. A successful, completed direct proof will therefore look like this:
Direct Proof n.
n+m
Shew....,A
[~
57
Natural Deduction
Obtaining A allows us to complete the direct proof only if two conditions are satisfied. First, A, on line n + m, must not already be enclosed in another set of brackets. Second, there must be no uncancelled Show statements in the area to be bracketed. We want to show that A follows from what has gone before. If A, on line n + m, were already enclosed in a bracket, or if we would be enclosing any uncancelled Show statements by drawing a new bracket to complete the proof, we could not be sure that A followed from what we were given. We would know only that A followed from what we were given together with some additional assumptions, for other proof methods introduce assumptions into the proof. Neither of the following, then, count as legitimate instances of direct proof: n. Sh8WP WRONG n. Sh8WP WRONG
;[r:
Q
: [:h~Q
On the left, we are trying to use the occurrence of 'P' on line k to complete the direct proof of 'P'. But 'P', on that line, is already enclosed in another set of brackets. Because the proof of 'Q' may have begun with an added assumption, we have no guarantee that 'P' follows from the formulas above line n alone. On the right, there is an uncancelled Show statement within the bracket. This too may have let in added assumptions. Since the bracketed lines constitute a proof that provides the justification for line n, no other justification is needed. The right column for line n can thus remain empty.
3.2
Rules for Negation and Conjunction
In this section we'll develop the rules of inference for negation and conjunction. First, however, we'll introduce a simple rule that allows us to construct hypothetical proofs. The assumption rule says that you may begin a proof by listing premises or assumptions. That is, before the very first Show line, you may write premises. The conclusion you derive will depend on these premises. Thus: AssumptionA
n.A A Here line n must precede the first Show line in the proof. Suppose that we want to prove 'Q' hypothetically from 'p ....... Q' and 'P V Q'. We would begin the proof by writing the premises, and then the Show line containing the conclusion:
l.P ....... QA
2. PvQ A 3. Show Q The rule of conjunction exploitation asks what use we can make of the information encoded in a conjunction. If a conjunction is true, in other words, what follows? Clearly, the truth of both conjuncts. If A&B is true, then both A and B must be true. The rule of conjunction exploitation thus takes two forms, since the truth of a conjunction implies the truth of both conjuncts:
r
Logic, Sets and Functions
58 Conjunction Exploitation( &E)
n.A&B n+m.A
&E,n
Conjunction Exploitation( &E)
n.A&B n+m.B
&E,n
Hereafter, we'll abbreviate rules having two forms by using parentheses. We can write &E as Conjunction Exploitation( &E)
n.A&B n+m.A(orB)
&E,n
This rule, often called Simplification, tells us that, from a conjunction, we can prove either or both conjuncts. The conjuncts may be written on any later line. When this rule is applied, we write a conjunct together with the explanation that the line comes by application of & E- conjunction exploitation- to the formula on line n. (The line just above the formula that results from applying this rule serves to separate the formula deduced from what must be present earlier; it won't appear in actual proofs.) To apply this rule to a formula, the conjunction must be its main connective. We can move from 'P&(Q ---. R)' to 'Q ---. R', but we cannot go from '(P&q) ---. R' to 'P ---. R'. Let's do a simple proof using this rule. Let's assume '((P&Q)&---,R) , and prove 'Q'. We begin with an assumption and a Show statement.
1. «P&Q)&-,R) 2. Show Q
A
Now, we can exploit the conjunction to derive the smaller conjunction '(P&Q)'. Note that we can't leap inside to derive 'Q' directly: the connective to which we apply the rule must always be the main connective of the formula.
I. «P&Q)&-,R) 2. Show Q 3. (P&Q)
A
&E,I
The right column tells us that the formula on line 3 comes from line 1 by applying conjunction exploitation. At this stage, we can easily apply that same rule again to obtain 'Q', which is what we want in order to finish the proof:
1. «P&Q)&-,R) 2. ~ Q 3. [(P&Q) 4. Q
A
&E,l &E,3
The rule of conjunction introduction tells us how to prove conjunctions. It's a very simple rule, since there is an extremely obvious strategy for proving a conjunction: prove each conjunct. If you
•
Natural Deduction
59
know that A is true, and that B is true, you can conclude that A&B is true as well. The rule thus states that from the two formulas A and B you can'derive A&B: Conjunction Introduction( &1)
n.A m.B p.A&B &I,n,m
The right column simply indicates that A&B comes from applying conjunction introduction to the formulas on lines nand m. The order in which A and B appear in the proof makes no difference. We could just as easily have concluded B&A. To see how this might be used in a proof, let's show that '(P&Q)' allows us to derive '(Q&P)'. Again we begin by using the assumption rule and writing a Show line:
1. (P&Q) A 2. Show (Q&P) Now, we need to show 'Q&P'. To do this, we need to separate out the two conjuncts. We can therefore derive them separately, using conjunction exploitation, and put them back together in the other order, using conjunction introduction:
1. (P&Q)
2.
~
3. [PQ
A
(Q&P) &E,1 &E,1 &1,4,3
4.
5.
(Q&P)
One negation rule serves as both an introduction and an exploitation rule. Because it always introduces or exploits two negation symbols at once, we'll refer to it as '-,-, '. It says, basically, that two consecutive negation signs "cancel each other out": -,-,A is equivalent to A. First formulated by the Stoics, this rule is often called Double Negation. One can add two consecutive negation signs, or delete them, without affecting truth values. Negation Introduction( -,-,)
n.A n + p.-,-,A
-,..." n
Negation Exploitation(-,-,
n.-,-,A n+p.A '-',n
We can express this in a more compact form by writing a double line between A and -,-,A. A double line indicates that the rule is invertible - that one can go from what is above the lines to what is below them, or vice versa. The rule works in both directions. So:
,.
60
Logic, Sets and Functions
Double Negation(-'-')
n.A
-'-',m
m.---,-,A
-'-',n
To illustrate the use of this rule, let's show that we can derive '-,-,P&-,Q' from '-,-,(P&-,-,-,Q)'.
1. ...,...,(P&""...,...,Q) 2. ~ (...,...,P&...,Q) 3. (P&...,...,...,Q) 4. P 5. ...,...,p 6. -,-,-,Q 7. ...,Q 8. (...,...,P&""Q)
r
A
"""",1
&E,3 ...,...,,4 &E,3
"""",6 &1,5,7
Another, more powerful negation rule is indirect proof, which is essentially another negation introduction rule. The indirect proof rule is complex. It says that we can write -,A if we can derive a contradiction from the assumption that A. Indirect proofs always introduce assumptions, called assumptions for indirect proof (AlPs), from which they try to prove contradictions. What point is there to that? If an assumption leads to a contradiction, it must be false. Consequently, indirect proofs always establish the negations of their assumptions. An indirect proof begins with a statement of what we want to prove: a formula prefaced by the word Show. This formula will have a negation as a main connective. The proof will then make an assumption. The assumption will be the same as the formula we're trying to establish, but with the main negation omitted. An indirect proof will thus begin: n. Show-,A n
+ 1. A
AlP
To complete the proof, it's necessary to prove a contradiction. The contradiction doesn't have to relate directly to the assumption; we do not, in other words, have to use the assumption A to prove A&-,A, or anything else containing A. Furthermore, we don't have to get our contradiction into a single formula; two formulas, one of which is the negation of the other, suffice. What we want, then, is to prove a formula /3 and also its negation -,/3. A completed indirect proof, then, looks like this:
Indirect Proof n.
Shew-,A
n+1.
[:
AlP n
+P
n+q
/3 -,/3
It makes no difference which of /3 and -,/3 is proved first. Notice that, once the proof is complete, we cancel the word Show to indicate that we've establish what, earlier, we had merely hoped for. Again, neither /3 nor -,/3 may already be enclosed in brackets on lines n + p and n + q, and we may enclose no uncancelled Show statements when we draw a bracket to complete the proof. To take an example, this rule allows us to prove the thesis that a contradiction implies anything. To see this, take a contradiction such as 'P&-,P', and a completely unrelated formula 'Q'. We can show that, from 'P&-,P', we can derive 'Q'. SO we begin with an assumption and a Show line:
61
Natural Deduction
1. (P&..,P) A 2. Show Q Now, how can we get to 'Q' ? There seems to be no way to get it by a direct proof, since it bears no relation to our assumption. Our only choice, at this point, is to use indirect proof. This bit of reasoning is not restricted to this example. Throughout this text we'll resort to indirect proof when nothing else suggests itself as a good proof technique. Of course, as we have stated the rule of indirect proof, we can prove only negated formulas. But our rule for negation tells us that 'Q' and '-,-,Q' are equivalent. Instead of proving 'Q', then, we can prove '-,-,Q', and then use the rule for double negation to obtain 'Q'. This strategy leads to the proof:
1. (P&..,P) 2. ~ Q ..,..,Q 3. 4. 5. 6. 7.
[[f
A AlP &E,1 &E,1 "'''',3
Notice that, in this proof, we used a line that contained the word ShoUT, that's how we got from ',-,Q' to 'Q'. Yet we can't use such lines all the time. If we could use Show lines anytime we wanted, we could easily prove anything. Consider, for instance, this "proof" that pigs fiy: 1.
ShewPigs fiy
2. 3.
["PigS fly Pigs fly
'-',1
Error
",2
Clearly we've got to restrict the circumstances in which we can use such lines. We'll say that any line we can use in a proof at a given point is free at that point. In sentential logic, every line is free, except (1) lines that begin with an uncancelled Show; and (2) lines that are imprisoned within a bracket. Lines beginning with an uncancelled Show contain formulas that we haven't yet proved. All we can say is that we hope to prove them. So the information on those lines is inaccessible to us in the proof. That's why the proof that pigs fly seems silly; we used what we wanted to prove in order to prove it. Once a Show has been cancelled, on the other hand, we've proved the formula, so we can use it throughout the rest of the proof. It's only a little harder to see the point of the second restriction. Lines that are enclosed in a completed bracket or "prison" may depend on a particular assumption. Both indirect and conditional proofs introduce assumptions; the assumptions themselves, and the following lines that depend on them, are true only given those assumptions. We can't assume that they are true in general. Above, for example, we introduced a contradiction- 'P&,P'- as an assumption. We proceeded to show that, if this were true, then anything would be true. Here, and in subordinate proofs, we wouldn't want to say outright that the contradiction was true; we would merely pretend that it was true. The contradiction, in other words, would serve only as a dialectical premise. If this proof were part of a larger proof, it would be a terrible mistake to come back and claim that, in general, we could prove 'P&,P' because it was introduced earlier, within a subordinate proof, as an assumption. The only free lines, then, are those that are neither prefaced with an uncancelled Show nor imprisoned within a bracket (i.e., a completed proof). These "proofs", consequently, make grave errors:
,.
62
Logic, Sets and Functions
1.
2. 3.
4. 5. 1.
Shew..P
[Shew«(P&--,P) --. P) [ (P& ..P) P
ACP
Error
&E,3 &E,3
--,p Shew(P&--,P)
2. [P 3...P 4. (P& ..P)
&E,l &E,l &1,2,3
Error Error
In summary, formulas preceded by an uncancelled Show have not yet reached legal age, and those within a completed prison are sentenced to life. Both kinds of formulas are inaccessible. We've already seen one rule in this section, the assumption rule, that is structural in that it makes no reference to any particular connectives. We'll now introduce another structural rule, which says that whatever is true, is true. It allows you to repeat yourself. The Reiteration rule allows us to repeat a formula that appeared earlier in the proof, so long as the line containing that formula is still free.
ReiterationR n ..A
n + p ..A
R,n
This rule supposes, in effect, that whatever we've shown to be true is still true. Repeating what we've established can never get us into trouble. To see how this rule works, consider a simple argument: It's not true that Congress will cut military spending and refuse to raise taxes, because Congress will raise taxes. We can symbolize this as
P :. -,(Q& ..P) We begin a proof to demonstrate the validity of the argument form by using the rule of assumption to introduce the premise, and then trying to show the conclusion.
1. P 2. Show
A ""(Q&""P)
The first rule of thumb for constructing proofs is this: choose a proof method by looking at the main connective of the formula being proved. In trying to prove a negation, use indirect proof.
1. P A 2. Show ...,(Q&...,P) 3. (Q&...,P) AlP We can apply conjunction exploitation to obtain '.. P', and then get a contradiction by reiterating the premise:
Natural Deduction
63
1. P A 2. ~ ..,(Q&..,P) AlP 3. [(Q&..,P) 4. ..,p &E,3
5.
P
R,l
In general, reiteration is useful chiefly for completing indirect and other subordinate proofs.
Problems Using deduction, show that each individual formula is provable, and that the conclusion of each argument form is provable from the premises. 1. P&Q;:. Q
2. P&(Q&R);:. Q 3. •(P&.P) 4. •(P&(Q&.P)) 5. •((P&.Q)&(Q&.P)) 6.. • P&.Q;:.•(P&Q) 7. (P&Q)&R; .. P&(Q&R)
8. P&.Q; .R&S; :. S&P 9. •(P&Q); Q;:.• p 10. Q;:.•(.P&.Q)
_."" /l1.P;:.• (Q&.(P&Q))
'\\~:\;"
.
12 . • ((P&.Q)&R);:.•((P&R)&.Q)
13 . •(P&.Q); P; .. Q 14. •(.P&Q); .(P&Q);:. • Q 15. P&Q; .(Q&.R); .(P&.S);:. R&S 16. •(.( .(P&.Q)&---.P)&---.P)
17. P&Q; ---.(---.R&Q); ---'(r&---.S);:. P&S 18.
* ---.(P&.Q); .(Q&---.P); ---.(P&Q); ---.(.P&---.Q);:.
19.
* ---.(---.S&Q); ---.(P&(---.Q&.R)); .(R&---.S);:. ---.(.S&P)
20.
* .(P&R); .(---.(P&Q)&.P);:. R
--. .S
R
'"
64
3.3
Logic, Sets and Functions
Rules for the Conditional and Biconditional
The rule of conditional exploitation is very simple. Many axiomatic systems, in fact, have only it as a rule of inference. Often called modus ponens, this rule sanctions the inference from 'p' and 'P - t Q' to 'Q'. It thus stands behind arguments such as 'If you're smart, you'll do well at logic; you're smart; so you'll do well at logic.' Conditional Exploitation
n.A - t B m.A p.B
-t
-t
E
E,n,m
To illustrate this rule, let's show that we can derive 'P&Q' from the hypotheses 'R 'R&P'.
1. (R-Q) 2. (R&P)
~: [~ 5. 6. 7.
P
Q (P&Q)
-t
Q' and
A A
(P&Q) &E,2 &E,2 -E,1,4 &1,5,6
Conditional introduction is a complex rule, which constitutes the method of conditional proof. A conditional proof, like an indirect proof, may use some premises or earlier lines of a proof, but it doesn't have to. It always establishes a conditional formula, that is, a formula with a conditional as its main connective. It begins with a statement of what we want to prove, and proceeds to make an assumption, called the assumption for conditional proof (ACP). This assumption is always the antecedent of the conditional we're trying to establish. A conditional proof, then, begins as follows:
n.
ShowA
n+l
A
-t
B ACP
The method mimics actual arguments for conditional statements in English. To argue for a conclusion such as If the fetus is a person, it has a right to life, we can begin by supposing that the fetus is a person, and seeing what follows. If we can show that it follows from the assumption that the fetus has a right to life, then we have established the original conditional. Summarizing: to show that if A, then B, assume A and try to show that B follows. A successful conditional proof has this form:
Conditional Proof n.
ShewA
n+l.
[:
n+p
B
-t
B ACP
Natural Deduction
65
Once again, obtaining the subordinate conclusion 13 allows us to cancel the Show above it and count the conditional statement as established. A proof proves it,s topmost Show line whenever the Show is cancelled. At that point, the conditional proof is complete. Drawing a bracket around the proof indicates that it is complete, and also allows us to see easily what lines constitute the justification for the conditional statement. Just as in direct and indirect proofs, we may not draw the bracket if 13, on line n + p, appears inside another bracket, or if we would be enclosing an uncancelled Show statement. To take an example, we can use conditional proof to show that 'R ~ Q' is derivable from 'P&Q'.
1. (P&Q) A 2. ~ (R--Q) ACP 3. &E,1 4. Q
[R
Another example points out that '(P
1. 2. 3. 4. 5. 6. 7. 8.
---->
""'Q)
--+
(Q
---->
....,P)' is provable:
~
((P---.Q)--+(Q--..,P)) (P--..,Q) ~ (Q--"'P)
ACP
Q
ACP ..,p
~
AlP -- E,2,6
P -.Q
Q
R,4
Here, an indirect proof is subordinate to a conditional proof that is, in turn, subordinate to another conditional proof. The rules for the biconditional are straightforward. First, consider biconditional introduction. Under what circumstances may one introduce a biconditional into a proof? What does one have to do, that is, to establish the truth of a biconditional? Recall that a biconditional is so-called because it amounts to two conditionals. This is crucial in devising a proof strategy. Mathematicians, for example, tend to prove biconditionals in two steps. They do the "left-to-right direction" and the "right-to-left direction" separately. In other words, they prove two conditionals in order to establish the biconditional. Our rule for biconditional introduction similarly requires two conditionals: Biconditional Introduction
n.A ----> 13 m.13 ~ A p.A ~ 13
~
~
I
I,n,m
The rule for exploiting biconditionals rests on the fact that a biconditional asserts that two sentences have the same truth value. If we know the truth of a biconditional, and also the truth of one of its components, we can deduce the truth of the other. Biconditional Exploitation n.A~B
m.A(orB) p.13(orA)
~
E, n, m
~
E
"
Logic, Sets and Functions
66
Notice that this rule differs from conditional exploitation by allowing us to deduce A from B or vice versa. If we have one component, we can derive the other; it makes no difference which appears on which side of the biconditional. The rule of conditional exploitation, however, works only in one direction. If we have the antecedEmt of the conditional, we can obtain the consequent. But we can't go from the consequent to the antecedent. To show how these rules work, let's show that 'P ....., Q' is derivable from '(P ~ Q)&(Q ~ P)', and vice versa.
1. (p.... Q) 2. ~ ((P-Q)&(Q-'P» 3. ~ (P-'Q) 4. 5. Q 6. ~ (Q-P) 7. 8. P 9. ((p--.Q)&(Q--'P»
[P
A
ACP -E,l,4
[Q
1. ((p--.Q)&(Q--.P» 2. ~ (P-Q) 3. [(P--'Q) 4. (Q--'P) 5. (p.... Q)
ACP -E,1,7 &1,3,6
A
&E,l &E,l -1,3,4
Problems Theophrastus (371-286), a pupil of Aristotle, cited these principles as hypothetical syllogisms. Use natural deduction to demonstrate their validity. 1. P
~
Q; Q
~
R; :. P
~
R.
;:i>~"';;
t.~>p ~
Q; Q ~ R; :. --.R
~
--.p.
3. P
~
Q; --.P
~
R; :. --.Q
~
4. P
~
Q; ,p
~
R;:. ,R ~ Q.
5. P
~
R; Q ~ -.R;:. P
~
R.
--.Q.
Show, using deduction, that these argument forms are valid. ~
6. P
f-t
(Q&P); .. P
7. P
->
Q; ,(P -. R);:. ,(Q
8. P
->
Q; P
9. P
f-t
Q; P ...... R; :. Q ...... R
10. P
->
Q; 'Q;:. -.P
->
R; :. P
11. p....., Q; ,P; .. -.Q 12. (P
->
Q)
~
P; .. P
Q
~
->
R)
(Q&R)
67
Natural Deduction
;
13.)P
Q; :. P ~ (P&Q)
-+
I:,.....>., ...• •
14. P&-.Q; R
(R&Q);....,R
-+
15. S
-+
(R&P);Q
-+
16. P
-+
(Q&(R
....,P)); Q
17. (P
-+
Q)
-+
-+
(R
-+
(-.R&....,P1 ); .. (Q&S)
* (S&-.R)
19.
*P
20.
* (P +-> Q) +-> (P +->
3.4
~
...,P; (Q
Q; :. -.P
+->
~
-+
Pz
-+
....,R;:. ....,p
S); ....,(P&Q)
18.
->
S; .. S
-+
-.S)
~
-+
S; :. ...,s
-+
-.R
~
-.P;:. P
~
(R&(S&Q))
-.Q
R); -.((-.P&-.Q)&-.R);:. P
+->
(Q
~
R)
Rules for Disjunction
Like most connectives, disjunction has both an introduction and an exploitation rule. The introduction rule is very simple. It says that one may introduce a disjunction into a proof if one has already obtained either disjunct. Disjunction Introduction(V I)
n.A(orB) n + p.A V B
V I,
n
To see how this rule works in practice, let's try to prove "the law of the excluded middle"Ip V -.P'. We can't prove either disjunct separately- neither Ip' nor '-.P' are valid- so we need
to use indirect proof. This means we need to prove '-.-.(P V -.P)'. Furthermore, introducing the assumption for indirect proof leaves us with very little to work with. We should be able to prove a contradiction, but it's not obvious how we can get anything out of I-.(P V ....,P)'. So, we can begin by trying to show '-.P'. (After all, from a contradiction, anything should follow.)
1. 2. 3. 4.
~ ~
(Pv-'P) -'-'(Pv-'P) -,(Pv-'P) AlP
~...,p
5. [P 6. 7. 8. 9.
(Pv-'P) -'(Pv-'P) (Pv-'P) (Pv-'P)
AlP vI,5 R,3 vl,4 -'-',2
This shows how useful disjunction introduction is, even in proving a fairly simple theorem. Disjunction exploitation is perhaps the most complicated rule in our entire proof system. How can we exploit the information encoded in a disjunction? That is, if we know a disjunction, how can we use it to obtain some conclusion? This is what a mathematician encounters in doing a "proof by cases". Often, one can say only that there are several possibilities, and then examine each individually. If the conclusion one is seeking holds in every case, then one can conclude, since those were the only possibilities, that the conclusion holds in general. So, if we have a disjunction, and
68
Logic, Sets and Functions
ways of getting from each disjunct to a conclusion, then we can obtain that conclusion. This leads us to the rule often called constructive dilemma: Disjunction Exploitation(V E)
n.A V 13 m.A -+ C p.13
-+
C
q.C
VE,n,m,p
Faced with a disjunction, then, we must usually prove our conclusion in each of the two cases the disjunction presents. We've got to "get down to cases." To see how disjunction exploitation works, let's derive 'Q' from 'P V Q' and 'P -+ Q'.
I. (PvQ) 2. (P-Q)
A A
3.~
Q (Q-Q)
4. 5. ~~ [Q 6. Q
ACP
vE,I,2,4
Note that we can prove 'Q same.
Q' in just one step, since the antecedent and consequent are the
-+
Problems Show that the conclusion formulas of these argument forms can be proved from the premise formulas. 1. P V Q; -.P;:. Q
2. P
V
Q; P
3. (P&Q)
R; Q
->
-+
S; :. R
(-,P&-.Q); :. P
V
4. -.P V -.Q; .. Q
->
V
~
S
Q
-.P
5. -.P V -,R; :. -,(P&R) 6. P
V
Q; -.P
V
-,Q;:. P
~
-'Q
7. (PVQ)VR;:.PV(QVR) /-f';.-
8. P&(Q
V
R); :. (P&Q)
V
9. (R&-.P) V (Q&R); :. (P
(P&R) -+
Q)&R
10. P V Q; R V S; -,(P V S); :. (Q&R) V T 11. P; -.S V -.P; P
12. P V Q; P
->
-+
R; :. -.S&R
R; -.R; .. Q
13. -.P V Q; -'Q V R; -.R; :. -.P
14. (P&...,Q)
->
-,R; R; :. ...,p V (P&Q)
69
Natural Deduction 15. -.5 V (5&P); (5 16. P
~
P)
~
R; :. R
-.Q; -,p V R; Q; (Q&R)
->
17. -.(P&-.Q) V -.P; :. P
P; ., -,R
~
Q
->
18. P&5; P ~ (-.5 V R); :. R ~
19. -.P V Q; -'Q; -.P
P; :. -.R V Q
20. P
~
Q; R
21. R
->
P; -,R
22. P
->
-.Q; R; R
23. P
~
(-.R
24. P
->
-.Q; -,p
->
V
->
~
Q; Q ->
S); P
~
25. P&-.5; R
R; :. R
(Q V -.S); :. S
~
-,S; :. P
-.R
->
->' -.R;
S; P
S; j :. P V S
R
~
-.S; :. -.Q
V
(Q
->
-.P
V
V
-.S
R); :. Q
26.
* P&Q;
R&-.S; Q ~ (P ~ Pd; PI ~ (R ~ (S V Ql)); :. Ql
27.
* P&Q;
P
28. 29.
* P V (Q V 5); SI&-,S2; -.(-.51 V S2) ~ -.P; (S ~ R)&-.R; :. Q * P&(-.Q&-.pl ); P ~ (S ~ R); S ~ (R ...... (PI V Q)); .. -.5
30.
*5
->
31.
*P
V (R V Q);
32.
*P
->
33.
*
34.
* (P V Q)&R;
~
(S V R); -'(R&Q); :. S
P; (S&P)
(r
~ ->
Q; R
V
(Q&R); :. -.Q Q
~
R V S; .. Q V SI
5t}&(Q ~ S2); (51 V S2) ~ (P V Q); -.P; .. Q
(S&R); (R V -.S)
(P&Q)
~ 51;
~ ->
S; :. -.P
(Q&Q1); Q1 ...... Q2; ., P
~
Q2
5 ~
(R
~
S)
35. *-.P V (Q&R); (R V -.Q) ~ (5&Qd; (Q1&Q2) V -.(Q1 V Q2); :. -.P V Q2 36.
3.5
* -.(-.P&Q)&(P ......
-.Q); :. P ...... (Q
~
R)
Derivable Rules
The series of proof methods and rules we've adopted allow us to prove the validity of any valid argument form in sentential logic. That is to say, the system is complete - every valid argument form can be proved valid in the system. This guarantees that we have enough rules. Our proof system is also sound in the sense that every argument form we can show to be valid is in fact valid. Our rules, in other words, never lead us astray. To speak in terms of formulas rather than argument forms: the system is complete, in that every valid formula is a theorem, and also sound, in the sense that every theorem is valid. The provable formulas are exactly the valid formulas. So this system matches precisely our semantics for the sentential connectives. Every aspect of the meanings of the connectives we've captured in some rule, proof method, or combination of these. Establishing the
f'
Logic, Sets and Functions
70
soundness and completeness of this chapter's natural deduction system requires a fairly sophisticated proof in our metalanguage; such a proof is possible, but we won't attempt it here. To cover all of sentential logic, therefore, we need no more rules or proof techniques. Nevertheless, this section will present some added rules and methods. Everything we can prove with them is still valid. But they are all derivable rules, since they force us to accept nothing new about the logical connectives. They are shortcuts; they abbreviate series of proof lines that we could write in terms of our basic rules. Though they are theoretically unnecessary, then, they save a great deal of time and effort. First, we can develop a new proof method, or, really, an amendment to a basic proof method. An indirect proof shows that something is false by showing that the assumption that it is true leads to a contradiction. As we've introduced it, this means that the conclusion of an indirect proof is always a negation. But it's easy to extend this method to any formula. If we want to show that a formula A is true, we can assume that A is false by assuming -.A, and showing that a contradiction follows. So our extended method of indirect proof will work as follows:
Indirect Proof (Second Form) Sh€>wA
n. n
+ 1.
[-.A
n
+p +q
l3 -.l3
n
AlP
Using this eliminates at least one application of negation exploitation. The first derivable rules pertain to negated formulas. Negations are somewhat difficult to exploit in our basic system, since we can eliminate negations only two at a time. It's very useful, therefore, to have ways of simplifying formulas that begin with a negation. These derivable rules all give equivalents for negated formulas. Because they do give equivalents, they are all invertible; they work in either direction. The first two are often called DeMorgan's Laws. Negation-Conjunction( -.&)
n.-.(A&l3) m.-.A v-,l3
-.&,m -.&,n
Negation-Disjunction( -.v)
n.-.(A V l3) m.-.A&-,l3
-.V,m -.V,n
Negation-Conditional (-. -+)
n.-.(A -+ l3) m.A&-,B
..., -.-.+,
m
....., ---+,n
Negation-Biconditional( -.
n.-.(A i--> B) m.-.A i--> B(orA
i--»
---'~,m i-->
-.B)
-, f-t,n
Natural Deduction
71
Closely related to these rules are several others that define a connective in terms of other connectives. The first allows us to transform disjunctions into conditionals, and vice versa: Conditional-Disjunction( -> V)
n.A -> B m .•A V B
V,m -> V,n
->
The next allows us to characterize the biconditional in terms of the conditional: Conditional-Biconditional( ->+-> )
n.A
+->
m.A
->
B B(orB
->
A)
--++->,
n
Next, four rules say that the order and grouping of subformulas is irrelevant in continued conjunctions and continued disjunctions. The first says that the order of conjuncts makes no difference: A&B is equivalent to B&.A. It thus indicates that conjunction is commutative. Commutativity of Conjunction(&C) n.A&B m.B&A
&C,m &C,n
The second of this group says that the grouping of conjuncts makes no difference or, in other words, that conjunction is associative. Associativity of Conjunction(&A) n.(A&B)&C
&A,m
m.A&(B&C)
&A,n
The third and fourth say the same for disjunctions. Commutativity of Disjunction(VC) n.A V B m.BvA
VC,m VC,n
Associativity of Disjunction(VA) n.(A V B) vC m.A V (B V C)
VA,m VA,n
~.,--
Logic, Sets and Functions
72
Finally, four rules abbreviate commonly used proof steps. The first is a variation of conditional exploitation; the second, a variation of biconditional exploitation; the third, a variation of disjunction exploitation. All these variations allow negations to function readily, without detours. The fourth expresses the principle that anything follows from a contradiction. Conditional Exploitation *( -> E*)
n.A -> 8 m.-,8 p.-,A
->
E*, n, m
Sometimes the above rule is called modus tol/ens. Biconditional Exploitation *( f--+ E*)
n.A f--+ B m. -,A( or-,8) p.-'8(or--.A)
f--+
E*, n, m
Disjunction Exploitation *(V E*)
n.AV8 m. -,A( or-,8) pB(orA)
vE*,n,m
Disjunction exploitation* is occasionally called disjunctive syllogism. Contradiction(!)
n.A m.--.A pB
!,n, m
In addition to these derivable rules, we'll adopt a principle of Replacement. In general, we can apply rules only to formulas with the appropriate main connectives. For instance, we can apply conjunction exploitation only to conjunctions- formulas with & as their main connectives. But invertible rules are justified by the equivalence of the formulas they link. And if we replace a subformula of any formula with an equivalent subformula, we obtain a formula equivalent to the original. If 'p' and '-,-,P' are equivalent, for instance, then so are 'P -> Q' and '-,-,P -> Q'. Consequently, we can apply invertible rules to subformulas as well as formulas. So we can use -,-, to move from '--.-,p' to 'P', but also from '-,-,P -> Q' to 'P -> Q'. The derivable rules are often tremendous time-savers, as attempts to show that they are derivable from the basic rules will demonstrate. We'll dose this chapter by summarizing strategy hints. Overall proof strategies derive, most significantly, from what one is trying to prove, and, secondarily, from what one already has. This table contains some of the most important strategies. In all cases, a direct proof is easiest when it
Natural Deduction
73
can be achieved. But, if it's not obvious how to prove the conclusion directly, then use the strategies listed.
Proof Strategies If you are trying to get:
1. -,A, try using indirect proof. 2. A&B, try proving A and B separately.
3. A V B, try indirect proof. 4. A
-4
B, try using conditional proof.
5. A+--> B, try to prove the two conditionals A
-4
Band B
-4
A.
If you have:
1. -,A, try using it with other lines that have A as a part, or use a derivable rule. 2. A&B, try using &E to get A and B individually.
3. A V B, try (i) getting the negation of one disjunct, and using using V E by taking each case separately. 4. A
-4
B, try (i) getting A and then reaching B by -4 E*.
-4
V E*
to get the other, or (ii)
E, or (ii) getting -,B and then reaching
-,A by
+--> B, try (i) getting either component and then reaching the other by ~ E, or (ii) getting the negation of either component and then the negation of the other by +--> E*.
5. A
These strategies indicate how to construct proofs of various kinds. They can serve as a helpful guide in a wide variety of situations. Sometimes, however, the obvious ploys may not work. When this happens, there are two "safety valves"- strategies that work well when pressure is high. First, when in doubt, use indirect proof. Anything provable can be proved with an indirect proof. Second, within an indirect proof, if it's not clear what to try to prove, choose a sentence letter, and try to prove it. The assumption for indirect proof should lead to a contradiction, so absolutely anything should follow. No matter what letter you select, therefore, you should be able to prove it.
Problems Construct a deduction to show that each of these arguments is valid. 1. If you are ambitious, you'll never achieve all your goals. But life has meaning only if you have ambition. Thus, if you achieve all your goals, life has no meaning.
2. God is that, the greater than which cannot be conceived. If the idea of God exists in our understanding, but God does not exist in reality, then something is conceivable as greater than God. If the idea of God exists in our understanding, therefore, God exists in reality. 3. God is omnipotent if and only if He can do everything. If He can't make a stone so heavy that He can't lift it, then He can't do everything. But if He can make a stone so heavy that He can't lift it, He can't do everything. Therefore, either God is not omnipotent, or God does not exist.
74
Logic, Sets and Functions 4. If the objects of mathematics are material things, then mathematics can't consist entirely of necessary truths. Mathematical objects are immaterial only if the mind has access to a realm beyond the reach of the senses. Mathematics does consist of necessary truths, although the mind has no access to any realm beyond the reach of the senses. Therefore the objects of mathematics are neither material nor immaterial. 5. If the President pursues arms limitations talks, then, if he gets the foreign policy mechanism working more harmoniously, the European left will acquiesce to the placement of additional nuclear weapons in Europe. But the European left will never acquiesce to that. So either the President won't get the foreign policy mechanism working more harmoniously, or he won't pursue arms limitations talks. 6. If we introduce a new product line, or give an existing line a new advertizing image, then we'll be taking a risk, and we may lose market share. If we don't introduce a new product line, we won't have to make large expenditures on advertizing. So, if we don't take risks, we won't have to make large expenditures on advertizing. 7. If we can avoid terrorism only by taking strong retaliatory measures, then we have no choice but to risk innocent lives. But if we don't take strong retaliatory measures, we'll certainly fall prey to attacks by terrorists. Nevertheless, we refuse to risk innocent lives. Consequently, terrorists will find us, more and more, an appealing target. 8. If God is all powerful, He is able to prevent evil. If He is all good, He is willing to prevent evil. Evil does not exist unless He is both unwilling and unable to prevent it. If God exists, He is both all good and all powerful. Therefore, since evil exists, God does not. 9. My cat does not sing opera unless all the lights are out. If I am very insistent, then my cat sings opera; but if I either turn out all the lights or howl at the moon, I am very insistent indeed. I always howl at the moon if I am not very insistent. Therefore, my lights are out, I am very insistent, and my cat is singing opera.
10. If we continue to run a large trade deficit, then the government will yield to calls for protectionism. We won't continue to run a large deficit only if our economy slows down or foreign economies recover. So, if foreign economies don't recover, then the government will resist calls for protectionism only if our economy slows down. 11. If companies continue to invest money here, then the government will sustain its policies. If they don't invest here, those suffering will be even worse off than they are now. But if the government sustains its policies, those suffering will be worse off. Thus, no matter what happens, the suffering will be worse off. 12. We cannot both maintain high educational standards and accept almost every high school graduate unless we fail large numbers of students when (and only when) many students do poorly. We will continue to maintain high standards; furthermore, we will placate the legislature and admit almost all high school graduates. Of course, we can't both placate the legislature and fail large numbers of students. Therefore, not many students will do poorly. Construct deductions to demonstrate the validity of these argument forms.
13. P V Q;-'P :. Q 14. P;-,P:. Q
15. P
-7
Q;-,Q :. -,p
Natural Deduction
75
16. P ...... Qj-.P .. -.Q 17. P ...... Qj-.Q .. -.P 18. p ...... Qj .. P ---. Q 19. P ...... Q :. Q ---. P 20. P&Q:. Q&P' 21. (P&Q)&R:. P&(Q&R)
22. P
V
Q :. Q V P
23. (PVQ)VR:.PV(QVR) Use deduction to solve each of these problems. 24. Holmes and Watson question three suspects: Peters, Quine and Russell. Hearing that their responses conflict, Holmes declares, "If Peters and Quine are telling the truth, then Russell is lying." Watson seemingly assents, saying, "Indeed, at least one of them is telling us a falsehood." Irritated, Holmes insists, "That's not all, my dear Watson! We know that Russell is the trickster, if the other two are telling us the truth!" Show that Holmes' irritation is unjustified by showing that his original statement is equivalent to Watson's. 25. Jones, feeling upset about the insecurity of the Social Security system, sighs that he faces a dilemma: "If taxes aren't raised, I'll have no money when I'm old. If taxes are raised, I'll have no money now." Smith, ever the even-tempered one, reasons that neither of Jones' contentions is true. Jones answers, "Aha! You've contradicted yourself!" Show that Smith's assertion that both Jones' claims are false is indeed contradictory. 26. Roger, a hapless accounting major, is trying to analyze a problem on an accounting exam. He needs to figure the tax liability of a corporation engaged in overseas shipping. Some of the fleet counts as American for tax purposes, and some does not. Poor Roger recalls the definition in the Tax Code of an American vessel as running like this: "Something counts as an American vessel if and only if (1) it is either numbered or registered in the U.S., or (2) if it is neither registered nor numbered in the U.S., and is not registered in any foreign country, then either its crew members are all U.S. citizens, or they are all employees of U.S. corporations." Show that this is the wrong definition, by showing that it implies that if something is registered in a foreign country, it counts as an American vessel. 27. On the way to the barber shop (adapted from Lewis Carroll): you are trying to decide which of three barbers- Allen, Baker and Carr- will be in today. You know Allen has been sick, and so reason that (1) if Allen is out of the shop, his good friend Baker must be out with him. But, since they never leave the shop untended, (2) if Carr is out of the shop, then, if Allen is out with him, Baker must be in. Show that (1) and (2) imply (a) that not all three are out; (b) that Allen and Carr are not both outj (c) that, if Carr and Baker are in, so is Allen. Show that each of our derivable rules is in fact derivable from the basic rules, by using only basic rules to prove valid these argument forms. 28.
* -.(P&Q)
:. -.P V -.Q
29.
* -.P V -.Q
:. -.(P&Q)
76
Logic, Sets and Functions
32.
* -.(P V Q) :. -.P&~Q * -.P&~Q :. -.(P V Q) * -.(P -> Q) :. P&~Q
33.
* P&-.Q :. -.(P
34.
36.
* P <-+ ...,Q :. -.(P <-+ Q) * -.(P <-+ Q) :. P <-+ ~Q * P -> Q .. -.P V Q
37.
* -.P V Q :. P
30. 31.
35.
->
->
Q)
Q
Show that the following argument forms are valid by using deduction. 38.
*
39.
42.
* -.(-.P&Q);(P V R) -> -.(Q,&S);Q <-+ S .. Q1 -> -.(Q V S) * P -> (Q V R);(-'Q&Q1) V (S -> P);-.(-.R -> -.P) :. -.S V Q * P -> (Q -> R);(~Q <-+ S) -> -,PiP V Q1 :. (R&S) -> Q1 * (P <-+ Q) <-+ R;-.(P <-+ -.R);(Q V -.S) -> S1;:' S1
43.
* -.5 -> -.SdS&5,)
44.
* (Q <-> -.P) -> -.R;(~Q&S) V (P&Q1);(S V Q1) -> R:. P -> Q * (P&-.R) <-+ (5 V ~Q);51&«~S&-.R) -> P);(S, -> Q) V (5, -> R)
40.
41.
45.
\.
-.(P
-> ~Q);R ->
(-.p V -.Q);(R V S)
->
(P
<-+
<-+
S1 :. S1
Q);-.(-.p V Q) :. S1
<-+
->
S
-.5,
.. Q&R
Chapter 4
Quantifiers Aristotle's theory of the syllogism treats only arguments having a very restricted form. Every sentence must have the structure 'All Fare G', 'No Fare G', 'Some Fare G' or 'Some F are not G', where F and G are general terms, expressions true or false of individual objects. Every syllogistic argument must have two such sentences as premises and one as conclusion, with the terms meshing in just the right way. As a result, syllogistic logic covers a rather limited domain. Sentential logic, by taking sentences as basic units of analysis, oversees a broader realm. It can handle arguments with any number of premises and sentences of any length and with any degree of complexity. But it too suffers from narrow horizons. Just as the theory of the syllogism can't solve problems characteristic of sentential logic, so sentential logic fails to solve syllogistic problems. Consider even a simple example: 1. All cows are mammals.
All mammals are animals. :. All cows are animals. This argument is surely valid. Yet sentential logic cannot explain why. It has no choice but to construe (1) as
2. C M
:. A which plainly fails to be valid. The same happens with any syllogism. The validity of arguments such as (1) depends on the structure within sentences, not on the structure relating distinct sentences. No theory that declines to analyze what sentential logic calls "atomic" sentences can hope to account for syllogistic reasoning. The split between syllogistic and sentential logic, which began in Greece in the third or fourth century B. C., persisted for more than two thousand years. Neither theory could account for arguments that the other took as paradigms of correct reasoning. In the nineteenth century, sentential logic became symbolic, attaining a fairly high level of mathematical sophistication. The symbolization of logic paved the way to a unification of the sentential and syllogistic realms. Two logicians working independently- Gottlob Frege and Charles Saunders Peirce- overcame the ancient divergence between syllogistic and sentential logic in 1879. They introduced symbols representing determiners, such as 'all', 'some', 'no', 'every', 'any', etc. Frege and Peirce used two symbols: the universal quantifier, which we will write "\:I' (Peirce's work has 'P'), and the existential 77
!'"
Logic, Sets and Functions
78
quantifier, '3 ' (in Peirce, "S'). The universal quantifier corresponds roughly to the English 'all', 'every' and 'each'; the existential quantifier, to the English 'some', 'a' and 'an'. To see how introducing quantifiers proved to be the crucial move in unifying sentential and syllogistic logic, notice that the theories use different basic units. Sentential logic treats sentences lacking connectives as unanalyzed, while syllogistic logic so treats general terms. Frege and Peirce combined a scheme relying on general terms with another relying on sentences by making general terms behave, in essence, as if they were sentences. Their move enabled sentential tools to work on syllogistic and, as it turned out, far more complex arguments.
4.1
Constants and Quantifiers
Sentential logic is limited precisely because it takes sentences as its basic, unanalyzed units of explanation. To expand this logic to include syllogistic arguments, we must look inside "atomic" sentences to see how they are put together. In general, they consist of a main or subject noun phrase and a main verb phrase. We can gain some insight into the consolidation of sentential and syllogistic logic by examining the character of verb phrases. Consider a few examples: 3.
(a) is a man (b) knows some people who live in Oklahoma City (c) sleeps very soundly (d) kicked the ball into the end zone (e) thought that Yosemite would be more fun to visit (f) gave Fred a copy of the letter
All verb phrases are general terms in the sense of syllogistic logic. They are true or false of individual objects. Pick any object you like; it will either be a man or not be a man. It will either sleep very soundly or not sleep very soundly. It will be true that either it gave Fred a copy of the letter or it didn't. Alone, verb phrases and other general tenus are not true or false. But we can think of them as yielding a truth vaiue when combined with an object. Objects of which the verb phrase or general term are true satisfy it; the verb phrase or general term applies to them. Equivalently, we can think of verb phrases and other general terms as classifying objects into two categories: those of which they are true, and those of which they are false. The first category- the set of objects of which the verb phrase or general term is true- is called its extension. Verb phrases combine with noun phrases to form sentences. Since verb phrases are expressions that are true or false of particular objects, noun phrases must specify an object, or a group of objects, and say something about the application of the verb phrase to them. Noun phrases take several forms. In this chapter, we'll concentrate on two of them. First, noun phrases may simply pick out a single object by naming it. A sentence containing a proper name as its subject will be true if the verb phrase is true of that object, and false otherwise. Each of these sentences results from combining a proper name with a verb phrase from (3): 4.
(a) Socrates is a man (b) Maria knows some people who live in Oklahoma City (c) Mr. Hendley sleeps very soundly (d) Barr kicked the ball into the end zone (e) Penelope thought that Yosemite would be more fun to visit
Quantifiers
79
(f) Nate gave Fred a copy of the letter To suggest that (4)a. is true just in case the verb phrase 'is a man' is true of Socrates, we could write it in the form (5)a., or, more simply, (5)b.: 5.
(a) is a man(Socrates) (b) Man(Socrates).
We can read (5)b. as saying that man applies to, or is true of, Socrates. Our formal language will thus need at least two kinds of symbol to represent sentences such as (4)a. First, we'll use lower case letters from the beginning of the alphabet, with or without numerical subscripts, to represent proper names; we'll call these symbolic names individual constants or, more simply, constants. Second, we'll use upper case letters from the middle of the alphabet, with or without numerical subscripts, as predicate constants, or more simply, predicates. Each predicate comes with a number assigned to it; predicates are called n-ary if they are assigned the number n. Every predicate yields a truth value when combined with a certain number of objects. The assigned number indicates of how many objects at once the predicate is true or false. Recall that a general term is true or false of individual objects. We will therefore symbolize simple general terms such as man with singulary predicates, that is, predicates to which 1 is assigned. These predicates are true or false of single objects, and produce sentences when combined with a single proper name. Other predicates yield sentences only when combined with two names; they are true or false of two objects taken together. Such predicates are assigned the number 2, and are called binary. They are useful for symbolizing, among other things, transitive verbs. Respect, for example, applies not to objects taken individually, but to objects taken in pairs. It requires a direct object. We can ask whether Robin respects Julia, but not simply whether Robin respects. Letting 'a' symbolize Socrates, and 'M' symbolize man, we can symbolize 'Socrates is a man' by
6. Ma, which is merely a symbolic version of 4(a). 'Ma' is a formula of quantificationallogic. One way of building formulas, then, is to combine individual constants with predicates. A second sort of noun phrase we'll consider in this chapter consists of a determiner, such as every or some, together with a common noun such as man or truck. The common noun may be modified by adjectives, adjectival phrases, prepositional phrases, or relative clauses. Whatever its grammatical structure, however, the modified noun will constitute a general term. Noun phrases of this more complex sort, when combined with the verb phrases in (3), yield the sentences 7.
(a) (b) (c) (d) (e) (f)
One reporter who covered the match is a man A few friends know some people who live in Oklahoma City Every endomorph sleeps very soundly Several prospects kicked the ball into the end zone A taxi driver thought that Yosemite would be more fun to visit Nobody gave Fred a copy of the letter
Nobody, in (7)f., is a special case; the word itself contains both a determiner and a general term, and is equivalent to no person. We can gain some insight into the structure of the sentences in (7) by examining some related sentences that contain subject noun phrases that, while complex, consists only of determiners and the rather colorless general terms thing and object. To begin, consider the sentence
80
Logic, Sets and Functions 8. Something is missing.
To symbolize this sentence, we can't use any individual constant for something. We don't want to say that missing applies to any object in particular. The sentence says that missing applies to some object. So, using our earlier strategy, we could try writing 9. Missing(something). But it is extremely useful to think of missing, a general term, as true or false of objects. That, after all, is what motivated us to introduce the pattern of (9) in the first place. We would like Missing(Ralph) to mean that Missing is true of Ralph. The name Ralph picks out a particular entity. But, if we write Missing(something), although we can continue to read it as saying that Missing is true of something, the something picks out no object in particular. The form of (9) tempts us to ask, "Well, then, what does something stand for?" Medieval logicians grappled with this problem for centuries without producing a fully satisfactory solution. The problem, according to modern logic, is that this is the wrong question. We don't want to say that Missing is true of the object denoted by something, but rather that Missing is true of some object. Quantification theory allows us to say, in effect, that x is missing is true for some object x. So, instead of thinking of a complex noun phrase such as something as picking out some object or objects satisfying a general term, we can indicate that general term is true of x for some value of x. The existential quantifier has exactly this function. To express the idea that x is missing is true for some x, we write the symbolic equivalent of 10. (for some x)(x is missing), which is 11. 3xMx.
We may read (11) as saying the following: 12.
(a) for some x, x is missing (b) some x is of such a kind that x is missing (c) there is an x of such a kind thatx is missing (d) an x is of such a kind that x is missing.
In good English, these become 13.
(a) something is missing (b) there is something missing (c) an object is missing.
Notice that the words thing and object serve in English much as variables such as x serve in quantification theory. Variables link quantifiers to the predicates they accompany. This strategy, then, requires two new sorts of symbol. First, we must introduce individual variables, or, more simply, variables. Variables will be lower case letters from the end of the alphabet, again with or without numerical subscripts. They act, in quantificationallogic, much as variables for numbers such as tn' or 'x' act in number theory or algebra. They stand for no objects in particular; instead, they mark places where names of particular objects could go. Second, the language of quantification theory must include quantifiers: '3 " the existential quantifier, and 'V " the universal quantifier.
Quantifiers
81
To take another simple example, suppose we want to say, recalling the title of a popular song, that everything is beautiful. To say that Pittsburgh is beautiful, or that Ingrid Bergmann is beautiful, we can introduce the predicate 'B' and the individual constants 'p' and 'b' and write 'Bp' and "Bb'. To say that everything is beautiful, however, we need to say that 'x is beautiful' is true for every object x. The universal quantifier does just this. We can prefix the quantifier to 'Bx', writing the symbolic equivalent of 14. (for every x)(x is beautiful), or
15. VxBx. (15) says
16.
(a) (b) (c) (d) (e) (f) (g) (h)
for every x, x is beautiful for all x, x is beautiful for each x, x is beautiful for any x, x is beautiful every x is of such a kind that x is beautiful all x are of such a kind that x is beautiful each x is of such a kind that x is beautiful any x is of such a kind that x is beautiful
or, in plain English, 17.
(a) (b) (c) (d)
everything is beautiful all things are beautiful each object is beautiful any object is beautiful.
All, every, each and, usually, any thus receive the same symbolization. In English, these words differ subtly but significantly in meaning. Note, for example, that though 'any object is beautiful' seems to mean just what 'everything is beautiful' does, 'anything is beautiful' sounds strange. Quantification theory cannot capture all the differences between these determiners, but it can capture some, as a later section of this chapter will show. The theory, despite its idealizations, succeeds surprisingly well in explaining the validity or invalidity of English arguments depending on these determiners. Variables, which in effect represent English words such as thing and object, have no meanings independent of their symbolic context. In (11) and (16), 'M' and 'B' represent missing and beautiful, respectively. They cannot be interchanged without changing the translation manual relating English sentences to their symbolic representations. Variables, in contrast, can be interchanged, with very few restrictions, without altering meaning. Something is missing could just as well be symbolized by 18. 3yMy
as by (16), and Everything is beautiful could just as well be represented by 19. VzBz
as by (21). This is true because the variables 'x', 'y' and 'z' themselves have no significance. They serve only to link quantifiers to predicates, and mark places where constants could be placed. Any sort of mark could do the same job.
Logic, Sets and Functions
82
4.2
Categorical Sentence Forms
Sentences such as something is missing or everything is beautiful carry one only so far. Most of the time, we want to say something about, for example, some people or every frog, not about just something or everything. We need to be able to handle sentences with subject noun phrases that contain more complicated general terms. If we could do this, it would be easy to represent any sentence having one of the four classic categorical forms in quantification theory. Consider first universal affirmative sentences, having the structure 20. All Fare G. We might want to represent, for instance, 21. All frogs swim. We already know that' Sa' can represent, say, Albert swims, and that 'VxSx' can represent Everything swims. To say that all frogs swim, then, we could begin by saying that swims is true for all frogs, writing (21) as 22. SWims(all frogs). But this leads to the puzzle we encountered earlier: to what does all frogs refer? If we want what follows swims in (22) to pick out a particular object, then simply writing all frogs there won't do. Quantification theory, in effect, focuses on the determiner all. It combines with a general term, in this case frogs, to form a noun phrase; that noun phrase in turn combines with another general term, the verb phrase swims. We can see all, then, as relating two general terms. The general terms, in this case, are simple, so we can symbolize them as singulary predicates 'F' and'S'. We can symbolize all with the universal quantifier. To link the quantifier and the predicates together, we can use a variable, say, 'x'. We need to search, then, for the proper logical relationship among 'Vx', 'Fx' and 'Sx' ; we need to say that every x of such a kind that x is a frog swims. In essence, the theory points out that All frogs swim says that, if an object is a frog, it swims. The Stoics recognized this equivalence between universal and conditional sentences, but not until Frege did anyone see its significance. So the relationship we are looking for is the conditional; we want to say something like 23. for all x, if x is a frog, then x swims which we can represent, using quantifiers and sentential connectives; as
24. Vx(Fx
-->
Sx).
This reasoning applies to all universal affirmative sentences. So we can represent anything having the form (21) as
25. Vx(Fx
-->
Gx).
Second, consider particular affirmative sentences, having the structure 26. Some Fare G. To take an example, let's try to represent 27. Some people are bothersome.
Quantifiers
83
Once again, we can represent Edna is bothersome as 'Be', and Some things are bothersome as '3xBx'. (27) seems to require that we express a relationship between the general terms people and bothersome in a way corresponding to the meaning of the determiner some. Here, conjunction expresses the right relationship: to say that some people are bothersome is to say that, for some objects x, x is a person and x is bothersome. So we can write (27) as
28. ?Jx(Px&Bx). In general, particular affirmative sentences translate into quantification theory as
29. ?Jx(Fx&Gx). Conjunction, together with the existential quantifier, succeeds in representing sentences having the form of (32). Particular negative sentences, which have the form 30. Some F are not G, are very similar to their affirmative counterparts. To say that some F are not G is to say that, for some object x, x is F but not G. So the combination of an existential quantifier and conjunction again does the trick; we can represent sentences having the structure (30) as
31. ?Jx(Fx&-.Gx). Finally, universal negative sentences, which have the form 32. No Fare G, have two equivalent and equally natural translations into the language of quantificationallogic. Suppose we want to symbolize No man is an island. We might think that we need a new quantifier to symbolize the determiner no. But we can symbolize the relation between general terms that no expresses by using either the existential or the universal quantifier, together with negation. No man is an island is a direct denial of the sentence Some man is an island, so we can represent it as the negation of a particular affirmative sentence form. Letting 'F' and 'G' represent man and island, respectively, this strategy yields
33. -,3x(Fx&Gx) which we can read 34.
(a) it is not the case that, for some x, x is a man and x is an island (b) there is no x such that x is a man and x is an island
or, more naturally, 35.
(a) It is not true that some man is an island (b) there is no man who is an island (c) there is nothing that is both a man and an island (d) no man is an island.
But no man is an island also bears some similarity to universal affirmative sentence forms, being equivalent to if an object is a man, it is not an island. It says about every man, in other words, that he is not an island. So, adopting the same translation manual as before, we can write
Logic, Sets and Functions
84
36. Vx(Fx
---+
,Gx),
which says 37.
(a) for all x, if x is a man, then x is not an island (b) every x is of such a kind that, if x is a man, then x is not an island
or, in plain English, 38.
(a) every man is not an island (b) no man is an island.
(33) and (36) both represent universal negative sentences. As we might hope, they are equivalent. Any categorical sentence form, therefore, has a representation in quantification theory. The logic of this chapter has the power to cover the entire realm of syllogistic reasoning. In fact, as the next few sections will demonstrate, it has the power to capture an extremely wide range of English arguments.
4.3
Polyadic Predicates
General terms are true or false of objects. Man, woman, animal and mortal, in addition to intransitive verbs such as swim and live, all apply or fail to apply to objects considered one by one. Because they are true or false of single objects, we can symbolize them using monadic or singulary predicates. Some English expressions, as we have seen, are true or false of objects taken in pairs, or even triples, quadruples, etc. Predicates symbolizing them are polyadic. Consider, for example, the English verb distrust. It makes no sense to ask whether distrust applies to an object considered alone. Does Socrates distrust? Does this book distrust? These questions seem incoherent. But we can ask whether Socrates distrusts Alcibiades (the answer: yes), or whether this book distrusts its author (the answer: no- books don't distrust anything). So distrust applies, or fails to apply, to objects taken in pairs. The verb needs, not only a subject, but also a direct object. The same holds of most transitive verbs. Predicates that are true or false of pairs of things are not only polyadic but, more specifically, dyadic (or binary). To represent general terms symbolically, quantification theory uses predicates. Thus'M a' might symbolize Alonzo is a man, and 'ej' might represent Joan lives in California. The same strategy works for polyadic predicates. We can symbolize 39. Hanna admires Bob Dobbs, which, because it asserts that admires applies to the pair of objects consisting of Hanna and Bob, has the structure 40. Admires(Hanno, Bob Dobbs), as
41. Mhb. Similarly, we can represent 42. Joanie loves Chachi
Quantifiers
85
as 43. Lje.
Given a dyadic predicate, say, 'D', symbolizing distrusts, we can construct a formula of quantification theory in several ways. First, we can combine the predicate with constants, obtaining formulas such as 'Dab' and 'Dec'. The first constant following the predicate marks the subject position; the second marks the object position. Notice that the same constant may appear more than once; this allows us to symbolize sentences such as Alan distrusts himself. Second, we may prefix a quantifier to the predicate and combine it with variables. We might do this in only one place, filling the other with a constant. To translate 44. Sam distrusts everyone, for example, we can write 45. 'VxDsx,
and, to translate 46. Everyone distrusts Sam, we can write 47. 'VxDxs.
Notice that the only difference between these formulas is the order of the variables and constants. This reflects the distinction in English between subject and object. Since' Dxy' symbolizes x distrustsy, the first position is always occupied by the subject of distrust, the second, by its object. So, in ''VxDsx', IS' is in the subject position. In ''VxDxs', IS' is in the object position. We may construct a formula from a predicate by prefixing a quantifier for each place. There may be two variables in the formula; each may link the predicate to either a universal or an existential quantifier. So using combinations of quantifiers and variables with 'D' might yield any of the following formulas: 48.
(a) 3x3yDxy
(b) 3y3xDxy (c) 3x'VyDxy
(d) 3y'VxDxy (e) 'Vx3yDxy
(f) 'Vy3xDxy (g) 'Vx'VyDxy
(h) 'Vy'VxDxy
(i) 3xDxx (j) 'VxDxx
86
Logic, Sets and Functions
To see what these formulas mean, let's examine them one by one. '3x3yDxy' says that for some x and some y, x distrusts y. This is an ugly way of saying that something distrusts something, or, if we assume for now that we are paying attention only to people, that somebody distrusts somebody. '3y3xDxy' says that for some y and some x, x distrusts y. In English, we can express this idea by saying that somebody is distrusted by somebody. We can capture the effect of reversing the variables by using the passive voice, which reverses subject and object in English. Of course, somebody distrusts somebody and somebody is distrusted by somebody are equivalent. So are '3x3yDxy' and '3y3xDxy'. In general, reversing the order of adjacent existential quantifiers produces an equivalent formula. 'VxVyDxy' says that, for all x and for all y, x distrusts y. Again restricting our attention to people, this means that everybody distrusts everybody. 'VyVxDxy' reverses the order of the quantifiers, saying that, for all y and all x, x distrusts y. Again we can use the passive voice to reverse subject and object in English; 'VyVxDxy' represents the English Everybody is distrusted by everybody. Everybody distrusts everybody and everybody is distrusted by everybody are also equivalent. Reversing adjacent universal quantifiers yields an equivalent formula. Combinations of existential and universal quantifiers, however, do not allow such switches. '3xVyDxy' says that there is an x such that, for all y, x distrusts y. This corresponds to the English Somebody distrusts everybody. '3yVxDxy' says that there is a y such that, for all x, x distrusts y; that amounts to the English Somebody is distrusted by everybody. 'Vx3yDxy' says that, for all x, there is a y such that x distrusts y. It thus represents Everybody distrusts somebody. 'Vy3xDxy' says that, for all y, there is an x such that x distrusts y. This corresponds to Everybody is distrusted by somebody. Finally, '3xDxx' says that some x is such that x distrusts x, corresponding to the English Somebody distrusts himself (or herself ), while 'VxDxx' says that every x is such that x distrusts x, corresponding to Everybody distrusts himself (or herself ). To put this into a table (again assuming that we are speaking only of people): 49. Formula English Sentence a. 3x3yDxy Somebody distrusts somebody b. 3y3xDxy Somebody is distrusted by somebody c. 3xVyDxy Somebody distrusts everybody d. 3yVxDxy Somebody is distrusted by everybody e. Vx3yDxy Everybody distrusts somebody f. Vy3xDxy Everybody is distrusted by somebody g. VxVyDxy Everybody distrusts everybody h. VyVxDxy Everybody is distrusted by everybody i. 3xDxx Somebody distrusts himself j. VxDxx Everybody distrusts himself No sentence in this table that contains both universal and existential quantifiers is equivalent to any other. Somebody distrusts everybody and everybody is distrusted by somebody differ in meaning. Somebody distrusts everybody tends to say that some one person (for instance, Sam) distrusts everybody. Everybody is distrusted by somebody, in contrast, tends to say that everybody is the object of somebody-or-other's distrust. The latter sentence, that is, allows that Fred distrusts you, Greta distrusts me, etc. The former requires that some one person distrust you, me and everyone. Much the same holds of everybody distrusts somebody and somebody is distrusted by everybody. The former tends to mean that everybody distrusts somebody or other; you may distrust Pat, I may distrust Lou, and so on. The latter, however, tends to require that one person is the object of everyone's distrust; you and I and everyone else distrust some one individual, say, Zeke. Quantification theory reflects these differences in meaning by giving quantifiers different positions. When an existential quantifier appears to the left of a universal, the formula is saying that some one
Quantifiers
87
object stands in a particular relation to every object. When an existential appears to the right of a universal, the formula says that, for each object, there is some object or other standing in some relation to it. English often distinguishes these senses by placing the existential first, for the former, or the universal first, for the latter. So the order of the English determiners frequently matches the order of the quantifiers. There are exceptions- each and any, for example, almost always go to the extreme left- and ambiguities, as the above sentences show. The linguistic rules governing this in English are highly complex and controversial. Consequently, we will often have to rely on our own intuitions about meaning to determine, in a given context, the quantifiers' correct order.
4.4
The Language Q
So far we've spoken of quantification theory in informal terms. Before going further, however, it is important to define the language of the theory more precisely. In the next section, we'll discuss some guidelines for representing English sentences as quantificational formulas. These guidelines naturally depend on the notion of what a formula of quantification theory is. The syntax of quantification theory comprises a vocabulary of symbols and a set of formation rules for combining the symbols to form formulas. The vocabulary of quantification theory includes the following:
Vocabulary Sentence Letter Constants: A, B, C, ... , Z.* with or without numerical subscripts n-ary Predicate Constants: An, Bn, cn, . . ., zn * with or without numerical ... Individual Constants: a, b, e, ... , s * with or without ... Individual Variables: t, u, v, w, x, y, zwith or without ... Sentential Connectives: -', ->, &, v, ...... Quantifiers: V,:J Grouping Indicators: (, ) This is the vocabulary of sentential logic, supplemented with predicates, constants, variables, and quantifiers. The rules for combining these symbols to construct formulas are more complex than in sentential logic. To understand them, we need to understand substitution. Suppose that we are given an expression A. We can construct a new expression, which we can call, abstractly, A[e/d] (A, with e substituted for d), by replacing every occurrence of the constant d with an occurrence of the constant e. Similarly, we can form A[x/y] by replacing every occurrence of the variable y with an occurrence of the variable x. We can also substitute constants for variables, and variables for constants as well. To take an example, suppose A is 'Fab'. Then A[e/a] is 'Feb'. Or, suppose A is 'VxFxa'j then A[y/a] is 'VxFxy'. The substitution operation is defined for all expressions, whether or not they are formulas. But applying it to a formula may yield a string of symbols that is no longer a formula.
88
Logic, Sets and Functions
Formation Rules Any sentence letter constant is a formula. An n-ary predicate followed by n constants is a formula. If A is a formula, -.A is a formula. If A and B are formulas, then (A -. B), (A&B), (A V B), and (A f-> B) are formulas. If A is a formula with a constant c, and v is a variable that does not appear in A, then 3vA[v/ c] and VvA[v/c] are formulas. Every formula can be constructed by a finite number of applications of these rules. The formation rules characterize the formulas of the language of quantification theory, Q. Notice that every formula of sentential logic is also a formula of Q. In addition, Q allows us to link quantified formulas with sentential connectives and to combine sentence letters with the quantificational apparatus. When we form VvA[v/ c] or 3vA[v/ c] in accordance with these rules, we'll say that the scope of Vv or 3v is all of VvA[v / c] or 3vA[v / c]. As in the case of sentential logic, we'll adopt a few simplifying conventions to make formulas more readable. First, we'll generally drop the superscript that indicates whether a given predicate is singulary, binary, etc. Second, we'll count the result of deleting a formula's outside parenthesis pair as an abbreviation of the actual formula. So, the first formula in each of these pairs abbreviates the second: 50.
(a) 3x(Fx
-+
Gx)
-+
Vx3yHxy ; (3x(Fx -. Gx)
(b) VzVw3t(Fzt&Gwz)
f->
-+
P; (VzVw3t(Fzt&Gwz)
Vx3yHxy) f->
p)
Third, we'll allow brackets and braces to count as sloppily drawn parentheses when that increases the readability of a formula. Fourth, because we'll usually drop predicate superscripts, we'll avoid using the same capital letter as both a monadic and a polyadic predicate, or as both a sentence letter and a predicate letter, within the same formula; though
51. 3xVy(Fxy
-+
Fa)
is correctly formed, using 'F' to represent two English expressions at once tends to be confusing; the formula, properly written out, would be '3xVy(F 2 xy -. FIa)', which makes it clear that there are two different predicates. Fifth, lower case letters can be constants or variables; the vocabulary specifies that 'a' through's' are constants, and 't' through 'z' are variables. We'll maintain this usage strictly only for variables, which are similar enough to constants in function to cause confusion. Several points about formulas of Q deserve mention. Note that only variables may appear with quantifiers; '3aFa' is not a formula. Neither is 'Vp(p -. q)' or '3FFa'. Intuitively, individual constants and variables take objects as values. In Q we can quantify over objects, speaking about all objects of a certain kind, or some objects of that kind, etc. We cannot do the same with sentences or predicates. Because it allows quantification over individuals alone, Q is a system of first-order quantification, sometimes called first-order logic. Other logical theories, called higher-order logics, do allow quantification over sentences and predicates, but at the price of substantial complication.
Problems Evaluate each of the following as (a) an official formula of Q, (b) a conventional abbreviation of a formula of Q, or (c) neither of the above. 1. FIX
Quantifiers
89
3. Fx
->
Fy
4. Fa
->
Fe
5. (Fx
Fy)
->
6. (Fla
FIb)
->
~xFx ->
7.
Fx
...~, ..
/8. '~x(Fx -> Fx)
..
\,.
9. ~xFIx
'
->
Fia
->
Fa)
......... .... '
.'lLpxFx '", ..'-«..;...1
12. VxVyFxy
13. VxVyF 2xy
14. VxVyFxy
Fa
->
->
17. (VxVyF 2xy
18. VxVyFxy
F 2ab
->
->
15. VxVy(Fxy
16.\:jxVyFxy
Fyx
->
Fyx)
VxVyFyx
->
->
VxVyF2 yX )
Fy
-> ~xFx
19. VxFx
-> ~xFx)
20. Vx(Fx
21. (VxF1x
....
->
3xF I x)
.?2~.~xVYGy 23. VxVyVb(Fxy V Fyb) 24.
Vx~F(Fx ->
Fa)
25. G1y 26. G 2 b
28. VxGxy
29. VyGly
->
30. Vy(G1y 31. VyGy
->
->
GI Z GIz)
Gy
Logic, Sets and Functions
90
32. Vy( Gly
Gly)
f--T
33. VXF2 xy
->
VyG 2yx
34. Vx(Fxy
-4
VyGyx)
35. VxVy(Fxy
->
Gyx)
36. Vy(VxFxy
-4
Gyx)
37. VxVy(3zFyz
38. Vx(Vy3zF 2yz 39. VxVy3z(Fyz
(Gy&Hzx))
f--T
f--T
f--T
(Gly&H 2 zx))
(Gy&Hzy))
40. (VxVy3z(Fyz&Gx)
f--T
Hzy)
Taking each expression below as A, write, where possible, (a) A[cld], (b) A[dle], (c) A[xle], (d) A[yld], and (e) A[ylx]' and say whether the result in each case is a formula. (Count abbreviations of formulas as formulas.) If the substitution is impossible, say so. 41. Hcd
42. Hee 43. Hex 44. Hxy 45. VxFx
f--T
Ge
46. Vx(Fx
f--T
Ge)
47. VxFxe
f--T
3xFdx
48. VxFxc
f--T
3yFdy
49. Vx(Fxe 50. Fxe
f--T
f--T
3yFdy)
3yFdy
51. 3xFxc&VxFxd 52. Fxe&VxFxd
4.5
Translation
With the addition of polyadic predicates, quantification theory has the power to express and evaluate a very large group of sentences and arguments. In this section we'll present a guide to representing English sentences in the theory. To translate even a simple sentence, we must distinguish its grammatical subject from its grammatical predicate. Because the word predicate takes on a different meaning in logic, we've called grammatical subjects subject (or main) noun phrases, and grammatical predicates main verb phrases. As we've indicated, it is easy to translate simple sentences such as All men are mortal, Some computers are not reliable, and Nobody admires everybody into Q. But noun and verb phrases may become far more complex. After listing some of the ways in which this can happen, we'll explain how quantification theory can incorporate them.
Quantifiers
4.5.1
91
Noun Phrases
Some noun phrases are easy to handle in Q. Proper names translate as individual constants; common nouns, such as woman and airplane, translate as monadic predicates. But here the simplicity ends. Determiners The determiners all, each, any and every translate as universal quantifiers, while some and a(n) generally translate as existentials. We say "generally" because even these rules have exceptions. First, a and an have a generic use, where they seem to talk about typical members of a kind. Thus 59. A whale is a mammal is not saying that some whales are mammals, but that all are. In these cases, a and an correspond roughly to universal quantifiers. Second, a, an and some all interact with conditionals whenthey are part of the antecedent. 60. If you steal something, you'll get into trouble
can be translated straightforwardly as a conditional with a quantified antecedent (where a represents you): 61. 3xSax
->
Ta.
But, when the consequent contains a word that refers back to something in the antecedent, this doesn't work. We could try to translate 62. If you steal something, you'll pay for it as 63. 3xSax
->
Pax
but this is not a formula; the final occurrence of 'x' is free (not within the scope of an appropriate quantifier phrase). Changing the parentheses so that the quantifier has scope over the entire formula in itself does not help. 64. 3x(Sax
->
Pax)
is a formula, but it says the wrong thing. Because of the nature of the conditional, it is equivalent to 65. 3x(--,Sax
V
Pax).
But this says that there is an object that either you don't steal or you pay for. And this is true if there is an object you don't steal. But the same does not hold of (62), the truth value of which is not determined by whether or not there are things you don't steal. To represent (62), we must use a universal quantifier with the entire formula as its scope: 66. Vx(Sax
->
Pax).
92
Logic, Sets and Functions
This says that everything you steal, you pay for, which is equivalent to (62). So, in certain cases in which they appear in the antecedent of a conditional, with reference back to the antecedent in the consequent, the determiners a, an and some correspond to universal quantifiers in Q. Notice that If you steal something, you'll pay for it is equivalent to If you steal anything, you'll pay for it. Any translates as a universal quantifier, but with the widest possible scope. Each usually takes wide scope among quantifiers; any, however, demands wide scope over connectives as well. EveTy and all make no such demand. This explains why the above sentences are not equivalent to If you steal everything, you'll pay for it and If you steal each thing, you'll pay for it. It also explains why any often seems similar to some or a. John didn't see any deer is equivalent to John didn't see a deer, not to John didn't see every deeT, because any appears as a universal quantifier to the left of the negation sign ('Vx-,(Sjx&Dx)'). Everything, anything, something, etc., all act like the corresponding determiners; thing functions, more or less, as a variable. Everybody, anybody, somebody, everyone, anyone, and someone all act like every person, any person, etc. They generally force the use of a quantifier together with the monadic predicate 'P' (for person) linked to the remainder of the formula in the appropriate way. No, and the related nobody, nothing, etc., all translate into Q in the ways suggested by representations of categorical sentence forms. They correspond to negations of existential quantifiers, or to universal quantifiers applying to negations. Only, though it is not really a determiner, functions much like all, except that it reverses the order of the relevant expressions. Only Fare G amounts to All G are F. English contains many other determiners that Q can't translate. Many, several, a few, fe, most, infinitely many, and much, for example, elude the powers of Q. Sentences containing them do not translate into quantification theory. We haven't discussed the semantics of Q. But an important part of interpreting formulas of Q is assigning them a universe of discourse or domain. This is a set of objects. Intuitively, the universe of discourse is a set containing the objects the formulas are talking about. The quantifiers range over the domain, in the sense that we interpret the universal quantifier as saying that something is true for all elements of the domain and the existential quantifier as saying that something is true for some element of the domain. We can often simplify translation by assigning an appropriate universe of discourse. If, within the context of an argument, we are speaking of nothing but people, then we can limit the domain to the set of people. If we do, then a universal quantifier will have the effect of for all x in the set of people, i.e., for all people x. The universal quantifier, in such a case, represents the English expressions anybody and everybody without using a predicate meaning person. Adjectives Adjectives, words such as good, red, friendly and logical, modify nouns. With a few exceptions, they translate into Q as monadic predicates, linked to the predicates representing the nouns they modify by conjunctions. Thus 67. All friendly cats purr becomes 68. Vx«Fx&Cx)
->
Px),
and 69. Some artists are unhappy people becomes
Quantifiers
93
70. :Jx(Ax&(Ux&Px)).
Notice that these sentences are basically categorical sentence forms. Adjectival phrases, consisting of an adjective modified by an adverb, for example, function in the same way. They must be treated as a single unit. Thus 71. John is a very wealthy logician translates into Q as 72. Wj&Lj,
where 'W' represents very wealthy. Most adjectives and adjectival phrases thus translate as conjunctions. The set of colorless gases is the set of things that are both colorless and gases. It is thus the intersection of the set of colorless things and the set of gases. For this reason, we'll call adjectives that work in the standard way intersective. Certain adjectives, however, do not translate into Q directly. They have meanings that relate in some way to the nouns they modify, so they and their nouns must translate as a single unit. Luckily there is an easy test for identifying these adjectives, which we'll call nonintersective. Wealthy logicians are both wealthy and logicians; red Chevrolets are both red and Chevrolets. But alleged criminals are not alleged and criminals. Good pianists are not simply pianists and good. Former Congressmen are not former and Congressmen; large mice are not large and mice. Good pianist means something like good as a pianist; large mouse, something like large for a mouse. This is why the following arguments fail: 73.
(a) Every pianist is a lover. :. Every good pianist is a good lover. (b) All mice are animals. :. All large mice are large animals.
To avoid trouble, then, nonintersective adjectives modifying nouns must translate as single monadic predicates. Note, however, that there is an important difference among nonintersective adjectives. The set of large mice is a subset of the set of mice; the set of good pianists is similarly a subset of the set of pianists. Nonintersective adjectives such as large and good are therefore called subsective. Other nonintersective adjectives, such as fake and former, are not even subsective. The set of fake diamonds is not a subset of the set of diamonds at all. Likewise, the set of former Congressmen is not a subset of the set of Congressmen. Nonintersective, nonsubsective adjectives must translate, together with the nouns they modify, as a single unit. Nonintersective but subsective adjectives may translate as conjunctions, but with an important difference from intersective adjectives. We can render large mouse asLx&M x where M represents mouse, but only if we construe L as translating, not large, but large for a mouse.
Relative Clauses Relative clauses are English expressions formed from sentences. They begin, generally, with that or a word starting with wh-, such as who, which, when or where, though these words are often omitted. Relative clauses frequently act like adjectives, modifying nouns or noun phrases. Thus that I used to attend, who once denounced Richard Nixon, and (when) I've placed my hopes in something behave as relative clauses in the sentences 74.
(a) A school that I used to attend has been closed. (b) Senator McCarthy, who once denounced Richard Nixon, is now retired.
Logic, Sets and Functions
94
(c) Every time (when) I've placed my hopes in something I've been disappointed. Like intersective adjectives, relative clauses are conjoined to the nouns they modify. They themselves, however, translate as open formulas rather than monadic predicates. That I used to attend, for example, derives from the open sentence I used to attend x. Similarly, who once denounced Richard Nixon derives from x once denounced Richard Nixon, and (when) I've placed my hopes in something, from I've place my hopes in something at (time) x. Using the obvious representations, we can symbolize these as 'Aix', 'Dxn', and '~yHiyx'. Conjoining them in the appropriate manner to representations of their nouns, we obtain: 75.
(a)
~x((Sx&Aix)&Cx)
(there is an x such that x is a school, I used to attend x, and x is
closed) (b) Dmn&Rm (Senator McCarthy once denounced Richard Nixon, and McCarthy is now retired) (c)
\lx((Tx&~yHiyx) ---+
Dix) (for every x, if x is a time and, for some y, I've placed my hopes in y at x, then I've been disappointed at x)
Relative clauses, in general, translate quite easily. Only one minor wrinkle ruins their simplicity. Some relative clauses restrict the group of things the noun phrase they modify applies to. If I tell you that everyone I know prefers Mexican to Chinese food, then I am speaking, not of everyone, but just of everyone I know. (74a.) and (74c.) contain such restrictive relative clauses. Other clauses, however, make almost parenthetical comments about their nouns or noun phrases. (74b.) contains such an appositive relative clause. Most relative clauses in actual discourse are restrictive. To tell whether a given clause is restrictive or appositive, ask whether the clause is helping to specify what the sentence is talking about or providing additional information concerning an already determinate topic. English does offer two linguistic hints. First, that often signals that a clause is restrictive; which, with some exceptions (for example, the phrases in which and with which), often signals an appositive. Relative clauses often begin with other wh-words, however, or with no special word at all. In these cases, there are no signals. Furthermore, the use of that and which is not very firmly established; these words are unreliable guides. Second, and more reliably, commas often do, and always can, set appositive clauses off from the rest of the sentence. Restrictives, on the other hand, reject commas in this role. So virtually all relative clauses set off by commas are appositive. For those not set off by commas, there is a simple test; try inserting commas. If the result sounds acceptable, the clause is probably appositive. Otherwise, it is restrictive. Restrictives and appositives, in symbolic representations, both connect to the remainder of the formula by conjunction. Most of the time, therefore, it makes no difference to the translation whether a given clause is restrictive or appositive. When universal quantifiers are involved, however, and the clause modifies the subject noun phrase, it does matter. Consider these sentences: 76.
(a) All the Democratic candidates for President, who are already campaigning, support labor unions. (b) All the Democratic candidates for President who are already campaigning support labor unions.
The only difference between them is the pair of commas setting off the relative clause in (76a.). In that sentence, the clause is clearly appositive. It asserts that all the Democratic candidates for President support labor unions, and remarks, on the side, as it were, that all those candidates are already campaigning. (76b.), in contrast, does not claim that all the Democratic candidates support
95
Quantifiers
labor unions; it asserts only that all those who are already campaigning do so. (76b.) is thus a weaker contention than (76a.). To translate these sentences, first translate the relative clause. Who are already campaigning derives from x is(are) already campaigning, which we can write as 'Cx'. Adopting the obvious representations (and letting 'Lx' correspond to x supports labor unions), then, we can symbolize the sentences in (76) as 77.
(a) Yx(Dx
->
Lx)&Yx(Dx
(b) Yx((Dx&Cx)
->
->
Cx)
Lx).
Note that the restrictive clause is conjoined, in effect, to the rest of the subject; the appositive clause, to the rest of the entire sentence. Prepositional Phrases
Prepositions are rather ordinary English words such as in, to, of, about, up, over, from, and so on. They combine with noun phrases to form prepositional phrases, which act as either adjectives or adverbs: thus up a creek, from Pennsylvania, and in the middle of Three Chopt Road. We'll discuss those acting as adverbs, which translate together with the verbs or adjectives they modify as single units, in a few pages. Here we'll talk about prepositional phrases modifying nouns, which have separate translations. In prepositional phrases that function more or less as adjectives, prepositions relate two noun phrases. They thus translate into Q as dyadic predicates. The representatives of prepositional phrases themselves connect to the symbolizations of the noun phrases they modify by conjunction. Consider these examples: 78.
(a) Everyone from Pittsburgh loves the Steelers. (b) If I don't meet you, I'll be in some jail.
(78a.) contains the prepositional phrase from Pittsburgh. Since from translates into a dyadic predicate, say 'F' (and since the Steelers here functions as the proper name of a team), (78a.) becomes
79. Yx((Px&Fxp)
->
Lxs).
(78b.) contains the prepositional phrase in some jail, which itself contains a determiner. The conjunction of prepositional phraSe to noun phrase, then, occurs within the scope of a quantifier:
80. -,Hab
->
3x(Jx&Iax)
(where 'a' and 'b' symbolize I and you, respectively) symbolizes (78b.). Prepositional phrases modifying nouns thus translate readily into quantification theory.
4.5.2
Verb Phrases
So far we've discussed how noun phrases and their modifiers translate into quantification theory. Since sentences consist of a subject noun phrase and a verb phrase, however, we also need to explain how to symbolize verb phrases and their modifiers in Q. In any verb phrase, of course, there is a verb. Verbs fall into several categories, depending on their ability to take certain kinds of objects. Some verbs are intransitive; they cannot take objects at all. Fall, walk and die are all intransitive. Transitive verbs take noun phrases as direct objects. Examples are throw, win and send. Some of these, such as give, also take noun phrases as indirect
96
Logic, Sets and Functions
objects. Other verbs take sentences, or grammatical constructions closely related to sentences, as objects. Believe, know and persuade are such clausally complemented verbs. The logic of verbs taking sentential complements remains the subject of much debate. Here, therefore, we'll consider only transitive and intransitive verbs. Note that many verbs fall into more than one category. Eat, for example, can have a noun phrase object (in, for example, We eat spaghetti every Wednesday night), but does not need one (Let's eat out). Believe can take a sentence (I believe that God exists) or a noun phrase (I believed him). Intransitive verbs translate into Q as monadic predicates. John walks, for instance, becomes 'Wj'; everyone who doesn't own a car walks becomes
81. 'ix((Px&,3y(Cy&Oxy))
--7
Wx).
Transitive verbs translate into Q as polyadic predicates. Usually, they become dyadic predicates; 'LmJ' represents Mary loves Fred, etc. Occasionally, however, a verb relates more than two noun phrases. Mike gave John War and Peace, for example, translates as 'Gmjp'. In general, predicates of more than two places prove very useful in symbolizing sentences with indirect objects or adverbial modifiers of certain kinds. This, however, raises the general issue of adverbs.
Adverbial Modifiers Adverbs, such as quickly, well, anytime and somewhere, modify verbs. They specify how, when or where a certain condition holds or a certain activity occurs. Unfortunately, most adverbs have no direct symbolizations in quantificationallogic. Q must represent them, together with the verbs they modify, as single units; expressions such as walks slowly or plays well become predicates such as 'W' or 'P'. Though some logicians have attempted to devise schemes for representing adverbs in Q, none has managed to find a way of translating all adverbs accurately. Some adverbs, however, do translate into quantificationallogic. We'll call always, anytime, whenever, wherever, anywhere, sometime, etc., adverbs of quantification. Consider the sentence I like Alfred sometimes. I like Alfred, normally, would become 'Lia'. So how do we represent 'Lia, sometimes'? Some is a determiner. So the sentence is saying, in effect, that, for some times x, I like Alfred is true at x. Instead of 'Lia', then, we need 'Liax', meaning I like Alfred at x. I like Alfred sometimes thus becomes
82. 3x(Tx&Liax).
)
Similarly, consider Everywhere I look there are timeshare resorts. This, in essence, amounts to for every x, if x is a place and I look at x then there are timeshare resorts at x, or, in symbolic notation,
83. 'ix((Px&Lix) --73y(Ry&Ayx)). Some adverbs of quantification, however, have no correlates in Q. Frequently, which amounts roughly to at many times, and rarely or seldom, which amount roughly to at few times, would translate into Q only if Q had a way of symbolizing many and few. Since quantification theory represents only a few determiners, it can represent only a few adverbs of quantification. Prepositional phrases, as we've seen, can modify nouns. They can also modify verbs. John ran down the street, We're singing in the rain and I'll have a hot dog on a paper plate all contain prepositional phrases functioning adverbially. Just as adverbs, in most cases, do not translate into Q except as parts of verb phrases that become predicates, so prepositions linking noun phrases to verbs or verb phrases translate together with the modified verb or verb phrase. They do not become dyadic predicates in their own right, as they do when modifying noun phrases. Nevertheless, because
l
Quantifiers
97
prepositional phrases contain noun phrases, their symbolic representations are more interesting than those of adverbs. Think about a sentence such as Laura lives on East 72nd Street. This becomes, when symbolized, 'Lle', where "e' represents East 72nd Street. Lives on translates as a single dyadic predicate. Note that we cannot apply the strategy appropriate to adjectival prepositional phrases; the above sentence is not equivalent to Laura lives and is on East 72nd Street. To take a more complex example, Richard has worked in every division of Reynolds Metals Company contains an adverbial prepositional phrase that itself contains a determiner. Work here is intransitive. It would usually translate into a monadic predicate. But in is a preposition that can combine with the verb for purposes of translation. Instead of using the simple open sentence x works, therefore, we can use x works in y, a dyadic predicate, to obtain 84. 'v'x(Dxc
-->
Wrx).
Notice that of in this sentence does not add another place to the predicate; of Reynolds Metals Company functions adjectivally, modifying division.
4.5.3
Connectives
Quantification theory includes sentential logic. Sentential connectives can link quantified sentences together; they can even inhabit noun and verb phrases. We've discussed some problems that this creates in Chapter 2. Here, we'll recapitulate them briefly and extend them to some more interesting cases. Recall that noun and verb phrases can be joined together by and, or and if not. Chapter 2 recommended a policy of splitting such phrases. Connectives linking noun or verb phrases usually accept transformation into connectives linking sentences. Thus Abraham Lincoln and Calvin Coolidge were Republican Presidents amounts to Abraham Lincoln was a Republican President, and Calvin Coolidge was a Republican President. Similarly, Fred likes hot dogs and hamburgers amounts to Fred likes hot dogs and Fred likes hamburgers. In quantification theory this advice becomes more important in many cases. All lions and tigers are cats is equivalent to All lions are cats and all tigers are cats. But the conjoined noun phrase can tempt us into a translation 85. 'v'x((Lx&Tx)
-->
Cx),
which says that everything that is both a lion and a tiger is a cat. Of course, nothing is both a lion and a tiger. So this symbolization is incorrect. Separating sentences results in the formula 86. 'v'x(Lx
-->
Cx)&'v'x(Tx
-->
Cx),
which captures the meaning of the original. When existential quantifiers are involved, or when the connectives are in the verb phrase, splitting makes little difference. But, in subject noun phrases, it is vital. As in the case of sentential logic, however, we must take care to split only those sentences for which the process preserves meaning. Harry loves knockwurst and sauerkraut may not be equivalent to Harry loves knockwurst and Harry loves sauerkraut; he may love the combination without being very excited about the individual components, or vice versa. Mary and Susan own the entire company is probably not saying that Mary owns the entire company, and that Susan does too, but that they own the entire company between them. Arguments relying on these special conjoined noun or verb phrases lie outside the bounds of quantification theory. Another problem pertains to connectives such as if, only if, etc. They cannot join together two noun phrases or two verb phrases, but they can appear within sentences in ways that do not reduce to simple sentential connection. Consider:
98
Logic, Sets and Functions
87. A formula is contingent only if it's not valid. We might think of this as a sentence with a formula as subject noun phrase and is contingent only if it's not valid as main verb phrase. The determiner a is clearly functioning generically, so it translates as a universal quantifier. The common noun formula appears as a monadic predicate. (87) thus looks like a complex version of a universal affirmative sentence form. Its symbolization begins with 'Vx(Fx -+ ... '. The main verb phrase contains two connectives, only if and not; the adjectives contingent and valid appear as monadic predicates. It acts much like a variable. So the symbolization of (87) turns out to be
88. Vx(Fx
-+
(Cx
-+
-..vx».
Alternatively, we might think of (87) as containing a connective, only if, joining together two sentences, a formula is contingent and it's not valid. On this approach, (87) resembles (62). It amounts to If a formula is contingent, it's not valid. This we might be tempted to translate as
89. 3x(Fx&Cx)
-+
-..vx ,
but, in (89), the final occurrence of 'x' is not in the scope of the existential quantifier. So this is not even a formula. Once again, the solution requires using a universal quantifier with the entire formula as its scope:
90. Vx((Fx&Cx)
-+
,Vx).
This, too, is an acceptable translation of (87). Fortunately, it is equivalent to (88). So the ways of construing the sentence's structure yield equivalent results. Naturally, connectives can also join together entire sentences that have no troublesome links between them: If we don't hang together, we'll surely all hang separately, Some political parties die out after a short time, but others last for centuries, and Unless everyone leaves, I'll refuse to come out all work as we might expect from sentential logic. Finally, quantification theory contains not only connectives but sentence letters. It might seem that any sentence can translate into Q by using just predicates, constants, variables and quantifiers. But a few, very simple sentences- it is raining and it's three 0 'clock, for exampl€r- resist this analysis. It, in these sentences, does not stand for an object, so it would be very odd to translate it is raining as, say, 'Ri'. To see this, try asking, "What is raining?" The question doesn't make very good sense. So sentences containing it in what linguists call its pleonastic use seem to be best translated with simple sentence letters.
Problems Translate the following sentences into Q, exposing as much structure as possible. If any translate only with difficulty, explain why. 1. All men are born good. (Confucius)
2. All that I know, I learned after I was thirty. (Georges Clemenceau) 3. Children are always cruel. (Samuel Johnson) 4. All who remember, doubt. (Theordore Roethke) 5. All big men are dreamers. (Woodrow Wilson) 6. Only the shallow know themselves. (Oscar Wilde)
I
Quantifiers
99
7. All are not friends that speak us fair. (James Clarke)
8. All's Well That Ends Well. (William Shakespeare) 9. No sound is dissonant which tells of life. (Samuel Taylor Coleridge) 10. If any would not work, neither should he eat. (II Thessalonians 3:10) 11. There is no detail that is too small. (George Allen) 12. Alas! It is delusion all. ... (George Gordon, Lord Byron) 13. . . . and now nothing will be restrained from them, which they have imagined to do. (Genesis 11:6) 14. All finite things reveal infinitude. (Theodore Roethke) 15. Poets are the unacknowledged legislators of the world. (Percy Bysshe Shelly) 16. . . . we are dust and dreams. (A. E. Housman) 17. Everything that man esteems endures a moment or a day. (William Butler Yeats) 18. To be beloved is all I need, and whom I love, I love indeed. (Samuel Taylor Coleridge) 19. Hope is a delusion; no hand can grasp a wave or a shadow. (Victor Hugo) 20. . . . the things which are seen are temporal; but the things which are not seen are eternal. (II Corinthians 4:18) 21. Some people with great virtues are disagreeable while others with great vices are delightful. (La Rochefoucauld) 22. Nothing which is true or beautiful or good makes complete sense in any immediate context of history. . . . (Reinhold Neibuhr) 23. They also live who swerve and vanish in the river. (Archibald MacLeish) 24. Nothing is done. Everything in the world remains to be done or done over. (Lincoln Steffens) 25. So then neither is he that planeth any thing, neither he that watereth; but God that giveth the increase. (I Corinthians 3:7) 26. Loafing needs no explanation and is its own excuse. (Christopher Morley) 27. Any mental activity is easy if it need not take reality into account. (Marcel Proust) 28. . . . it is not poetry, if it make no appeal to our passions or our imagination. (Samuel Taylor Coleridge) 29. When a man is wrong and won't admit it, he always get angry. (Thomas Haliburton) 30. My only books were women's looks, and folly's all they've taught me. (Thomas Moore) 31. All things fall and are built again, and those that build them again are gay. (William Butler Yeats) 32. All men have aimed at, found, and lost. . . . (William Butler Yeats)
Logic, Sets and Functions
100
33. Great is the hand that holds dominion over man by a scribbled name. (Dylan Thomas) 34. He that stays in the valley shall never get over the hill. (Jonn Ray) 35. It is always the secure who are humble. (G. K. Chesterton) 36.. Every country can produce good men. (Gotthold Lessing) 37. A dull axe never loves grindstones. (Henry Ward Beecher) 38. To whom nothing is given, of him nothing can be required. (Henry Fielding) 39. There has never been any 30-hour week for men who had anything to do. (Charles F. Kettering) 40. A thing of beauty is a joy forever. ... (John Keats) 41. Work is a grand cure of all the maladies that ever beset mankind. (Thomas Carlyle) 42. He that has no patience has nothing at all. (Italian proverb) 43. Every man without passions has within him no principle of action, no motive to act. (Claude Adrien Helvetius) 44. You can't have a better tomorrow if you are thinking about yesterday all the time. (Charles F. Kettering) 45. Nothing will ever be attempted if all possible objections must be first overcome. (Jules W. Lederer) 46. There is a singer everyone has heard. . . . (Robert Frost) 47. All man's friend, no man's friend. (John Wodroephe) 48. To do nothing is in every man's power. (Samuel Johnson) 49. Nobody ever did anything very foolish except from some strong principle. (William Lamb) 50. We receive only what we give. (Samuel Taylor Coleridge) 51. If you build a castle in the air, you won't need a mortgage. (Philip Lazarus) 52. In every work of genius we recognize our own rejected thoughts. (Ralph Waldo Emerson) 53. Nothing is more boring than a man with a career. (Aleksandr Solzhenitsyn) 54. . . . the Bears were good Bears, who did nobody any harm, and never suspected that anyone would harm them. (Robert Southey) (Use a predicate 'B' for is one of the Bears.)
4.6
Interpretations
First, we need to develop a precise notion of what an interpretation in quantificational logic is. Recall that in sentential logic an interpretation is simply an assignment of truth values to atomic formulas, i.e., to sentence letters. Quantificational logic includes sentential logic, so interpretations within it will incorporate such truth value assignments. But quantificational interpretations are more complex.
Quantifiers
101
Definition 4.1 An interpretation M of a set S of formulas of quantificationallogic is an ordered pair (D, ¢), where D (M's domain, or universe of discourse) is a nonempty set and ¢ is a function assigning (1) truth values to sentence letters in S, (2) elements of D to constants in S, and (3) sets of n-tuples of elements of D to n-ary predicates in S. An interpretation thus has two components. The first is a set that specifies what objects the formulas in question are talking about. The quantifiers range over this set, in the sense that we construe "for all x and "for some y" as meaning "for all x in D" and "for some y in D" or, in other words, "for all elements of D" and 'for some element of D." The last section already introduced the notion of a universe of discourse in the context of translation. If we are speaking about nothing but people, or automobiles, or income groups, then we can avoid populating all our formulas with a predicate representing this subject matter by taking the set of people, or of automobiles, or of income groups as the universe of discourse. We thereby count "for all x" as having the significance of, for example, 'for all people x". Specifying a domain can thus simplify translations. In interpreting quantified formulas, however, specifying a domain is a necessity. We can either say that the domain is the set of people, or income groups, or whatever, or enumerate the members: for example, {Tom, Dick Harry}. The second component of an interpretation M is an interpretation function. This function, in effect, assigns meaning to the constants, predicates and sentence letters in the formulas we're interpreting. It assigns truth values to sentence letters, telling us whether the sentences they represent are true or false. It assigns elements of the domain to constants, telling us which objects they stand for. Finally, it assigns sets of n-tuples of objects to n- ary predicates. Consider a unary or monadic predicate, 'R', which informally means red. The interpretation function assigns 'R' a set of I-tuples. Intuitively, it tells us which objects satisfy 'R'; it tells us, in other words, which objects are red. The function assigns to a binary or dyadic predicate such as 'L', meaning loves, a set of ordered pairs. The function tells us, then, who loves who. If ¢(L) = {(Bob, Carol), (Carol, Ted), (Ted, Alice), (Alice, Bob)}, then Bob loves Carol, Carol loves Ted, Ted loves Alice, and Alice loves Bob. This table summarizes how the interpretation function works. Symbol . Interpretation truth value sentence letter constant object in the domain n-ary predicate set of n-tuples of objects in domain The definition of an interpretation only performs part of the task that we need to accomplish. We want to be able to produce interpretations that make various formulas true and others false. To do this, we need to know how to evaluate the truth value of a formula on an interpretation. It's easy to judge the truth value of a sentence letter on an interpretation; just see what value the interpretation function assigns to it. The interpretation M = (D, ¢rangle makes a sentence letter P true just in case ¢ assigns truth to P: (1)[P]M
=T
{::} ¢(P)
= T.
Other atomic formulas, consisting of an n-ary predicate followed by nconstants, are also easy to evaluate. The sentence Bob loves Carol is true if and only if Bob loves Carol. The sentence is true, in other words, just in case the interpretation we assign to the predicate loves includes the pair (Bob, Carol). In general, then, an atomic formula of the form Ral ... an will be true on an interpretation Mjust in case the set the interpretation function assigns to R includes the n-tuple consisting of the objects that al, . .. ,an stand for:
Logic, Sets and Functions
102
(2)[Ral ... anlM = T
¢:}
(¢(at), ... , ¢(an )) E ¢(R).
Assessing the truth values of formulas with sentential connectives as main connectives is also easy, provided that we know the truth values of the components. We can proceed exactly as in sentential logic: • [-AIM
=T
¢:}
[AIM = F.
• [(A&B)]M = T • [(A V B)]M
¢:}
=T
¢:}
[A]M = [B]M = T. [A]M = Tor [B]M
= T.
• [(A
-+
B)]M
=T
¢:}
[AIM
= For [B]M = T.
• [(A
<-+
B)IM
=T
¢:}
[AIM
= [BIM.
The real task is in defining the truth value of quantified formulas. Consider an existentially quantified formula, say, '3xFx. This formula will be true if the predicate 'F' is true of some object in the domain. We can think of ourselves as considering an instance of the formula, such as 'Fa', and asking whether 'a' could name something that would make 'Fa' true. Similarly, consider a universally quantified formula, say, 'Vx(Fx -+ Gx)'. We can think of an instance of this formula, such as 'Fa -+ Ga'; this instance should be true no matter what object in the domain 'a' stands for, if the original formula is true. So, to judge the truth value of a quantified formula on an interpretation M, we can look at the truth values of an instance of that formula on interpretations which are just like M, except that they may assign different objects from the domain to the constant substituted for the quantified variable. (Of course, M may assign no object to this constant, if the constant doesn't appear anywhere in the formulas that M is interpreting. In that case, we'll be looking at interpretations that differ from M by extending it to cover the constant.) • [3vA]M = T ¢:} [Ac/x]w = Tfor some constant c and interpretation M' differing from M at most in assigning a different element of D to c. • [VVA]M = T ¢:} [AC/X]M' = T for every constant most in assigning a different element of D to c.
C
and interpretation M' differing from M at
Another way of thinking about the truth values of quantified formulas is this. A universally quantified formula VvA is true just in case A would be true no matter what object in the universe of discourse v would stand for. An existentially quantified formula 3vAwouid similarly be true if there's an object in the domain such that, if v stood for it, A would be true. To see how this works in practice, let's consider the formulas 'VxFnx' and 'Vx3y(Fxy&Gyx)'. Are these formulas true or false on the following interpretation M = (D, ¢)?
• D = { New York, Los Angeles, Chicago} • ¢(n) = New York • ¢(F) = {( New York, Los Angeles), (New York, Chicago), (Los Angeles, Chicago)} • ¢(G)
= {( Los Angeles, Chicago),
(Los Angeles, New York), (Chicago, New York)}
Quantifiers
103
The domain here consists of the three largest U.S. cities. We can think of 'F' as meaning has a larger population thanand 'G' as meaning is west of Informally, then, the first formula we're analyzing means that New York is more populous than each city in the set. Is this true? The instance 'Fna' of this formula should be true no matter which city 'a' stands for. So we should find, in the interpretation of 'F', all possible pairs of the form (New York,_). Do we? Included in 4>(F) . are ( New York, Chicago) and ( New York, Los Angeles ), but not ( New York, New York). The formula is therefore false; New York is not larger than itself. The second formula says that, for each of the three largest U.S. cities, there is another that is both smaller than and west of it. This should obviously be false; no city in the set is west of Los Angeles. If we analyze the formula according to our definitions, we can see that the instance '3y(Fay&Gya)' should be true no matter which city 'a' stands for. So suppose 'a' stands for Los Angeles. We should be able to find another city which, as a referent for 'b, would make 'Fab&Gba' true. Can we? First let's contemplate the' Fab' part. Does the interpretation of 'F' contain a pair of the form ( Los Angeles, _)? Yes: ¢>(F) contains the pair ( Los Angeles, Chicago). Furthermore, this is the only pair of that form in ¢>(F). Moving to the 'Gba' part, does ¢>(G) contain the inverse pair ( Chicago, Los Angeles) ? The answer is no. So we can't find another city that would stand in the proper relations to Los Angeles. But the universal formula would be true only if we could do this for every city in the universe of discourse, so the formula as a whole is false on M. Notice, incidentally, that we couldn't find a city with the proper relations to Chicago either, because no city in the set is smaller than Chicago; ¢>( F) contains no pair of the form ( Chicago, _). We could, however, find a city with the proper relations to New York. We find both ( New York, Chicago) and ( New York, Los Angeles) in ¢>(F), and the inverses of both pairs, (Chicago, New York) and (Los Angeles, New York), occupy ¢>(G). This reflects the fact Chicago and Los Angeles are both smaller than and west of New York. It also indicates that, if the initial quantifier of the formula were existential rather than universal, the formula would be true on M.
Problems Let D = {a}, 4>(a) interpretation?
a, and ¢>(F)
{(a)}. What is the truth value of these formulas on this
1. Fa
2. -.Fa 3. Fa&Fa 4. Fa
V
-.Fa
5. Fa
->
6. negFa
-.Fa ->
Fa
->
3yFy)
7.~F:
8. 'r/xFx 9. Vx(Fx
10. Vx(-,Fx
->
-,3yFy)
Let D = {a}, ¢>(a) = a, ¢>(F) this interpretation? 11. Fa
= 0, and
¢>(G)
= {(a)}.
What is the truth value of these formulas on
Logic, Sets and Functions
104 12. Ga
13. Fa
~
Ga
14. Ga
~
Fa
15. 3xFx 16. VxFx 17. 3xGx 18. VxGx 19. Vx(Fx ~ Gx) 20. forallx«Fx V Gx) ~ Gx)
Let D = {a}, ¢(a) = a, ¢(F) on this interpretation?
= {(a)}, and ¢(R)
= {(a, a)}. What is the truth value of these formulas
21. Raa 22. Vx(Fx
~
Rxx)
23. VxRxx 24. VxRax
25. Vx(Rax&Rxa) 26. VxVy(Rxy ~ Ryx)
27. Vx(Fx ~ 3y(Fy&Rxy)) 28. VxVyVz«Rxy&Ryz) 29. VxVy(Rxy
-7
-7
Rxz)
3z(Rxz&Rzy))
30. 3x(Fx&Rxx)
f-->
Let D = {a,b},¢(a) this interpretation?
Vx(Fx&Rxx)
= a,¢(b) = b,
and ¢(F) = {(b)}. What is the truth value of these formulas on
31. Fa 32. Fb
33. 3xFx 34. VxFx 35. existsx-,Fx 36. Vx-,Fx 37. 3xFx&3x-,Fx 38. 3x(Fx&-.Fx)
'\
Quantifiers
105
39. 3x(Fx
->
VyFy)
40. Vx(Fx
->
VyFy)
Let D = {a,b},¢(a) = a,¢(b) = b, and ¢(R) formulas on this interpretation?
=
{(a,a),(b,a)}. What is the truth value of these
41. Rab 42. Rba 43. Rbb 44. 3xRax 45. 3xRbx 46. 3xRxa 47. 3xRxb 48. VxRxx 49. VxVy(Rxy
->
Ryx)
50. Vx3yRxy
Let D = {a,b},¢(a) = a,¢(b) = b,¢(F) = {(b)}, and ¢(R) of these formulas on this interpretation?
= {(b,b), (b,a)}.
What is the truth value
51. 3x(Fx&Rxx)
52. existsx(Fx&Rxa) 53. Vx(Fx
->
Rxx)
54. Vx(Fx
->
Rax)
55. VxVy(Rxy
->
Ryx)
56. VxVy(Rxy ---; Rxx) 57. VxVy(Rxy
->
Ryy)
58. VxVyVz((Rxy&Ryz)
->
Rxz)
59. VxVyVz((Rxy&Ryz)
->
-.Rxz)
60. VxVy((Fx&Fy)
-+
(Rxy&Ryx»
Let D = {a,b},¢(a) = a,¢(b) = b,¢(R) = {(a,a),(b,a)},¢(S) = {(b,b),(a,b)}, and ¢(F) = {(a)}. What is the truth value of these formulas on this interpretation? 61. Vx(Fx
-+
62. Vx3yRxy 63. Vx3ySxy
Rxx)
106
Logic, Sets and Functions
64. 3xYyRxy 65. 3xYySxy 66. 3xYySyx 67. 3xYyRyx 68. YxVy(Rxy V Sxy) 69. VxYy(Rxy
-+
-.sxy)
70. YxYy(Rxy
-+
(Fx V Fy))
Let D = {a, b}, ¢(a) = a,¢(b) = b,¢(S) these formulas on this interpretation?
= 0 and
¢(R)
= {(a,a), (b,b)}.
What is the truth value of
71. 3xRxx 72. 3xSxx 73. VxRxa 74. 3xRxa 75. Vx(Rxb
Sxb)
-->
76. YxYy(Rxy
-+
Sxy)
77. VxVy(Sxy
-+
Rxy)
78. VxYy((Rxy&Sxy)
-->
79. VxYy((Rxy&-.sxy) 80. YxYy((Rxy&Ryx)
Syx)
-+
-->
Ryx)
3z(Rxz V Sxz))
Let D = {a,b,c},¢(a) = a,¢(b) = b,¢(F) = {(b), (en, and ¢(R) truth value of these formulas on this interpretation?
81. 3xRxx 82. Vx(Fx
-->
3yRxy)
83. 3x(Fx&YyRxy) 84. 3x(Fx&YyRyx) 85. Vx(Rxx
-+
Fx)
86. Vx3yRxy 87. 3xYyRyx 88. (Fa V Fb)
-+
(Raa V Rab)
89. Yy(3xRxy
-+
3zRyz)
90. Vy(3zRyz
-+
3xRxy)
= {(a, c), (b,c), (e,e)}.
What is the
Quantifiers
107
Let D = {O, 1, 2, ... , 1O}, with '<' and ':S' having their usual interpretations. What truth values do these formulas have on this interpretation? 91. Vxx
<x
92. Vxx
:S x
93. VxVy(x
--+
-'y
< x)
94. :Jx:Jy(x:S y&y ::; x)
< y&y :S x)
95. :Jx:Jy(x
96. VxVyVz((x
< y&y < z)
--+
x
< z)
97. VxVyVz((x
:S y&y :S z)
--+
x ::; z)
9S. VxVyVz((x
< y&y :S z)
-+
x
< z)
--+
:Jz(x
100. VxVy(x ::; y
--+
:Jz(x ::; z&z ::; y))
99. VxVy(x
< z&z < y))
101. Vx:Jyx < y 102. Vx:Jyx
:S y
103. Vx3yy
<x
104. :JyVxx
105. :JyVxy::; x
Let D = {O, 1,2, ... }, with '<' and I::; , having their usual interpretations. What truth values do these formulas have on this interpretation? 106. Vxx
<x
107. Vxx
:S x
lOS. VxVy(x
--+
-'y
< x)
109. :Jx:Jy(x:S y&y :S x) 110. :Jx3y(x
< y&y :S x)
111. VxVyVz((x < y&y < z)
--+
x
< z)
112. VxVyVz((x
:S y&y :S z)
--+
x::; z)
113. VxVyVz((x
< y&y :S z)
-+
x
< z)
114. VxVy(x
--+
3z(x
< z&z < y))
115. VxVy(x
:S y
-+
3z(x
:S z&z :S y))
116. Vx3yx
117. Vx3yx
:S y
118. Vx3yy
<x
119. 3yVxx
120. 3yVxy::; x
Chapter 5
Quantified Natural Deduction The system of natural deduction for sentential logic extends easily to quantificational logic. All the rules of sentential deduction will apply in quantification theory. But, to deal with quantifiers, the system adds three new rules and a new method of proof. The deduction system that emerges shares the virtues of its sentential cousin. It is sound, for every provable formula is valid, and every conclusion that can be proved from a set of premises is implied by them. The system is also complete, for every valid formula of quantification theory can be proved in the system, and every valid argument can be shown to be valid within it. Furthermore, the system mirrors very closely the processes of reasoning that people use in a wide variety of contexts. It reflects particularly well reasoning in mathematics and related disciplines.
5.1
Deduction Rules for Quantifiers
The deduction rules needed for quantificational logic are very straightforward. Recall that A[c/x] is the result of substituting c for every occurrence of x throughout the formula A. If "ivA and 3vA are formulas, then A[c/v] is called an instance of them. Conversely, VvA and 3vA are geneTics of A[c/v]. Each connective has two rules. One introduces it into proofs, while the other allows us to exploit its presence. The existential quantifier similarly comes with two rules: an introduction rule and an exploitation rule. The introduction rule, in essence, allows us to move to an existentially quantified formula from any instance of that formula. 1 The rule, often called existential generalization, thus takes the form: 3I(Existential Introduction)
n.A[c/v] n+p.3vA 3I,n Here c may be any constant. Existential Introduction allows us to infer an existentially quantified formula from any instance of it. It sanctions the step from an instance to its corresponding existential generic. Suppose that our universe of discourse consists entirely of people. If Jones, for example, is a spy, then we may conclude that someone is a spy. If Susan suspects Harry, then Susan suspects someone; of course, it's IThis rule was first formulated by the English philosopher William of Ockham in the fourteenth century.
109
Logic, Sets and Functions
110
also true that somebody suspects Harry and that somebody suspects somebody. Finally, if Frank doesn't trust himself to work around large sums of money, then Frank doesn't trust somebody to work around large sums of money. Additionally, somebody doesn't trust Frank to work around large sums, and someone doesn't trust himself to do so. Each of the following is an acceptable application of existential introduction. Premise Fa Gab Gab Hcc Hcc Hcc 3xFxa VxFx --> Gb
Conclusion 3xFx 3xGax 3xGxb 3xHcx 3xHcx 3xHxx 3y3xFxy 3z(VxFx --> Gz)
In each case, the premise is an instance of the conclusion. The rule of existential exploitation allows us to move from an existentially quantified formula to an instance of it. 2 It is almost exactly the reverse, then, of the existential introduction rule. But it does impose a restriction: the instance must involve a constant new to the proof. The rule says that we may drop an existential quantifier serving as a main connective in a formula, and substitute for the quantified variable a constant that hasn't appeared earlier in the proof. The constant must have appeared nowhere in the deduction, not even in a Show line. (Actually, no harm would result from allowing us to use, for 3E, constants that appear earlier only on already bracketed lines. But, to minimize confusion, we'll always use completely new constants.) 3E(Existential Exploitation)
n.3vA n + p.A[c/v] 3E, n Here c must be new to the proof. Suppose that we have the information that someone in our department is selling trade secrets to a competitor. We don't know who this person is- or, perhaps, who these people are- but we do want to reason from what we know in order to find out. We know that at least one person has been selling secrets; our reasoning and our communication will proceed much more readily if we give this person some name- John Doe, say, or just the mole - so that we can refer to him or her in various contexts. We can't simply say, "Someone has been selling our trade secrets. Someone must have joined the department around the middle of 1981, because that's when secrets began to leak." Nothing here indicates that the two "someones" are the same. To tie these assertions to the same individual, we must have a way of referring to that person. Introducing a name accomplishes this. It's critical that the name we choose be new. If Sarah Freeland is the head of the department, and we decide to call the seller of trade secrets Sarah Freeland, then utter confusion will result. The system of rules in this chapter is sound in the sense that the rules never lead us astray; they never allow us to prove a formula that isn't valid, or permit us to establish the validity of an invalid argument form. Nevertheless, the existential exploitation rule seems unsound; it justifies the inference from 3vA to A[c/v], so long as c is new to the proof. Our demand for a new constant 2The American philosopher W. V. Quine first formulated existential exploitation in this way in 1950, in his Methods of Logic (Cambridge: Harvard University Press, 1950, 1982).
Quantified Natural Deduction
111
prevents the rule from doing any harm. Most of our rules have been truth-preserving: the truth of the rules' premises guarantees the truth of their conclusions. Existential exploitation, however, is not truth-preserving. We should hardly be able to argue, "Some philosophers have been Nazis. Therefore, ristotle was a Nazi." This form of inference is fallacious. But introducing a new name avoids such problems by containing our use of the name within a portion of the proof. We could not use existential exploitation to establish the validity of the above argument. On the left is the attempted proof within quantification theory; on the right is the English equivalent. 1. 3x(Px&Nx)
2. Pa 3. Show Na 4. Pb&Nb
A A 3E,1
Some philosophers have been Nazis. Aristotle was a philosopher. Show that Aristotle was a Nazi. b was a philosopher and a Nazi.
It seems clear that we will never be able to reach' N a'. Because the constant' a' appears in both lines 2 and 3, we can't use existential exploitation to obtain the instance 'Pa&Na' that we would need to reach the conclusion. Although 3E is not truth-preserving, therefore, it never gets us into trouble. The rule is conservative in that any formula without the new constant that follows from the conclusion of the rule also follows from the rule's premise. Mathematicians very often introduce existential assertions and names for the objects asserted to exist in one breath. Consider these examples from a calculus text: Let f be continuous at a. For every number e > a we can choose a number d > that 1f(x) - f(a) 1< e for all x E A with 1x - a 1< d .... Since U is open, there is an open rectangle B with f(a) E B C U. 3
a so
Or, consider this, from a text on Lebesgue integration: Let {Sn} converge to s. Take e = 1 in the definition of convergence; then there exists an integer N such that I Sn - S 1< 1 for all n :::: N, i.e. S - 1 < Sn < S + 1 for all n :::: N. 4 These passages combine the introduction of an existential assertion with its exploitation in one step. They name an object asserted to exist in the very act of making that assertion: "there exists an integer N," "there is an open rectangle B" and "we can choose a number d." Note that, in every instance, the names introduced in this way haven't appeared before; we have no independent information about N, B or d. It would be very different- and outrageous- if these passages were to say instead, "there exists an integer 43" or "we can choose a number Jr." The third rule for quantifiers is universal exploitation. If we know that something is true about every object, then we can conclude that it is true for each particular object that we consider. If God loves everyone, then God loves me, you, and the Earl of Roxburgh. If Jane likes everyone she meets, then she likes you, if she's met you; she likes me, if she's met me; and so on. The rule of universal exploitation says that, from a universally quantified formula, we may infer any of its instances. V'E(Universal Exploitation)
n.V'vA n + p.A[c/v] V'E, n Here c is any constant. 3 Angus E. Taylor, Calculus (Englewood Cliffs: Prentice-Hall, 1959), p. 71. 4Alan J. Weir, Lebesgue Intergration and Measure (Cambridge: Cambridge University Press, 1973), p. 108.
Logic, Sets and Functions
112
This rule does not require us to use a new constant. In fact, it will generally be silly to use a new constant in applying VE. There is no point to introducing a new name unless no constants appear in the proof at all up to this line. If we have constants 'a' and 'b' appearing earlier, and a formula 'VxFx', we can infer 'Fa', or 'Fb', or both. If we also have a formula ''r/xGxx', we can infer 'Gaa' or 'Gbb'. And, if have 'VxVyHxy', then we can obtain 'VyHay' or ''r/yHby' and, in another step, any of 'Haa', 'Hab', 'Hba' and 'Hbb'. We could also, of course, infer similar formulas with other constants. Unless we are forced to introduce those constants in other ways, however, using them to exploit a universal formula will serve no purpose. To see how these rules work, let's demonstrate the validity of a simple argument: Something's upsetting John; whatever upsets John upsets Edna; so something's upsetting Edna.
1. 3xFxj 2. 'ty(Fyj-+Fye) 3. ~ 3zFze j 4. 5. [Fa (Faj-+Fae) 6. Fae 7. 3zFze
A A 3E,1
'tE,2 -+E,5,4 31,6
Problems Use deduction to show that these arguments are valid. 1. God created everything. So God created Pittsburgh. 2. God created everything. So God created Himself. 3. God created everything. So God created something. 4. God created everything. So something created God. 5. Nothing coherent ever baffies me. This course baffies me, so this course is incoherent. 6. All writers who express nationalism in their writing are trying to achieve political aims. Some American writers express nationalism in what they write. Thus, some writers trying to achieve political aims are Americans. 7. Everyone who understands the nature of the radical Islamic movement recognizes that it threatens the existing Arab regimes of the Middle East. Some people in the State Department understand the nature of radical Islam, so some State Department personnel realize that the movement threatens existing Arab regimes. 8. Some plants that are widely cultivated in this area are not able to survive on rainfall alone. All plants native to this area can survive on rainfall alone. It follows that some plants that are not native to this area are widely cultivated here. 9. Some utility companies are predicting brownouts in their service regions during this summer. No utilities that can easily and affordably purchase power from other utilities are predicting brownouts for this summer. Hence, some utility companies cannot easily and affordably purchase power from other utilities. 10. Some computer programs used for processing natural language are written in PROLOG. Nothing written in PROLOG relies heavily on the notion of a list. Consequently, not all computer programs used for processing natural language rely heavily on lists.
Quantified Natural Deduction
113
11. Each person who came to the party was observed by the company president. Some people who came to the party became obviously drunk. Thus, the company president observed some obviously drunk people. 12. Nothing written by committee is easy to write or easy to read. Some documents written by committee are nevertheless extremely insightful. So some extremely insightful documents are not easy to read. 13. Some of the cleverest people I know are clearly insane. Any of the cleverest people I know could prove that this argument is valid. Hence, some people who could prove this argument valid are clearly insane. 14. There are cities in the Sunbelt that are experiencing rapid growth but that are not ranked as very desirable places to live. Every city that experiences rapid growth has to raise taxes. Therefore some cities that will have to raise taxes are not ranked as very desirable places to live. 15. Some analysts insist that we are in the middle of a historic bull market, but others say that the market will soon collapse. Nobody who is expecting the market to collapse is recommending anything but utility stocks. None who believe that Ml controls the direction of the economy contend that we are in the midst of a historic bull market. Thus, some analysts are recommending only utility stocks, but some don't believe that M1 controls the economy's direction. 16. The Longhorns can beat everyone who can beat everyone the Longhorns can. Therefore the Longhorns can beat themselves. 17. One prosecutor can convict another only if he or she can convict everyone that prosecutor can. There are prosecutors who can convict each other. So there are prosecutors who can convict themselves. 18. The unrestricted axiom of abstraction states that there is a set of all and only those objects satisfying any open sentence. Bertrand Russell proved the axiom inconsistent by using the open sentence . . . is not a member of itself Show, following Russell, that '3xVy(y E x +4 Y fj. y)' is contradictory, by proving its negation. 19. Say that an archetypal pig is something such that, if anything at all is a pig, it is. Show that there is an archetypal pig. 20. Say that something is truly ugly just in case, if it is beautiful, then anything is. Show that some things are truly ugly. Use deduction to establish the validity of these argument forms. 21. 3xFx:. VxGx
.. ....
.:-
-4
3x(Fx&Gx)
:
:'22. 3x(Fx&Gx)Vx(Gx
-4
.Hx) :. 3x(.Hx&Fx)
"".'"
23. 3yFyy3xVzGxz:. 3x3y(Gyx&Fxx) 'j
",\
~~24. ;3x3yFxyVxVy(Fxy "'" '~"
+4
(Gx&.Gy)) :. 3xGx&3x·Gx
~-'
25. 3xGx&3x.GxVxVy(Fxy 26. Vx(Fx
-4
+4
(Gx&.Gy)) :. 3x3yFxy
Gx) :. 3x.Fx V 3xGx
Logic, Sets and Functions
114 27. VxVy(Fxy.-. (Gx&-,Gy))3x3y(Fxy&Fyx) .. 3x(Gx&-,Gx)
28. 3x3y(Fx&Gyx)VxVy(Gxy 29. VxVyVz«Fxy&Pxz) 30.
5.2
~
~
(Hx&Jyx)) :. 3x3y«Fx&Hy)&Jxy)
Fyz)3x3y(Pxy&-,Pyx) :. -,VxPxx
* Vz(-,Hz.-. Vx(Fx&Gz))V£Jy(Gy&Fx)
:. -NxHx
Universal Proof
Introducing a universal formula, in this system, requires a new method of proof. We already have three proof techniques: direct proof, indirect proof, and conditional proof. Quantificational logic adds universal proof.
Universal Proof ShewVvA n + 1. [ShewA[C/V]
n.
[: To prove a universal conclusion, in other words, prove an instance of it. The instance must result from substituting a constant new to the proof for the quantified variable. Since no information will appear anywhere earlier in the proof regarding the new constant, it seems to stand for no object in particular. It represents, as it were, an arbitrarily chosen object. Because the proof puts no constraints on it, absolutely any object could play this role. Consequently, though we prove something about c, we have shown how to prove it about anything. And this justifies our drawing a universal conclusion. This too corresponds closely to mathematical practice. Consider this example of a theorem and proof from a standard calculus text: If a function has a derivative which is zero at each point of an interval, the function is constant on that interval. ... [Proof:] Suppose f'ex) = 0 for each x on an interval. . 5
Or this, from a well-known high school geometry textbook: Theorem 9-2. In a plane, two lines are parallel if they are both perpendicular to the same line. Proof. Given that LI-l.L at P and L 2 1.L at Q. It is given that L 1 and L 2 are coplanar. We need to show that they do not intersect. Suppose that L 1 intersects L 2 at a point R. Then there are two perpendiculars from R to L. By Theorem 6-4, this is impossible. Therefore'L} II L 2 . 6 These are universal proofs. The former establishes a result about all functions by proving something about an arbitrary function f. The latter derives a conclusion about any two lines from reasoning about two arbitrarily selected lines, L 1 and L2. (Within this universal proof there is an indirect proof. The proof also combines an existential exploitation with another step: ". . . at a point R.") 5Michael Spivak, Calculus on Manifolds (Menlo Park: W. A. Benjamin, Inc, 1965), pp. 13 and 12. 6Edwin E. Moise and Floyd L. Downs, Jr., Geometry (Menlo Park: Addison-Wesley, 1967), p. 230.
Quantified Natural Deduction
115
It might seem that this form of proof allows us to prove very silly arguments valid. It lets us derive a universal formula from one of its instances. So, can't we show that, if Fred loves the Go-Gos, everybody loves the Go-Gos? Fortunately, no, because of the new-constant requirement. 1.Lfg A Fred loves the Go-Gos. 2. Show VxLxg Show everybody loves the Go-Gos. 3. Show Lag Show a loves the Go-Gos. We can't go from the information that Fred loves the Go-Gos to the conclusion that some arbitrarily selected a does. To take an example of a universal proof, consider this inference pattern: All Fare G; everything is F; so everything isG. To establish its validity, we construct a universal proof:
I. \fx(Fx -Gx) 2. \fxFx 3. ~ \fxGx
4.[~ 5. Fa 6. 7.
Ga
(Fa-Ga) Ga
A O A
\fE,2
'VE,1 -E,6,5
To show that everything is G, we show that some arbitrarily chosen object a is G.
Problems Using deduction, show that these arguments are valid. 1. Anything you think you can achieve, you can achieve. You should try to achieve everything you can achieve. So, anything you think you can achieve, you should try to achieve.
2. Every team that finishes in last place declares its next year a rebuilding year. So, no teams from Chicago will declare next year a rebuilding year only if no Chicago team finishes in last place. 3. No Ivy League colleges have tuitions of under $8,000 a year. Every state-affiliated college has a tuition under $8,000 per year. A college is private if and only if it is not state-affiliated. It follows that every Ivy League college is private. 4. No mammals but bats can fly. Every commonly-kept house pet is a mammal, but none are bats. So nothing that can fly is a commonly-kept house pet. 5. Nothing stupid is difficult. Everything you can do is stupid; anything that isn't difficult, I can do better than you. So anything you can do, I can do better. 6. Anybody who is reflective despises every demagogue. Anybody who is a demagogue despises everyone. Thus, since there are demagogues, every reflective person despises someone who, in turn, despises him or her. 7. There are no good books that do not require their readers to think. Every book that has inspired acts of terror has been inflammatory. No inflammatory books require their readers to think. Therefore all books that have inspired acts of terror are no good. 8. Anyone with some brains can do logic. Nobody who has no brains is fit to program computers. No one who reads this book can do logic. So no one who reads this book is fit to program computers.
116
Logic, Sets and Functions
9. A person is humble if and only if he or she doesn't admire him- or herself. It follows that nobody who admires all humble people is humble.
10. All Frenchmen are afraid of Socialists, and Socialists fear only Communists. Thus, every French Socialist is a Communist. 11. All horses are animals. So all heads of horses are heads of animals. 12. A psychiatrist can help all those who cannot help themselves. So a psychiatrist can help someone who can help himself. 13. All Don Juans love all women. All who have a conscience treat all whom they love well. If some Don Juans have consciences, therefore, all women are treated well. 14. An Olympic athlete could outrun everyone on our team. Since none on our team can outrun themselves, no one on our team is an Olympic athlete. 15. A person is famous if and only if everyone has heard of him or her. So all famous people have heard of each other. 16. If nobody comes forward and confesses, then someone will be punished. So someone will be punished if he doesn't confess. 17. Mary was Jesus's parent. Mary's parents were born with the taint of original sin. But Jesus was not tainted by original sin. Therefore, it's not true that everyone who has a parent born with the taint of original sin is also so tainted. 18. Popeye and Olive Oyl like each other, since Popeye likes everyone who likes Olive Oyl, and Olive Oyl likes everyone. 19. I like everyone who likes everyone I like. So there are people that I like. 20. The government chooses to do x rather than y just in case it doesn't choose to do y over x. A person has veto power just in case the government can choose to do x over y only if that person doesn't prefer y to x. A person is a dictator just in case the government chooses to do x rather than y if he prefers x to y . Consequently, everyone with veto power is a dictator. These are the syllogistic patterns that Aristotle considered valid, together with their medieval names. Show that each is valid in quantification theory. (Some require extra assumptions; they are listed in parentheses.) 21. Barbara: Eyery M is L; Every S is M; :. Every S is L. 22. Celarent: No M is L; Every S is M; :. No S is L. 23. Darii: EveryM isL; Some S is M; :. Some S is L. 24. Ferio: No M is L; Some S is M; :. Some S is not L. 25. Cesare: No L isM; Every S is M; :. No S is L. 26. Camestres: EveryL is M; No S is M; :. No S is L. 27. Festino: No L is M; Some S is M; :. Some S is not L. 28. Baroco: Every L is M; Some S is not M; :. Some Sis notL.
Quantified Natural Deduction
117
29. Darapti: Every M is L; Every M is 5; (There are Ms;) :. Some 5 is L. 30. Felapton: No M isL; Every M is 5; (There are Ms;) :. Some 5 is not L. 31. Disamis: Some Mis L; Every M is 5; :. Some 5 is L. 32. Datisi: Every M is L; Some M is 5; :. Some 5 is L. 33. Bocardo: Some M is not L; Every M is 5; :. Some 5 is not L. 34. Ferison: No M is L; Some M is 5; :. Some 5 is not L. Medieval logicians added other syllogistic patterns to those Aristotle explicitly held valid. Show that these "subaltern moods" are valid, at least with the added assumptions in parentheses. 35. Barbari: Every M is L; Every 5 is M; (There are 5s;) :. Some 5 is L.
36. Celaront: No M is L; Every 5 is M; (There are 5s;) :. Some 5 is not L . 37. Cesaro: No L is M; Every 5 is M; (There are 5s;) :. Some 5 is not L. 38. Camestros: Every L is M; No 5 is M; (There are 5s;) :. Some 5 is not L. Theophrastus, who succeeded Aristotle as head of the Lyceum, also added additional syllogistic principles. Show that these too are valid, with the added assumptions in parentheses. 39. Baralipton: Every M is L; Every 5 is M; (There are 5s;) :. Some L is 5. 40. Celantes: No M is L; Every 5 is M; :. No L is 5. 41. Dabitis: Every M is L; Some 5 is M; :. Some L is 5. 42. Fapesmo: Every M is L; No 5 is M; (There are Ms;) :. Some L is not 5. 43. Frisesomorum: Some M is L; No 5 is M; :. Some L is not 5. These arguments both have the form of argumenta a recto ad obliquum, in the terminology of Joachim Junge, who discussed such arguments in his 1638 textbook. They played an important role in leading logicians beyond Aristotelian logic to a comprehensive theory of relations. Show that each is valid. 44. Knowledge is a conceiving; :. The object of knowledge is an object of conception. (Aristotle, Topics) 45. A circle is a figure; :. Whoever draws a circle draws a figure. (Junge)
These principles concern the distribution of quantifiers over sentential connectives. Quine has referred to them as rules of passage. Show that each is valid. (Throughout, A is any formula not containing x.)
46. V'x(Fx&A)
f-t
(V'xFx&A)
47. 3x(Fx&A)
f-->
(3xFx&A)
48. V'x(Fx
A)
f-->
(V'xFx
V
A)
49. 3x(Fx V A)
f-->
(3xFx
V
A)
V
Logic, Sets and Functions
118
50. Vx(Fx 51. Vx(A
--+
52. 3x(Fx 53. 3x(A
A)
<-4
(3xFx
Fx)
<-4
(A
A)
<-4
(VxFx
Fx)
<-4
(A
-->
-->
-->
--+
-->
--+
A)
VxFx) -->
A)
3xFx)
Chapter 4 asserted that the order of existential quantifiers within a string of such quantifiers makes no differencej similarly for universal quantifiers. Illustrate this by showing valid both these principles:
54. 3x3yFxy
<-4
3y3xFxy
55. VxVyFxy
<-4
VyVxFxy
Although, in general, we can't switch existential and universal quantifiers in order to reach an equivalent formula, we can do this in special circumstances. Without using any rules of passage above, show that each of these switches is legitimate. ... ~"ft
. 56.}3xVyFxYj .. Vy3xFxy 57.
* Vx3y(Fx&Gy)j :.
58.
* Vx3y(Fx V Gy)j:. 3yVx(Fx V Gy)
59.
* Vx3y(Fx
60.
* V£Jy(Gy --> FX)j:.
-->
3yVx(Fx&Gy)
Gy)j:. 3yVx(Fx 3yVx(Gy
-->
Gy)
-->
Fx)
Use deduction to establish the validity of these argument forms.
61. 3xFx
-->
3xGx;:. 3x(Fx
-->
Gx)
62. 3xFx V 3xGx; .. 3x(Fx V Gx) 63. 3x(Fx V GX)j:. 3xFx V 3xGx 64. 3xFx
-->
Vy(Gy
-->
Hy); 3xJx
65. 3xFx V 3xGx; Vx(Fx 66. Vx((Fx&Gx)
-->
--+
-->
3xGx;:. 3x(Fx&Jx)
-->
Gx);:. 3xGx
HX)j Ga&VxFx; .. Fa&Ha
,,:... ............
'67. Vx(3yFyx
-->
VzFxz); .. VyVx(Fyx
-->
Fxy)
"''0.;
68. Vx(Fx
-->
Vy(Gy
-+
Hxy»; 3x(Fx&3y--.,Hxy);:. 3x--.,Gx
69. -.3x(Fx&Cx)j:. Vx(Fx
70. Vx(Fx
-+
--+
-.Gx)
-.Cx);:. -.3x(Fx&Gx)
3zHz
Quantified Natural Deduction
5.3
119
Derived Rules for Quantifiers
The system resulting from adding these three rules and universal proof to our natural deduction system for sentential logic already has all the power it needs. It can demonstrate the validity of any valid argument form in quantification theory. Nevertheless, adding some further rules can increase its efficiency and naturalness considerably. This section will present some derived rules which, while theoretically superfluous, make the proof system more pleasant to work with. First, we can apply quantifier rules several times in a single step. Suppose, for example, that we want to take instances of 'VxVyFxy'. The universal exploitation rule requires us to move first to the instance 'VyFay', say, and then to its instance' Faa'. But any constant can substitute for both 'x' and 'y' here, since both are universally quantified. So it's easy to perform the operation in one step. We can move directly from 'VxVyFxy' to an instance of an instance of it, 'Faa'. Similarly, we could move to 'Fab', 'Fba', 'Fbb', 'Fac', etc. We can record that we've exploited a series of n universal quantifiers at once by writing not 'V E' but 'VEn'. In this case, then, the application would look like:
m. m+p.
VxVyFxy Faa
VE 2 ,m
We can readily do the same for series of existential quantifiers, provided that we replace each quantified variable with a new constant. So we may move immediately from '3x3y3z(Fxy&Fyz&Fzx)' to 'Fab&Fbc&Fca', citing the rule we've applied as 3E3 . Finally, we can also compress a sequence of universal proofs into a single proof in a similar way. We can prove a formula with an initial string of n universal quantifiers by proving an instance of an instance. . . of an instance of the formula, using n different new constants to replace the n quantified variables. For example, we can prove 'VxVy(Fxy --+ --.Fyx) , by proving 'Fab --+ --.Fba', where neither 'a' nor 'b' have appeared earlier in the proof. Although these combinations of rules or proofs are very convenient, it's not a good idea to combine exploitations of universal and existential quantifiers. While VE 2 and 3E4 are fairly easy to follow, something like VE 2 3E 3VE would be extremely difficult to apply or understand. Second, variables have no independent meanings. The formulas 'VxFx' and 'VzFz' function in logically similar ways; so do '3yGy' and '3wGw'. Indeed, it's easy to show that these pairs are equivalent. Here are the proofs in one direction:
1. VxFx
A VzFz 3. r~ Fa 4. [Fa VE,l 2.~
1. 3yGy 2. ~
3. [Ga
4.
3wGw
A 3wGw
3E,1 31,3
The proofs of the other directions follow exactly the same pattern. Thus, we can substitute one variable for another throughout a formula. The only restriction we must observe is that we should not introduce into a formula a variable that is already there; otherwise we could go from the legitimate formula 'VxVyFxy' to the very different nonformula 'VxVxFxx'. The derived rule, then, is this.
Logic, Sets and Functions
120
V R(Variable Rewrite)
n.A m.A[v/u]
VR,n
(wherevis foreign toA)
Third, it's extremely useful to have a direct way of dealing with negations of quantified formulas. Our rules allow us to attack formulas with quantifiers as main connectives in one step. But, if a quantifier is preceded by a negation sign, the proof strategy becomes much more complicated. Luckily, negations of quantified formulas are equivalent to formulas with quantifiers as main connectives. Two rules, called quantifier negation rules, relate quantified formulas to their negations. In the process, these rules relate the universal and existential quantifiers. In fact, they show how to define each quantifier in terms of the other. QN(Quantifier Negation)
n.--,3vA mVv--,A
QN,m QN,n
QN(Quantifier Negation)
n.--,VvA m.3v--,A
QN,m QN,n
Both versions of quantifier negation are invertible. That is, the premise and conclusion are equivalent, so they can be used in either direction. We can infer '3x--,Fx' from '--,VxFx', and vice versa. Similarly, just by adding a negation sign, we can see that '3xFx' is equivalent to '--,Vx--,Fx', and that 'VxFx' is equivalent to '--,3x--,Fx'. So we can define the quantifiers in terms of each other. Deriving even simple applications of these rules from our basic quantifier rules shows how much work they can save. These two proofs are necessary to show, for example, that '3x--,Fx' and '--,VxFx' are equivalent.
1. 3x...,Fx 2.~
3. 4. 5.
["XFX ...,Fa
A
...,VxFx AlP 3E,1 VE,3
Fa
1. ...,VxFx
A
3x...,Fx ...,3x...,Fx ~ VxFx
2.~
3. 4. 5. 6. 7. 8. 9.
~~ ...,Fa
Fa
3x...,Fx ...,3x...,Fx ...,VxFx
AlP AlP 31,6 R,3 R,l
Quantified Natural
Deduciittm~
121
\.....
"'*~ . /., Deriving the other equivalence is similar. Because QN takes the for:- of an equivalence, the replacement principle allows us to apply it to portions of formulas as well as entire formulas. Each of the following is thus a legitimate application ofQN, -,3xFxx -,'t/x't/yGxy 't/x't/y-,'t/z(Fxz&Fyz) 3x-,3yGyx 't/x-,Fxx 3x-,'t/yGxy 't/x't/y3z-,(Fxz&Fyz) 3x't/y-,Gyx This rule, too, can be applied several times in a single step: we can abbreviate -,'t/x't/y't/z((Rxy&Ryz) --+ 3x-,'t/y't/z((Rxy&Ryz) --+ 3x3y-,'t/z((Rxy&Ryz) --+ 3x3y3z-,((Rxy&Ryz) --+
Rxz) Rxz)QN Rxz)QN Rxz)QN
-,'t/x't/y't/z((Rxy&Ryz)
Rxz) 3x3y3z-,((Rxy&Ryz)
to --+
--+
Rxz)QN3
Problems Establish the validity of these argument forms by means of deduction. .,\, 1. 't/x(Fx
-->
't/y(Gy
-->
2. 't/x((Fx V Hx)
-->
3. 3x(Fx&'t/y(Gy
-->
Hxy));'t/x(Dx
-->
't/y(Hxy
-->
Cy));:. 3x(Fx&Dx)
Hxy));:. 3x(Fx&(Ga
-->
Hx));'t/zGaz;:. 3xFx
-->
~
Hx); 3x3y((Fx&Mx)&(Gy&Jyx)); 3xHx
'
9. '
3xGx
-->
Gx);:, 't/xFx
11. 3x(Fx&'t/y(Gy 12.,'3x'
..,., ....-
13, 't/x(Kx
-->
-->
-->
Hy));'t/x(Fx
-->
16. 't/x(Fx
-->
--+
(-,Lx
-->
-->
3zLzx));'t/x(3zLzx
Hx); 3xHx
-->
-->
Lxx);-,3xLxx; .. 't/x(Kx
-->
-,Jyz);:. 3x(Gx&Hx)
-->
3xLx
-->
-->
't/y-,Lxy)
Gxx);:. 3xGxx
-,3yGy;:. 't/x(Gx
(Gx V Hx)); 't/x((Jx&Fx)
18. '
-->
-->
-,3zFzz)
-,Gx); 't/x(-,Fx -->
-->
-,Jx);:. 't/x(Jx
Fxe);:. -,(Hea&Fec)
Cx); 3y(Dy&3x((Fx&Gx)&Byx));'
'19,!'t/x't/y't/z((Fxy&Fyz)
-->
Fxz); -,3xFxx;:. 't/x't/y(Fxy
,20.)Vx't/y't/z((Fxy&Fyz)
--+
-,Fxz);:. 't/x-,Fxx
.~.
-->
--,3z(Kz&Hz)));:. 3x(Kx&Gx)
17. ....,3x(Hxa&....,Gxb);'t/x....,(Fxc&Fbx);'t/x(Gdx
........ '~.
't/y't/z(-,Hy
Fyx);'t/x3yFxy; .. 3x't/yFyx
(3yLxy
-->
-->
3xGx
14. -,'t/x(Hx V Kx); '
-,Hx);:. 3xHx
't/yFyx); 3x3yFxy;:, 't/x't/yFxy
't/x(Fx ...... Gx);:. 't/xFx
10. 3x(Fx
Cy)
3xHx
6. 't/x(~x
48.
3y-,Gy); 3x3yFxy; 't/x(Gx
-->
-->
-->
-->
Hxa))
5. 't/x(3yFxy
7. '
't/y(Gy
(Gx&Kx)); -,'t/x(Kx&Gx);:. 3x-,Hx
4. 't/x't/y(Gxy ~ (Fy
--+
-->
--+
-,Fyx)
-->
-->
Hx)
t- a
Ax);:. 3x(Cx&Dx)
~ Logic, Sets and Functions
122 .
~ ~
-',
. 2t(VxVyVz((Fxy&Fyz) ~ Fxz);VxVy(Fxy ~ Fyx); 3x3yFxy; : , x x 22. VxVyVz((Fxy&Fxz)
~
~
Fyx); VxFxx;:. VxVyVz((Fxy&Fyz)
~
Fxz)
23. VxVyVz((Fxy&Fyz)
~Fxz); VxVy(Fxy ~
Fyx); VxFxx;:. VxVyVz((Fxy&Fxz)
~
Fyz)
24. VxVyVz((Fxy&Fxz)
~
Fyz); VxVy(Fxy
Fyz); .. VxVyVz((Fxy&Fxz)
~
Fzy)
25. 3xVy-.Fxy;:. 3xVyVz(Fxz ~ Fzy)
26. Vx(Fx 27. Fa
f-+
~
(3xGx
28. Vx(Dx 29. 3xFx
VyGy);:. VxFx V Vx-.Fx
~
~
~
Fx);:. Da
~
Vy((Fy V Gy)
~
~
~
(Vy(Fy ~
Hx);Vx(-,Jx ~
Gy)
~
(3y(Ay&Gy)
3xGxb; :. 3x(-.Fxc
~
Gxb)
34. Vx(Fx
,
~
(35. VxVy(Fxy
~
Vy(Py
~
Fyx); .. VxVy(Fxy
36. Vx-,Fxx;:. -,3xVy(Fyx
~
+-+
~
(.Fa VVx-.Gx)
Bxx)); 3z(Az&Tz);:. 3xBxx
Fy))
f-+
~
.Lx); .. 3x(Nx&Mx)
3y(Fy&Hxy));:. VxVy(Hxy
~
Hyx)
~
Vx-,(Fx&Gx)
Fyx)
3zVw((Fwz
~
37. VxVy((Ax&By) Cxy); 3y(Fy&Vz(Hz Bx); .. VzVy((Az&Hy) ~ Czy)
~.ii
~
~
Ly); 3x(Px&Nx); Vx(Px
.3y(Gy&Hxy));Vx(Fx ~
~
Hy); 3xHx; -,Vz-.Fz;:. 3x(Fx&Hx) ~
33. 3x(Px&-.Mx)
-,Hx);:. -,Jb
Ga)
32. VxFx;:. -,3xGx +-+ -.(3x(Fx&Gx)&Vy(Gy
:,;;;'~
~
Gy)); Vx(Fx
30. 3x(Fx&Vy(Ty 31. Vx-.Fxc
Gb);Vx(Gx
Fwy)&-,Fzy))
~
Cyz)); VxVyVz((Cxy&Cyz)
~ CxIVX(FX ~
Chapter 6
Identity and Function Symbols This chapter will extend our system of quantificational logic to include a new binary predicate and a new way of forming singular terms. The new predicate, identity, represents the familiar notion of equality in mathematics and an equally familiar concept in natural language. English most often expresses this concept using the word is. New singular terms will be formed with the help of function symbols. Function symbols occur frequently in mathematics; '+', 'x',"-', and '7 ',not to mention "f' and '8', all represent functions. English expressions such as mother of, diameter of, length of, kinetic energy of, hometown of and grade of also represent functions. We've discussed functions at several points in earlier chapters; we will now develop ways of representing them explicitly in our formal language. Quantification theory as it stands can already represent identity by using a dyadic predicate and adopting some special axioms concerning it. The theory can furthermore represent functions and definite descriptions by using polyadic predicates and some accompanying assumptions. But adding identity and function symbols permits quantification theory to formalize very naturally a wide variety of English arguments and mathematical theories. In certain subtle ways, it also adds to quantificationallogic's power.
6.1
Identity
To symbolize identity, we'll use the symbol '=', which mathematicians generally use for the same concept. (For negations of identity formulas, we'll use the symbol '=1=' to abbreviate '...,... = .. .'. Thus, we'll generally write ....,a = b as a =1= b.) Obviously, having such a predicate in our logic allows us to work with mathematical formulas with a minimum of translation or alteration. It also allows us to render many English sentences very simply and naturally: 1.
(a) Tully is Cicero. (b) Austin is the capital of Texas. (c) Ronald Reagan is the President of the United States. (d) The wealthiest town in the United States is West Hartford, Connecticut. (e) Baltimore was Babe Ruth's birthplace.
In each of these sentences, the verb to be expresses identity. This, of course, isn't that verb's only job. In Dan's car is red, Phyllis is angry and Peter is coming to town tomorrow it expresses predication. Aristotle was perhaps the first person to recognize the distinction between these two 123
124
Logic, Sets and Functions
roles of to be. Philosophers today sometimes refer to this as the contrast between the is of identity and the is of predication. A primary use of the identity symbol in logic is to formalize the is of identity. Adding the identity symbol to our language requires only one small revision in our formation rules. We must add the rule that, where c and d are constants, c = d is a formula. (In fact, it's an atomic formula.) The required addition to our semantic principles is similarly trivial. A formula of the form c = d is true on an interpretation just in case the interpretation function assigns the same object to both c and d. Identity allows us to translate a wide and surprising variety of English constructions. First, we can easily express numerical determiners. In quantificational logic, the existential quantifier means, in essence, at least one. But how can we express, for instance, at least two? We can use two existential quantifiers, but this, in itself, doesn't suffice; '3x3y(Fx&Fy)' doesn't rule out the possibility that 'x' and 'y' stand for the same object. To say that there are at least two Fs, we need to say that there is an object x and there is another object y, both of which are Fs. We can say that x and yare different by writing 'x =1= y'. So the appropriate symbolization becomes '3x3y(Fx&Fy&x =1= y)'. To say that there are at least three Fs, we need to write three existential quantifiers and three negated identities to guarantee that all the objects are distinct. The correct translation is correspondingly "3x3y3z(Fx&Fy&Fz&x =1= y&y =1= z&x =1= z)'. We can apply the same strategy for any number of objects.
There are at least . . . Fs Translation One Two Three Four
3x 3x3y 3x3y:3z :3x3y:3z:3w
Fx (Fx&Fy&x =1= y) (Fx&Fy&Fz&x =1= y&y =1= z&x =1= z) (Fx&Fy&Fz&Fw& x =1= y&x =1= z&x =1= w&y =1= z&y =1= w&z =1= w) n 3Xloo.:3xn (FXI& ...&Fxn& Xl =1= X2&Xl =1= X3& .. ·&Xl =1= Xn&X2 =1= X3& .. ·&x2 =1= Xn &... &Xn-l =l=x n ) We can thus symbolize numerical determiners of the form at least n. Notice, however, that the complexity of the resulting formulas increases very quickly. We can also translate sentences containing determiners of the form at most n. Suppose, for example, that we want to say that there is at most one omnipotent being. This assertion does not imply that there is an omnipotent being; it merely states that there is no more than one. The assertion is the negation of There are at least two omnipotent beings, so we might try to negate a formula of the kind we've just seen, obtaining '-":3x:3y(Ox&Oy&x =1= y)'. This is equivalent to 'VxVy((Ox&Oy) --> x = y)', which says that, for any x and y, if x and yare both omnipotent, then x and yare identical. It says, that is, that all omnipotent beings are the same. But that means that there is at most one omnipotent being. Similarly, if we want to translate There are at most two superpowers, we can treat it as the negation of There are at least three superpowers, obtaining '-'3x3y3z(Sx&Sy&Sz&x =1= y&x =1= z&y =1= z)'. But this is equivalent to 'VxVyVz((Sx&Sy&Sz) --> (x = y V x = z Vy = z»', which says that, if x, y and z are all superpowers, two of them must be identical. This strategy works for any value of n.
Identity and Function Symbols
125
There are at most . . . Fs Translation One Two
VxVy VxVyVz
Three
VxVyVzVw
n
Vxll..·Vx n
((Fx&Fy) ---> x = y) ((Fx&Fy&Fz) ---> (x=yVx=zVy=z)) ((Fx&Fy&Fz&Fw)---> (x = y V x = z V x = w Vy = z Vy = w V z = w)) ((FXI& ...&Fxn)---> (Xl = X2 V Xl = X3 V . VXI = Xn V X2 = X3 V .. VX2 = Xn V ... V Xn-l = xn))
Finally, we can translate numerical determiners of the form exactly n. Since There are exactly n Fs is equivalent to a conjunction of There are at least n Fs and There are at most n Fs, we can devise such translations simply by conjoining the translations given in the above tables. But there is an easier way. If we want to say that there is exactly one god, we can say that there is a god, and anything that is a god is identical with it. If we want to say that there are exactly two great American writers, we can say that there are at least two and, moreover, that any great American writer must be identical with one or the other. This suggests the strategy:
There are exactly Zero One Two Three
Vx 3xVy 3x3yVz 3x3y3zVw
n
3XI ... 3xnVy
.. Fs Translation --,Fx (Fx&(Fy ---> y = x)) (Fx&Fy&x =1= y&(Fz ---> (z = x V z = y))) (Fx&Fy&Fz&x =1= y&x =1= z&y =1= z& (Fw ---> (w = x V w = y V w = z))) (FXI& ...&Fxn &XI =1= X2&XI =1= X3&'" &XI =1= Xn&X2 =1= X3&'" &X2 =1= xn&",&Xn-1 =1= Xn &(Fy ---> (y = Xl V ... V Y = xn)))
Second, identity allows us to symbolize sentences containing the words other, another andelse. Typically, these words are anaphoric in much the way pronouns are. That is, they carry with them an implicit reference to something introduced by prior discourse. If we say John admires himself. But he hates everybody else, we mean that John hates everybody other than John. A good symbolization of the latter sentence, then, would be 'Vx(x =1= j ---> Hjx)', which says that everyone who isn't John is hated by John. Of course, the antecedent discourse may have introduced some other referent. If the preceding sentence were John loves Susan, we could take everybody else as meaning "everybody other than Susan" or "everybody other than John and Susan." An.appropriate symbolization would be 'Vx(x =1= s ---> Hjx)' or 'Vx((x =1= s&x =1= j) ---> Hjx)'. Unfortunately for the translator, natural languages contain referential ambiguities, and determine the range of possible interpretations in very complex and still poorly understood ways. In many cases, only one interpretation is possible. If we say A passenger shouted, "This is a hijacking!" and another waved a pistol overhead, we clearly mean that a passenger other than the one who shouted did the pistol-waving. Introducing the obvious predicates, then, we could translate this sentence as '3x3y3z(Px&Sx&Py&y =1= x&Gz&Wyz)' with all quantifiers in front, or, more naturally, as '3x((Px&Sx)&3y(Py&y =1= x&3z(Gz&Wyz)))'. Third, identity allows us to translate superlatives such as fastest and most interesting into formulas containing predicates that represent comparatives such as faster and more interesting. Suppose that' F' symbolizes is faster than. To express This is the fastest car, we can say that it's faster than every other car. That is, we can say that this car is faster than every car not identical with it. So an appropriate translation would be 'Vx((Cx&x =1= a) ---> Fax)', where 'a' stands for this car. (We might also want to add a subformula 'Ca', to say that this is a car.) If we want to say that this is
126
Logic, Sets and Functions
the slowest car, we can say that every other car is faster than it: 'Vx«Cx&x =f. a) -+ Fxa)'. Notice that, if this is the fastest car, then it follows that no car is faster than it. But these two statements aren't equivalent. Suppose there are several cars that tie for the title of fastest car in the universe. No car will be faster than any of them, but none will be the fastest car; each will be one of the fastest, but not the fastest. Superlatives thus translate readily into formulas containing dyadic predicates representing correlated comparatives. One must be very cautious in translating, however, to avoid confusing the direction of the comparison and mixing up more and less: a is more F thanb Fab a is the mostF Vx(x =f. a -+ Fax) a is the least F Vx(x =f. a -+ Fxa) a is less G than b Gab a is the most G Vx(x =f. a -+ Gxa) a is the least G Vx(x =f. a -+ Gax) The order of the variable and constant in this formula makes all the difference between more and less, most and least. In our examples above, we had to add a predicate to these schemas; we wanted to say, not that this was the fastest thing, but that this was the fastest car. Finally, identity permits us to symbolize many sentences with the word only. Recall that we can translate sentences of the form Only F G as 'Vx(,Fx -+ ,Gx)' or, equivalently, 'Vx(Gx -+ Fx)'; as equivalent, that is, to No non-F G or All G F. Identity allows us to translate as well sentences involving only in combination with singular terms. Consider, for example, Only Elmo got drunk. We could paraphrase this as Only those identical with Elmo got drunk. So, applying the strategy we devised for only -general-term combinations, this should be equivalent to All who got drunk were identical with Elmo, or, in symbolic terms, 'Vx(Dx -+ x = e)' . Equivalently, we could say that nobody but Elmo got drunk: 'Vx(x =f. e -+ ,Dx)'or'-,3x(Dx V x =J- e)' or 'Vx(-,Dx V x = e)'. These formulas seem surprising, for they do not imply that Elmo got drunk. Only Elmo got drunk is at least extremely misleading, if not false, if Elmo stayed sober. The translation of Only Elmo got drunk as 'De&Vx(Dx -+ x = e)'- "Elmo, and only Elmo got drunk"- thus better harmonizes with certain linguistic intuitions.
Problems Symbolize these sentences in Q with identity. If a sentence is ambiguous, explain why. 1. April is the cruelest month. (T. S. Eliot)
2. Twice no one dies. (Thomas Hardy) 3. Philosophy is the highest music. (Plato) 4. To work for the common good is the greatest creed. (Albert Schweitzer) 5. This poem is the reader and the reader this poem. (Ishmael Reed) 6. Liberty is always dangerous, but it is the safest thing we have. (Harry Emerson Fosdick) 7. Action is the last resource of those who know not how to dream. (Oscar Wilde) 8. The seed ye sow, another reaps; the wealth ye find, another keeps; the robes ye weave, another wears; the arms ye forge, another bears. (Percy Bysshe Shelley) 9. The most precious thing a parent can give a child is a lifetime of happy memories. (Frank Tyger)
Identity and Function Symbols
127
10. That action is best, which procures the greatest happiness for the greatest number. (Francis Hutcheson) 11. The worst-tempered people I've ever met were people who knew they were wrong. (Wilson Mizner) 12. The most valuable executive is one who is training somebody to be a better man than he is. (Roobert Ingersoll) 13. I hold that man in the right who is most closely in league with the future. (Henrik Ibsen) 14. An executive organization, like a chain, is no stronger than its weakest link. (Robert Patterson) 15. In cases of difficulty and when hopes are small, the boldest counsels are the safest. (Livy) 16. Behold, the people is one, and they have all one language; . . . and now nothing will be restrained from them, that they have imagined to do. (Genesis 11:6) 17. Once upon a time there were Three Bears, who lived together in a house of their own, in a wood. (Robert Southey) 18.
* He prayeth best, who loveth best all things both great and small.
(Samuel Taylor Coleridge)
Sentences 19, 20, 23 and 24 are cited by James McCawley; each involves only in combination with singular terms. 19. Only Lyndon pities himself. 20. Only Lyndon pities Lyndon. 21. Lyndon pities only Lyndon. 22. Lyndon pities only himself. 23.
* Only Lyndon
24.
* Only
25.
* Only Lyndon
6.2
pities only Lyndon.
Lyndon pities only himself. pities only those who pity only Lyndon.
Deduction Rules for Identity
The two principles of self-identity and the indiscernibility of identicals underlie natural deduction rules for identity. The rule for identity introduction is extremely simple: whenever you wish, in the course of a proof, you may record c = c, for any constant c. Identity Introduction( = 1)
n.c - c
= I
128
Logic, Sets and Functions
This rule allows a very quick proof of the principle that everything is self-identical:
1. ~ 'tJxx=x 2. r~ a=a 3. [a=a =1 The rule of identity exploitation x states that, for any constants c and d, c = d justifies substituting c for d, or d for e, in any formula free at that point in the proof. Identity Exploitation( = E) n.c = d
m.A p.A[cl Id](orA[dl Ie])
= E, n, m
Here A[cl Id] is any result of substituting c for some or all occurrences of d throughout A. If A is 'Fdd', for instance, Alclld] could be 'Fed', 'Fdc', or 'Fcc'. This rule allows us to prove that identity is symmetric: that the order of the terms in an identity stat ement, in other words, makes no difference.
1. ~ 'tJx'tJy(x=y---->y=x) 2. ~~ 'tJy(a=y---->y=a) 3. ~ (a=b---->b=a) 4. a=b 5. a=a 6. b=a
ACP =1
=E,5,4
We can reach line 4 here either by introducing 'a = a' by identity introduction or by substituting 'a' for 'b' in 'a = b'. Identity exploitation also allows us to prove that identity is transitive.
1. 2. 3. 4. 5. 6. 7. 8.
~ ~ ~
'tJx'tJyVz«x=y&y=z)---->x=z) 'tJy'tJz«a=y&y=z)---->a=z) 'tJz«a=b&b=z)--a=z) ~ «a=b&b=c)-a=c) (a=b&b=c) a=b b=c a=c
ACP &E,5 &E,5 =E,6,7
We begin by using a triple universal proof. We reach the crucial line, 6, by using identity exploitation. We can think of ourselves as substituting 'a' for 'b' in 'b = e', or as substituting 'e' for 'b' in 'a = b'. Either way, we end up with 'a = c'. To take a final example, suppose we are faced with the argument, John is selfish. But everybody else is selfish too. So all people are selfish. Assuming that John is a person, we can let our universe of discourse consist consist just of people, and so translate this Sj \/x(x =I- j -. Sx) :. VxSx
To show that this argument form is valid, we can construct a universal proof, which contains an indirect proof.
129
Identity and Function Symbols
A A
1. Sj 2. '{Jx(,xlj-Sx) 3. ~'VxSx 4. ~ Sa 5. ....,Sa 6. (,a=j~Sa) 7. ".a=J 8. a=J 9. Sa
AlP VE,2 -E*,6,5 ",7 =E,1,8
We want to show that some arbitrary person- say Ann- is selfish. So we assume, for purposes of indirect proof, that she isn't selfish. Since everybody other than John is selfish, but Ann isn't, Ann must be John. But John is selfish as well; so if Ann is John, then, by identity exploitation, she's selfish. But this is contradicts our assumption that she isn't selfish.
Problems We can say that there is one and only one God, in symbolic terms, by writing the formula ':JxVy(y = x ~ Fy)'. Show that each of the following is a consequence of this formula. 1. :Jx(Fx&Gx) ...... Vx(Fx --. Gx)
2. VxFx --. :Jx(Fx&Gx) 3. 3x(Fx&Gxx) ...... 3x3y(Fx&Fy&Gxy)
4. 3x(Fx&(Ga --. Hx)) ...... (Ga --. 3x(Fx&Hx)) 5. 3x(Fx&(Gx
->
Ha))
~
6. 3x(Fx&-.Gx) ...... Vx(Fx
(3x(Fx&Gx) ->
->
Ha)
...,Gx)
7. :Jx(Fx&VyGyx) ...... Vx:Jy(Fy&Gxy)
8. VxVy((Fx&Fy) --. x 9. VxFx
->
VxVyx
10. 3x3yVz(x
6.3
= y)
=Y
t- y&(z = x V z = y)) ...... 3xVy(y = x ....... -.Fy)
Function Symbols
Function or operation symbols combine with constants, variables and other function symbols to form singular terms. They are extremely common in mathematics. The basic arithmetic operations of addition, subtraction, multiplication and division, for example, are all functions. Function symbols also translate many English possessives and of constructions: John's birthday the length of this line Barbara's father the truth value of the sentence Jill's house the mass of an electron somebody's BMW the trunk of my Camara nobody's honor the President of the United States Each of these expressions translates into quantification theory with the help of function symbols.
128
Logic, Sets and Functions
This rule allows a very quick proof of the principle that everything is self-identical:
1. ~
2. r~ 3. [a=a
Vxx=x a=a =1
The rule of identity exploitation x states that, for any constants c and d, c = d justifies substituting c for d, or d for c, in any formula free at that point in the proof. Identity Exploitation( = E) n.c = d
m.A p.A[clld](orA[dllc])
= E,n,m
Here A[cl I d] is any result of substituting c for some or all occurrences of d throughout A. If A is 'Fdd', for instance, A[elld] could be 'Fed', 'Fdc', or 'Fcc'. This rule allows us to prove that identity is symmetric: that the order of the terms in an identity stat ement, in other words, makes no difference.
1. ~ VxVy(x=y--y=x) 2. ~~ Vy(a=y-+y=a) 3. ~ (a=b-+b=a) 4. a=b 5. a=a 6. b=a
ACP
=1 =E,5,4
We can reach line 4 here either by introducing 'a = a' by identity introduction or by substituting 'a' for 'b' in 'a = b'. Identity exploitation also allows us to prove that identity is transitive.
1. 2. 3. 4. 5. 6. 7. 8.
~ ~ ~
VxVyVz«x=y&y=z)-+x=z) VyVz«a=y&y=z)-+a=z) Vz«a=b&b=z)-+a=z) ~ «a=b&b=c)--a=c) (a=b&b=c) a=b b=c a=c
ACP &E,5 &E,5 =E,6,7
We begin by using a triple universal proof. We reach the crucial line, 6, by using identity exploitation. We can think of ourselves as substituting 'a' for 'b' in 'b = e', or as substituting 'e' for 'b' in 'a = b'. Either way, we end up with 'a = c'. To take a final example, suppose we are faced with the argument, John is selfish. But everybody else is selfish too. So all people are selfish. Assuming that John is a person, we can let our universe of discourse consist consist just of people, and so translate this 8j 'v'x(x f j -.8x) :. 'v'x8x
To show that this argument form is valid, we can construct a universal proof, which contains an indirect proof.
129
Identity and Function Symbols
A
1. Sj 2. Vx(~xlj-Sx) 3. ~·VxSx 4. ~ Sa 5. ~Sa
A
AlP VE,2 -E* ,6,5
(~a=j:-Sa) ~~.a=J
6.
7. 8. 9.
~~,7
a=J Sa
=E,I,8
We want to show that some arbitrary person- say Ann- is selfish. So we assume, for purposes of indirect proof, that she isn't selfish. Since everybody other than John is selfish, but Ann isn't, Ann must be John. But John is selfish as well; so if Ann is John, then, by identity exploitation, she's selfish. But this is contradicts our assumption that she isn't selfish.
Problems We can say that there is one and only one God, in symbolic terms, by writing the formula '3xVy(y = x ...... Fy)'. Show that each of the following is a consequence of this formula. 1. 3x(Fx&Gx) ...... Vx(Fx
2. VxFx
-->
-->
Gx)
:Jx(Fx&Gx)
3. :Jx(Fx&Gxx) ...... 3x3y(Fx&Fy&Gxy)
4. 3x(Fx&(Ga
-->
Hx))
(Ga
5. 3x(Fx&(Gx
-->
Ha))
(:Jx(Fx&Gx)
6. :Jx(Fx&-.Gx) ...... Vx(Fx
-->
-->
3x(Fx&Hx)) -->
Ha)
-.Gx)
7. 3x(Fx&VyGyx) ...... Vx:Jy(Fy&Gxy)
8. VxVy«Fx&Fy)
-->
9. VxFx
=Y
-->
VxVyx
10. 3x3yVz(x
6.3
=1=
y&(z
x = y)
= x V z = y)) ...... :JxVy(y = x ...... -.Fy)
Function Symbols
Function or operation symbols combine with constants, variables and other function symbols to form singular terms. They are extremely common in mathematics. The basic arithmetic operations of addition, subtraction, multiplication and division, for example, are all functions. Function symbols also translate many English possessives and of constructions: John's birthday Barbara's father Jill's house somebody's BMW nobody's honor
the the the the the
length of this line truth value of the sentence mass of an electron trunk of my Camara President of the United States
Each of these expressions translates into quantification theory with the help of function symbols.
Logic, Sets and Functions
130
All functions take a certain number of arguments, i.e., inputs. A function laking just one input is a singulary function. Binary functions take two inputs; in general, n-ary fun~tions take n inputs. Addition, multiplication, subtraction and division are ail binary functions. Taking a square and taking a square root are both singulary functions. Let's agree to call constants, variables, and function terms singular terms, or, more simply, terms. To form a function term, we must concatenate an n-ary function symbol with n constants, variables, or other terms. The string of n terms must be enclosed in parentheses and separated by commas. Where'!" 'g' and 'h' are singulary, binary and ternary function symbols, respectively, we can form the function terms
f(a) f(x) f(j(b)) f(j(y)) f(g(a, a)) f(h(x, y, z))
g(a,a) g(a,x) g(y,b) g(x,y) g(a,g(a,a)) g(j(x), h(b, f(y), z))
h(a, b, c) h(x,y,z) h(a,z,b) h(j(a) , f(b), f(c)) h(g(x,x),x,x) h(h(x,y,z),f(z),g(z,x))
Notice that function terms may appear inside other function terms. Function symbols of three or more places are usually written in front of their argument terms, with parentheses marking the boundaries. But singulary and binary function terms may be written in other ways. We may write binary function symbols between their argument terms, supplying parentheses only when necessary to avoid ambiguity. Thus mathematicians generally write '2 + 3' rather than '+(2,3)', 'a U b' rather than 'U(a, b)', etc. Indeed, we've done the same in sentential logic. Sentential connectives express truth functions, but we've written the binary connectives between the formulas they link together. Thus we've written the formulas on the left rather than the versions on the right:
pvQ (P -+ Q)&R R ...... (P&--,R)
V(P,Q) &( -+ (P, Q), R) ...... (R, &(P, ,Q))
In fact, since we know which connectives are singulary and which are binary, we don't need the parentheses in the expressions on the right. So we could express the formulas more simply:
pvQ (P -+ Q)&R R ...... (P&--,R)
v(P,Q) &(-+ (P,Q),R) ...... (R, &(P, ,Q))
vPQ & -+ PQR ...... R&P,Q
The third column shows the form in which computers generally read logical formulas and other instructions. A group of Polish logicians developed and wrote in this notation earlier in this century; for this reason, it's called Polish notation. To incorporate function terms into our language, we need to change its formation rules slightly. Officially, we'll maintain the standard notation for function terms, though we'll allow ourselves to use forms in more frequent use when that proves convenient. Lower-case letters from the middle of the alphabet- '1', 'g', 'h', 'i', etc., with or without subscripts- will serve as our function symbols. We'll say that a function term is open if its contains some occurrences of variables, and closed if it doesn't. A closed term is either a constant or a closed function term. We may amend our formation rules to read: 1. An n-ary predicate followed by n closed terms is a formula. 2. If t and t' are closed terms, then t = t' is a formula. 3. If A and B are formulas, then so are ,A, (A&B), (A V B), (A -+ B), and (A ...... B).
Identity and Function Symbols
131
4. If A is a formula with a closed term t and v is a variable not in A, then 3vA[v/t] and \;IvA[v/t] are formulas.
5. Every formula may be constructed by means of finitely many applications of rules (1)-(4). Semantically, function symbols stand for functions on the domain; closed function terms stand for individual elements of the domain. Function terms thus differ sharply from formulas. Formulas are either true or false; closed function terms stand for objects. Connectives may join together formulas, but not function terms: 'f(a) --> g(b,c)', which, in English, would correspond to something like If Ann's mother, then Bill and Carlota's anniversary, makes no sense. Similarly, function terms may flank the identity sign, but formulas may not. A formula such as 'f (a) = g( b, c)' could represent an English sentence such as Ann's birthday is Bill and Carlota's anniversary, but '3xFx = Ga' is incoherent. The closest English rendering of it would be something like There is a fish is Al is crazy. That function symbols stand for functions on the domain has a very important consequence. Functions are relations between objects. But two things distinguish functions from other relations. First, given some input to the function, we always obtain an output. Any object stands in the appropriate relation to some object. Second, given an input, we always obtain a unique output. Each object stands in the appropriate relation to only one object. Together these requirements imply that, given any input, we must obtain exactly one output. We may refer to the first requirement as the existence requirement; the second, as the uniqueness requirement. Because function symbols stand for functions, we may use a function symbol to represent an English expression only when these requirements, intuitively speaking, are satisfied. Mathematical expressions are easiest to evaluate in light of these requirements. Consider addition. Any two numbers have exactly one sum, so the requirements are satisfied. Or, consider multiplication. Any two numbers have a unique product, so multiplication is a function, and we may represent it with a function symbol. Subtraction and division are slightly more complex. If we include negative numbers, then subtraction is a function: any two numbers, taken in a given order, have a unique difference. Without negative numbers, however, the existence requirement fails; no positive number is 3 - 5. Division is almost a function, but not quite, even if we include fractions, since, for any n, is undefined. It is a function only on the nonzero numbers. Satisfying these requirements is important, because, armed with a function symbol, we can prove existence and uniqueness. The use of a function symbol thus presupposes them. To see this, we need to know how to handle function terms within quantification theory. We need no new rules. Recall our definition:
*
Definition 6.1 A formula A[t/v] is an instance of a formula 3vA or \;IvA iff it results from dropping the main connective of that formula and substituting a closed term t for every occurrence of the variable v throughout A. We can form instances, then, by dropping quantifiers and substituting closed function terms as well as constants for variables. Both 'Fa' and 'Ff(a}' count as instances of '3xFx'. Since \;IE and 31 speak of terms in general, \;IE and 31 both allow us to work with any instance of a formula. Universal exploitation permits us to deduce from a universal formula any of its instances, while existential introduction allows us to deduce an existential from any of its instances. We cannot use function terms in applying our existential exploitation rule or our universal proof method, however. Existential exploitation requires that we use a constant completely new to the proof in taking an instance. Universal proof similarly allows us to prove a universal formula by proving an instance with a new constant. Whenever our rules have required a new constant, then, they will continue to do so: 3E and universal proof all require a new constant.
132
Logic, Sets and Functions
It's easy to see that we can prove the existence and uniqueness of values of functions, given any set of inputs in the domain. The existence requirement states that a function must yield a value for any input. We can symbolize this as 'Vx3yy = f(x)'. This is valid, as the following proof shows:
1. ~ \fx3yy=f(x) 2. [~ 3yy=f(a) 3. f(a)=f(a) =1 31,3 4. 3yy=f(a) Notice that, in applying 31 in the proof, we substituted a variable for a closed function term. The instance of the relevant quantified formula resulted from substituting a closed function term rather than a constant for the quantified variable. Uniqueness is almost as easy to prove valid. The uniqueness requirement states that for any input there is a unique output to the function. We can symbolize this by construing it as stating that there is at most one output to the function, given any input: 'VxVyVz«y = f(x)&z = f(x)) -+ y = z)'.
1. 2. 3. 4. 5. 6. 7. 8.
~ ~ ~
\fx\fy\fz«y=f(x)&z=f(x))--+y=z) \fy\fz«y=f(a)&z=f(a))->y=z) \fz«b=f(a)&z=f(a))->b=z) ~ «b=f(a)&c=f(a))--+b=c) (b=f(a)&c=f(a)) b=f(a) c=f(a) b=c
ACP
&E,5 &E,5 =E,6,7
To see why observing these requirements in translation from natural language is so important, consider these arguments. 2. Your mother is your parent. Your father is your parent. :. Your mother is your father. Although we can reasonably translate mother and father as function symbols, we can't so translate parent; among humans, parents come in pairs. Parent thus violates the uniqueness requirement. Argument (2) is certainly invalid. But it would appear valid if we were to render parent as function symbol. All that we need to establish the validity of the argument form is the principle of the indiscernibility of identicals, expressed in the rule = E:
1. m(a)=g(a) A 2. f(a)=g(a) A 3. ~ m(a)=f(a) 4. [m(a)=f(a) =E,1,2 To see the trouble that arises if we ignore the existence requirement, consider this terrible argument, surely belied by the baby boom: 3. No babies have children. :. There are no babies. We shouldn't translate child of as a function symbol, since it fails to meet either requirement. Many people have no children, while others have more than one. If we nevertheless symbolize the argument using a function symbol, the resulting argument form will be valid:
Identity and Function Symbols
1. 'v'x(Bx--ayy=g(x» 2. ~ ...,3xBx 3. 3xBx 4. Ba 5. (Ba-...,3yy=g(a» 6. ...,3yy=g(a) 7. 'v'y""y=g(a) 8. ""g(a)=g(a) 9. g(a)=g(a)
133
A
AlP 3E,3
'v'E,l -E,5,4 QN,6 'v'E,7 =1
Using function symbols in translation is legitimate, strictly speaking, only if the existence and uniqueness requirements are satisfied. This is why many mathematicians demand a proof of existence and uniqueness before they admit a function or operation symbol into a mathematical theory. Nevertheless, we can relax these requirements somewhat when translating natural language arguments. Many function symbols satisfy existence and uniqueness on only a portion of the domain. If we use' l' to symbolize the Social Security number of, for example, we must restrict its application to constants denoting people. The domain will contain both people and numbers, and we don't want to commit ourselves to talking about the Social Security numbers of numbers. This is an informal restraint; we could impose it formally by restricting our logical rules when we use function symbols. Alternatively, we might let any intuitively silly function term, such as the Social Security number of 17, name some specially designated "dummy" object.
Problems Symbolize these sentences in Q with identity and function symbols. 1. A man's judgment and his conscience is the same thing. (Thomas Hobbes) 2. Our birth is but a sleep and a forgetting. (William Wordsworth) 3. The painter's brush consumes his dreams. (William Butler Yeats) 4. If the mind of the teacher is not in love with the mind of the student, he is simply practicing rape, and deserves at best our pity. (Adrienne Rich) 5. Nobody loves me but my mother, and she could be jivin' too. (B. B. King) 6. A foIl despises his father's instruction. (Proverbs 15:5) 7. The friend I can trust is the one who will let me have my death. The rest are actors who want me to stay and further the plot. (Adrienne Rich) 8. If you see in any situation only what everybody else can see, you can be said to be so much a representative of your culture that you are a victim of it. (S. 1. Hayakawa) 9. In a hierarchy every employee tends to rise to his level of incompetence. (Laurence J. Peter) 10. Common sense holds its tongue. (Proverbs 10:19) 11. Not all those who know their minds know their hearts as well. (La Rouchefoucauld) 12. All that we send into the lives of others comes back into our own. (Edwin Markham) 13. The cleverly expressed opposite of any generally accepted idea is worth a fortune to somebody. (F. Scott Fitzgerald)
Logic, Sets and Functions
132
It's easy to see that we can prove the existence and uniqueness of values of functions, given any set of inputs in the domain. The existence requirement states that a function must yield a value for any input. We can symbolize this as 'Vx3yy = f(x)'. This is valid, as the following proof shows:
I. ~
'v'x3yy=f(x)
2. ~~ 3yy=f(a) 3. f(a)=f(a) 4. 3yy=f(a)
=1
31,3
Notice that, in applying 31 in the proof, we substituted a variable for a closed function term. The instance of the relevant quantified formula resulted from substituting a closed function term rather than a constant for the quantified variable. Uniqueness is almost as easy to prove valid. The uniqueness requirement states that for any input there is a unique output to the function. We can symbolize this by construing it as stating that there is at most one output to the function, given any input: 'VxVyVz((y = f(x)&z = f(x)) ----. y = z)'.
I. ~
2. 3. 4. 5. 6. 7. 8.
'v'x'v'y'v'z((y=f(x)&z=f(x))--+y=z) 'v'y'v'z((y=f(a)&z=f(a))--+y=z) 'v'z((b=f(a)&z=f(a))--+b=z) ((b=f(a)&c=f(a))--+b=c) (b=f(a)&c=f(a)) b=f(a) c=f(a) b=c
~ ~ ~
ACP &E,5 &E,5 =E,6,7
To see why observing these requirements in translation from natural language is so important, consider these arguments. 2. Your mother is your parent. Your father is your parent. :. Your mother is your father. Although we can reasonably translate mother and father as function symbols, we can't so translate parent; among humans, parents come in pairs. Parent thus violates the uniqueness requirement. Argument (2) is certainly invalid. But it would appear valid if we were to render parent as function symbol. All that we need to establish the validity of the argument form is the principle of the indiscernibility of identicals, expressed in the rule = E:
1. m(a)=g(a) A 2. f(a)=g(a) A 3. ~ m(a)=f(a) 4. [m(a)=f(a) =E,1,2 To see the trouble that arises if we ignore the existence requirement, consider this terrible argument, surely belied by the baby boom: 3. No babies have children. :. There are no babies. We shouldn't translate child of as a function symbol, since it fails to meet either requirement. Many people have no children, while others have more than one. If we nevertheless symbolize the argument using a function symbol, the resulting argument form will be valid:
Identity and Function Symbols
1. 'Vx(Bx-+.3yy=g(x» 2. ~ .3xBx 3. 3xBx 4. Ba 5. (Ba-+..,3yy=g(a» 6. ..,3yy=g(a) 7. 'Vy"'y=g(a) 8. 'g(a)=g(a) 9. g(a)=g(a)
133
A
AlP 3E,3
'VE,1 -+E,5,4 QN,6 'VE,7 =1
Using function symbols in translation is legitimate, strictly speaking, only if the existence and uniqueness requirements are satisfied. This is why many mathematicians demand a proof of existence and uniqueness before they admit a function or operation symbol into a mathematical theory. Nevertheless, we can relax these requirements somewhat when translating natural language arguments. Many function symbols satisfy existence and uniqueness on only a portion of the domain. If we use' l' to symbolize the Social Security number of, for example, we must restrict its application to constants denoting people. The domain will contain both people and numbers, and we don't want to commit ourselves to talking about the Social Security numbers of numbers. This is an informal restraint; we could impose it formally by restricting our logical rules when we use function symbols. Alternatively, we might let any intuitively silly function term, such as the Social Security number of 17, name some specially designated "dummy" object.
Problems Symbolize these sentences in Q with identity and function symbols. 1. A man's judgment and his conscience is the same thing. (Thomas Hobbes)
2. Our birth is but a sleep and a forgetting. (William Wordsworth) 3. The painter's brush consumes his dreams. (William Butler Yeats) 4. If the mind of the teacher is not in love with the mind of the student, he is simply practicing rape, and deserves at best our pity. (Adrienne Rich) 5. Nobody loves me but my mother, and she could be jivin' too. (B. B. King) 6. A foIl despises his father's instruction. (Proverbs 15:5) 7. The friend I can trust is the one who will let me have my death. The rest are actors who want me to stay and further the plot. (Adrienne Rich) 8. If you see in any situation only what everybody else can see, you can be said to be so much a representative of your culture that you are a victim of it. (S. 1. Hayakawa) 9. In a hierarchy every employee tends to rise to his level of incompetence. (Laurence J. Peter) 10. Common sense holds its tongue. (Proverbs 10:19) 11. Not all those who know their minds know their hearts as well. (La Rouchefoucauld) 12. All that we send into the lives of others comes back into our own. (Edwin Markham) 13. The cleverly expressed opposite of any generally accepted idea is worth a fortune to somebody. (F. Scott Fitzgerald)
134
Logic, Sets and Functions
14. A clever man is wise and conceals everything, but the stupid parade their folly. (Proverbs 13:16) 15. The man with no inner life is the slave of his surroundings. (Henri Frederic Amiel) Use natural deduction to verify the validity of these arguments. 16. Nobody's perfect. Consequently, nobody's mother is perfect. 17. Tanya and Tatiana are the same person. So Tanya's Social Security number is the same as Tatiana's. 18. Everybody is his or her own best friend. Therefore, each person's best friend is their best friend's best friend. 19. Suppose that the name of the name of a thing is always just the thing's name itself. Then it follows that anything that names anything is a name of itself. 20. Suppose that anything that names anything is a name of itself. It follows that the name of the name of a thing is always just the thing's name itself. 21. For every man there is a priest who is his intermediary with God. Every priest is a man. No one is intermediary between God and God. Therefore, God is neither man nor priest. 22. Augustus is Superman's father. Augustus knows everybody. Somebody can beat up Superman if and only if he or she can do anything. All failure results from a lack of self-knowledge. Therefore, Superman's father can beat him up. 23. No matter who you are, the being with the ultimate power over your actions is God. One person has responsibility for another's actions just in case that person has ultimate power over the other's acts. A person can be blamed, from a moral point of view, only if he or she has responsibility for his or her own actions. Thus nobody but God can be morally blamed. 24. Nothing is demonstrable unless the contrary implies a contradiction. Nothing that is distinctly conceivable implies a contradiction. Whatever we conceive as existent, we can also conceive as nonexistent. There is no being, therefore, whose nonexistence implies a contradiction. Consequently there is no being whose existence is demonstrable. (David Hume) [Note: this is really two arguments: (i) Nothing that is distinctly conceivable implies a contradiction. Whatever we conceive as existent, we can also conceive as nonexistent. :. There is no being whose nonexistence implies a contradiction. (ii) There is no being whose nonexistence implies a contradiction. Nothing is demonstrable unless the contrary implies a contradiction. Consequently there is no being whose existence is demonstrable. To give these a fair hearing, assume that the nonexistence of a thing is the contrary of its existence.] Using natural deduction, show that these argument forms are valid. 25. VxFx;:. Vx(Gxa 26. Vx3y3zx
=
~
Fh(x, a».
f(y, z);:. VxVyGf(x,y)
~
VxGx.
27. VxVyx = y;:. Vxx = f(x). 28. Vxx
= f(a);:. Vxa = f(x).
29. Vxg(J(x» = x;:. VxVy(J(x)
'7
y ~ g(y) = x).
Identity and Function Symbols 30.. VxVy(f(x)
135
= y -> g(y) = x); .. Vxg(f(x)) = x.
31. VxFf(x)x; .. VxFf(f(x))f(x).
32. 3x(x
1- a&Gx);:.
3xGx&(Ga
3x3y(Gx&Gy&x
->
1- y))
33. Vxx = f(x); .. Vxf(x) = f(f(f(x))).
= y);:. Vx3yx = f(y)· * Suppose there are functions f and g such that (a) Vxf(x,O) = x, (b)Vxf(x, g(x)) = 0, and (c) VxVyf(x,y) = f(y,x). Show that 3xx = g(x) and that 3xf(x,x) = x. * Suppose that there is a binary function L satisfying the axioms (a) VxVyy = x£y and (b)
34. 3x3y(x 35. 36.
1- y&Vz(z = x V z = y)); VxVy(f(x) = f(y)
->
x
VxVyx£y = y£x. Show that the domain of such a function must include at most one object: VxVyx = y. Leibniz developed a calculus concerning the combination of concepts using a single function for combination and a single predicate "is in" or "is contained by", defined in terms of the combination function. Using Leibniz's symbol '0' for combination and '<' for "is in", we can present Leibniz's theory in one definition and two axioms. Definition. VxVy(x < y ..... 3zx 0 z = y) Axiom 1. VxVyx 0 Y = Y 0 x Axiom 2. Vxx 0 x = x Leibniz proceeds to prove the following propositions. Show that they follow from these axioms, together with the associativity principle: Axiom 3. VxVyVzx 0 (y 0 z) = (x 0 y) 0 z
39.
* Vxx < x * VxVyVz(x < y -> zox < zoy) * VxVy(xoy = x -> y < x)
40.
* VxVy(x <
41.
45.
* VxVyVz«x < y&y < z) -> x < z) * VxVyVz(x 0 y < z -> y < z) * VxVyVz(y < z -> y < zox) * VxVy«x < y&y < x) -> x = y) * VxVyVz«x < y&z < y) -> x 0 Z < y)
46.
* VxVyVzVw«x <
37.
38.
42. 43.
44.
y
->
yo x = y)
y&z < w)
->
x
0
Z
< yow)
Chapter 7
Sets Sets are collections or groups of objects. As natural as this idea might seem to us today, the study of sets is relatively recent. A German mathematician, Georg Cantor (1845-1918), wrote a series of papers, beginning in 1874, that began the branch of mathematics known as set theory. Later work by Ernst Zermelo (1871-1953), Abraham Fraenkel (1891-1965), John von Neumann (19031957), Bertrand Russell (1872-1970), Alfred North Whitehead (1861-1947), and others demonstrated that set theory suffices for the construction of the entire body of classical mathematics. That is, every mathematical statement can be translated into a statement of set theory; furthermore, every mathematical theorem, under this translation, becomes a theorem of set theory. The study of sets is thus the study of mathematics in an extremely general form. Sets have practical as well as theoretical importance. virtually every branch of mathematics uses sets extensively. Sets similarly form part of the foundation of computer science, linguistics, advanced areas of logic, and other disciplines that emphasize rigor. It is no exaggeration to say that all these realms of inquiry presuppose set theory. In this chapter we'll present a fairly simple theory of sets. It constitutes a fragment of a much more powerful theory, Zermelo- Fraenkel set theory with urelements (ZFU), which is the set theory most commonly used by mathematicians and computer scientists (though not mathematical logicians). The language of our theory is simple. It includes all the logical symbols of QL, the language of quantificational logic with identity. It contains three nonlogical predicate symbols, a binary predicate meaning "is a member of," and one singulary predicate, meaning "is a set." It also contains a primitive symbol similar to a function symbol that is intuitively understood to pick out collections of objects. Later we'll introduce other predicates and function symbols, but all will be defined with the symbols already mentioned. The predicate "is a set" and the function symbol defining collections, therefore, allow us to define any other idea used in our version of set theory or in the whole of mathematics.
7.1
Extensionality
The theory we'll present in this chapter is a system of axiomatic set theory. The rules of inference of this theory will include those of our quantified natural deduction system. Unlike a system of natural deduction, however, our theory will have axioms. These axioms convert to new rules that we will be using. They serve as starting points for the theory. All other theorems are proved by starting with the axioms and using the rules. Our theory will divide the world into two kinds of entity: sets and nonsets. We want to be able to talk about collections of grapefruit, or birds, or real numbers, or cities, or other sets. Because we want our theory to be as general as possible, we won't worry about what, specifically, the nonsets are. We'll refer to them as urelements, the basic objects that 137
Logic, Sets and Functions
138
we will count as given. Our theory will have little to say about these objects. It will take them as unanalyzable, and, in particular, as not being decomposable into elements in the way sets are. We will introduce the singulary predicate S to mean "is a set". The formula Sa thus says that a is a set; -,Sa says that a is not a set, that is, that a is an urelement. The most basic concept of set theory is membership. The binary predicate E , read "is a member of", "is an element of" or "belongs to", is the only other undefined predicate in the language ZL. We'll write this symbol between its arguments. Thus, a E b means that a belongs to, or is an element of, b. To abbreviate -,a E b, we'll write a t/. b. Sets, in general, decompose into elements; urelements don't. So, whenever we know that something is an urelement, we can conclude that it has no members. Equivalently, if something has members, it must be a set. So we'll allow ourselves to apply the rules: Sets(S)
-,Sx Vyy rt x
S
Sets(S) yEx Sx
S
Sets, as we've said, are collections of things. The order in which these things appear is irrelevant; so is any repetition. The only thing that makes a difference, in defining a set, is what objects are members of the set. This is the content of the first axiom of set theory, the axiom of extensionality.
Axiom 7.1 (Extensionality) Sets with the same members are identical. Symbolically:
Axiom of Extensionality VxVy((Sx&Sy)
-t
(Vz(z E x
<->
z E y)
-t
X
= y»)
This axiom gives us an identity criterion for sets. We have written our axiom of extensionality so that it rules out identifying urelements by means of their elements. If we had not, that would have had disastrous consequences for our theory of urelements: if extensionality held of urelements, there would only be one urelement! That is, we have taken a bold stance and claimed that all objects except for urelements are like sets in that their identity is defined by their members. How do we determine whether set a is the same as set b? We look at their members. This might seem like a trivial matter. But notice that the same collection may be described in very different ways. We might describe the collection containing just the two cities Pittsburgh and Philadelphia, for example, as the set of the two largest cities in Pennsylvania; as the set of Pennsylvania cities with metropolitan area populations of over two million; as the set of cities with National League pennant-winning teams in 1909 and 1915; as the set of cities with baseball teams finishing first in the National League East between 1974 and 1979; and so on. No matter which description we use, the set described is the same. Extensionality will yields a useful rule of inference, which we'll call extensionality exploitation. This rule allows us to replace an expression of the form Vz(z E a f--> Z E b) with one of the form a = b, provided that a and b are sets.
Sets
139
Extensionality Exploitation( Ext)
Sa Sb Vz(z E a .-.. z E b) a= b
Ext
Of course, we can perform the converse inference, from a = b to Vz(z E a .-.. z E b), solelY.rusiI!g identity exploitation. ~ Membership is the most important relation between collection. But there are other very frequentlyused notions that can be defined in terms of membership. Definition 7.1 Something is a subset of another iff it is a set and every member of the it is also a member of the other. Symbolically,
VxVy(x
<;:::
Y f-> (Sx&Sy&Vz(z E x
~
.. ".
z E y))
The rules we derive from this are as follows :
<;:::E x<;;;y Sx&Sy&Vz(z E x
<;;;E ~
z E y)
<;;;E
<;;;E -.Sx (or-.Sy) -.x <;;; y
<;;;E
Definition 7.2 One collection is a proper subset of another iff the first is a subset of the second, but is not identical to it. In symbols,
VxVy(x C Y .-.. (x <;;; y&x
=t y))
Definition 7.3 One set is a superset of another set iff the second is a subset of the first. In symbols,
VxVy(x ;2 y .-.. y <;;; x) Note that every set is a subset of itself, since every member of a set is, of course, a member of that set. For the same reason, every set is a superset of itself. But no set is a proper subset of itself. This means that the subset and superset relations are reflexive. The proper subset relation, in contrast, isirreflexive. To see how a simple proof in set theory works, let's show that every set is a subset of itself. We'll begin with a formal proof, and then present an informal translation.
140
Logic, Sets and Functions
Theorem 7.1 Every set is a subset of itself. Vx(Sx
1.
~
Vx($x-+xCx)
2.
~
($a-aCa)
3.
$a
4.
~
5.
6.
k~
~
(x <;;; x))
ACP VZ(ZEa-+ZEa) (bEa-bEa)
ACP
$a
R,3
8.
($a&$a)
&1,3,7
9.
(( $ a&$a)& VZ(ZE a-+ZE a»
&1,8,4
aCa
~,9
7.
1O.
(Informally: Every set is a subset of itself. We must show that an arbitrary set a is a subset of itself; this requires us to show that every member of a is a member of a. But this is trivial.) It's easy to confuse the membership and subset relations. But they are very different. The subset relation is reflexive; as we'll see later, membership is irreflexive. No set is an element of itself. The set Z of integers is a subset of the set R of real numbers, but Z is not a member of R. (The set of integers is not itself a real number.) Similarly, the number 1 is a member of R, but is not a subset of R. To take some more ordinary examples, each of you is a member of the set of logic students, but no one of you is a subset of that set. None of you are sets at all! The set of logic students, in contrast, is a subset of the set of students, but isn't a member of the set. The set of students has only people, not other sets, as members. To see a more complex proof in set theory, let's show that two sets that are subsets of each other are identical. Again, we'll present a formal, and then an informal, proof.
Theorem 7.2 Two things that are subsets of each other are identical. VxVy((x <;;; y&y <;;; x)
y)
-+
x =
Sets
141
1.
~
VxVy«xCy&yCx)-+x=y)
2.
~
Vy«aCy&yCa)-+a=y)
3.
~
«aCb&bCa)-+a=b)
4.
(aCb&bCa)
ACP
5.
aCb
&E,4
6.
bCa
&E,4
7.
«$a&$b)&Vx(xEa-+xEb»
C,5
8.
«$b&$a)&Vx(xEb-+xEa»
C,6
9.
($a&$b)
&E,7
10.
Vx(xEa-+xEb)
&E,7
11.
Vx(xEb-+xEa)
&E,8
12.
~
13.
~
Vz(zEa-zEb) (cEa-cEb)
14.
(cEa-+cEb)
"IE, 10
15.
(cEb-+cEa)
VE,II
16.
(cEa-cEb)
-1,14,15
17.
$a
&E,9
18.
$b
&E,9
19.
a=b
Ext, 12, 17, 18
Informally: Two sets that are subsets of each other are identical. Suppose that a and b are sets and are subsets of each other. Then every member of a is a member of b, and vice versa. So something is a member of a iff it is a member of b. By the axiom of extensionality, therefore, a and b are identical. These two proofs demonstrate an important feature of proofs within set theory: Definitions and axioms furnish rules of inference. Above we've written definitions in English, in formulas of ZFU, and in the form of rules. Steps in a proof will often cite a definition or an axiom together with the number of a preceding line. This is perfectly legitimate; it represents an abbreviation of a more complete proof, with the axiom or definition introduced on a separate line. Consider, for example, this inference from the last proof: 4.a~b
6. Sa&Sb&Vz(z E a -. z E b)
&E,3 ~ E,4
We could have written this, more fully, as: 4. a ~ b 4.1 VxVy(x C eqy +-+ (Sx&Sy&Vz(z E x -. z E y)) 4.2 a C eqb +-+ (Sa&Sb&Vz(z E a -. z E b) 6. Sa&Sb&Vz(z E a -. z E b)
&E,3 ~E
VE 2 , &E, 4.1 +-+ E,4,4.2
So long as we know that we can fill out such steps to obtain official proofs, however, this sort of abbreviation is acceptable. There is no reason to write the proof out in the official form in which
142
Logic, Sets and Functions
every line is either an axiom, a definition or a consequence of previous lines by a rule of inference. It's helpful, therefore, to think of the axioms and definitions as rules. In most proofs, the axiom and definitions of this chapter will appear as rules of inference. The following is a useful derivable rule:
c* xEy yc;;.z xEz
.,
c;;.*
Problems
·d
From here on, please take special note of the exercises followed by an '". '!' indicates that the problem is a very useful theorem for later proofs. Prove the following theorems. 1. Every set is a superset of itself. (!)
2. No set is a proper subset of itself. (!) 3. For any sets x, y and z, ifx c;;. y and y
~
z, then x c;;. z. (!)
4. For any sets x, y and z, if x C y and y C z, then x C z. (!) 5. For any sets x, y and z, if x
2 y andy 2 z, then x
;2 z.
6. No two sets are proper subsets of each other. (!) 7. Two sets that have the same members belong to the same sets: 'v'x\fy[(Sx&Sy) -> ('v'z(z E x..-. z E y) -> 'v'w(x E W ..-. YEW))]. (This is the Bestimmheit, or determinateness, principle.) 8. The principle that sets that are subsets of each other are identical implies the extensionality axiom. (Don't use extensionality in your proof!) 9.
7.2
* Say that a set is
transitive iff all members of its members are also members of it: \fx(Sx -> (xis transitive..-. 'v'y\fz((y E z&z E x) -> Y E x))). Then a set of sets a, LeVx(x E a -> Sx), is transitive iff all members of it are subsets of it.
Abstraction
We now know when two sets are equal to each other. The axiom of extensionality gives us an identity criterion for sets. It tells us that sets are determined by their members. But we still know very little about what sets and in particular what sets are. We now need to explore ways of talking about sets that will lead us finally to distinguish between sets and sets. One way we might talk about a particular set is to list its members. At times, however, it is very inconvenient to list the members of a set in order to be able to talk about it. Suppose we want to talk about the members of Congress. We could list everyone who is a member of Congress and surround this list with braces. But this would be laborious. Normally, we just speak of the set or collection of members of Congress. That is, we don't list the members of the set. We instead provide a description or rule that specifies the members of the set.
Sets
143
This tactic is critically important when we speak of infinite sets, such as the set of prime numbers or the set of algorithms. In these cases, we couldn't write out a full list of the elements; if we tried, we would never finish. Providing a description or rule constitutes the only direct way of referring to an infinite set. Mathematicians tend to use this method, the method of abstraction, in the following form. They specifY the set of prime numbers as {x : x is a prime number}, or, equivalently, {xix is a prime number} or .xx(x is a prime number) . Within the braces {,} is a variable, followed by a colon, followed by an open formula. The resulting expression picks out the collection of all objects satisfying the open formula. So,{x: Fx}, "the collection of all xs such that Fx", is the collection of Fs. Note that the particular variable we use is irrelevant. We may refer to the members of Congress by writing {x : x is a member of Congress}, {y : y is a member of Congress}, {z : z is a member of Congress}, etc. As this may suggest, the abstraction notation functions much as a quantifier does. Both bind variables. The abstraction operator {... : ... }, however, converts formulas into terms that pick out collections. It is therefore also similar to a function symbol, although function symbols do not bind variables. Expressions like the abstraction operator are often called variable-binding operators. So we'll count x, y and z as bound in the expressions above. To be justified in using abstraction, no matter what open sentence we had in mind, we would have to know that, for any open sentence, there is a set of objects satisfying it. This seems like a natural principle to adopt. Indeed, Cantor and Gottlob Frege (1848-1925), the German logician and co-inventor of quantificationallogic, did adopt this "naive" or "unrestricted" principle of abstraction. Unfortunately, it leads to contradictions. Bertrand Russell devised a predicate for which there is no corresponding set: is not a member of itself We might think that some sets, such as "the set of sets," belong to themselves, although most sets, such as the set of elephants, obviously don't. (The set of elephants, after all, is not itself an elephant.) According to the naive principle, we should be able to speak of the set of sets that are not members of themselves. But is this set (call it A) a member of itself or not? If it is, then A must be one of the sets that are not members of themselves; so A is not a member of itself. Suppose, then, that A is not a member of itself. A is the set of all such sets; so A must be a member of itself. Forming an expression to refer to the set of objects satisfying the predicate 'is not a member of itself', therefore, leads to a contradiction. To see this reasoning in symbolic form, 1. 3x(Sx&V'y(y Ex+-; Y ¢ y» Naive abstraction 2. Sa&V'y(y E a +-; y ¢ y) 3E,1 3. a E a +-; a ¢ a &E,V'E,2 4. (a E a&a ¢ a) V (a ¢ a&a E a) 3, +-; V Modern mathematicians therefore adopt a restricted form of the axiom of abstraction. Abstraction is a very powerful method for creating sets; left unrestricted, it is too powerful. But mathematics would be virtually impossible without it. Zermelo solved this problem by allowing abstraction within the boundaries of an already-created set. That is, we can weaken the naive principle to the assertion that, given an open formula and a set, there is a set containing just those entities in the given set that satisfy the open formula. Recall that {x : A(x)} represents the set of As, the things satisfying the open formula A. As Russell's paradox shows, we cannot assume that there is such a set for every open formula. Instead, we must restrict abstraction by applying it only within an already-constructed object. We should always use the abstraction notation in the form {x E a: A(x)}, where a is either an ur-element (in which case the constructed set is empty) or some already-created set (in which case the constructed set is a subset of a ). The expression {x E a : A(x)} represents the set of things in a which are As. In most applications of set theory, there is a set U which serves as the "universe of discourse" for that application. All abstraction terms can then be restricted to this set U: e.g., {x E U : A(x)} .
144
Logic, Sets and Functions
In such cases, mathematicians simply omit explicit mention of the restrictor term and proceed as if we were using naive abstraction. Whenever there is such a universe of discourse, such uses of naive abstraction will not result in any contradictions. Throughout the rest of this book, all applications of set theory will in fact involve such universes of discourse. Thus, despite the inconsistency of naive abstraction when used without restriction, we will simply take it for granted that no such inconsistency will appear. This will enable us to use the simple definition of naive abstraction terms and the simple naive abstraction rule.
Naive Abstraction Terms: If v is a variable, t is a closed term, and A is a formula not containing v, then: {v : A[vlt]} is a closed term.
The abstraction notation uses an expression, {v : A[vIt']}, that is very similar to a function symbol, although it binds variables. Expressions of this kind are often called variable-binding operators. Like function symbols, we can introduce such operators legitimately only if we can show that there is a unique object answering to each term that results. To guarantee this, we must adopt another axiom of set theory. Axiom 7.2 (Naive Abstraction axiom schema) S{z: A(z)}&'
E
{z : A(z)}
<-t
A(y))
This is not an axiom but an axiom schema, that is, a rule for producing axioms. Replacing 'A(z)' with any open formula with the single free variable 'z' yields an axiom. Consequently, there are really infinitely many axioms of abstraction; there is one for each open formula with a single free variable, and there are infinitely many of these. In this chapter we'll generally restrict the open formulas that can enter into the schema to the language of set theory itself. The following theorem-schema and inference rule will be useful: Theorem 7.3 S{z : A(z)} Naive Abstraction v E {z : A(z)}
Abst
A(v)
Abst
To see how the abstraction rule works, let's prove the simple theorem that any set x is identical to the set of x's self-identical members, {y : y E x&y = y}. Theorem 7.4 Any set is the set of its self-identical members. Proof: we want to show that any set is identical to the set of its self-identical members. We must show that (i) for any set a the collection of a's self-identical members is a set; (ii) that a and the set of a's self-identical members have the same members. The first follows immediately from Theorem 7.3, and the second follows from the abstraction rule together with the reflexivity of identity (=1). Then, by extensionality, a is identical to the set of its self- identical members. Note that this proof uses a strategy we've already seen. To show that two sets are identical, it is often necessary to use the axiom of extensionality. The proof therefore proceeds by showing that the two sets have the same members. This strategy will apply again and again throughout the problems in this and future chapters. We can use the principle of abstraction in its more sophisticated form
Sets
145
(not naive abstraction) to show that there is no universal set, no set which contains absolutely every set there is. Suppose that U is the universe of discourse of a particular application of set theory. We can show that there is set which does not belong to U, namely, the Russell set {x E U : x (j. x}. 1. To show {x E U : x (j. x} (j. U: 2. Assume (for IP) that {x E U: x (j. x} E U 3. Show {x E U : X (j. x} (j. {x E U : x (j. x} 4. Assume, for another indirect proof, that {x E U : x (j. x} E {x E U : x (j. x} 5. {x E U: x (j. x} E U&{x E U: x (j. x} (j. {x E U: x x} by Abstraction, from line 4. 6. {x E U : x x} {x E U : x x}, by &E on line 5.
rt
rt
rt
rt
rt
rt
7. Line 6 contradicts line 4. Hence, we may conclude that {x E U : x x} {x E U : x (j. x}, by indirect proof. 8. {x E U: X (j. x} E U&{x E U: X (j. x} {x E U: x (j. x}, by &1,2,7. 9. {x E U: x (j. x} E {x E U: x x}, by abstraction from line 8. 10. Lines 9 and 2 are contradictory. Hence, we have completed our indirect proof, begun on line 1.
rt
rt
Problems 1. Assuming that Sa, prove that {z : z E a} ~ a.
2. Suppose that, for any x, {A(x} only if S(x}. Show that {y : A(y)} ~ {z : S(z)}. 3. Suppose that, for any x, A(x} iff S(x}. Show that {y : A(y}}
7.3
= {z : S(z)}.
Pair Sets, Unit Sets and Enumeration
Earlier, we mentioned that we might talk about a set by listing all its members. In this chapter we make this notion more precise. Often, sets are designated by expressions such as {Pittsburgh, Philadelphia}, {Fred} and {Washington, Adams, Jefferson, Madison, Monroe}. Intuitively, the braces, { and }, indicate that we are speaking of a set; inside the braces appears a list of the members of the set. Since there is nothing more to the identity of a set than its elements, an expression of this form properly specifies a set. Using the naive abstraction axiom schema, it is quite simple to define unit sets and pair sets. If we were being more careful and used only restricted abstraction, we would have to appeal to special axioms to assert the existence of such sets.
Definition 7.4 (Unit Set) 'v'x({x}
= {y: y = x}}
Definition 7.5 (Pair Set) 'v'x'v'y( {x, y} = {z : z = x
V
z = y}}
Corresponding to these definitions are the following inference rules: Pair Set U
E {y, z}
u=yVu=z Unit Set
x E {y} x=y
Unit Unit
Pair set Pair set
146
Logic, Sets and Functions
With our axiom schema of abstraction and axioms of unit sets and pair sets, we can now show the following theorems:
Theorem 7.5 VxVyS{x, y} Theorem 7.6 VxS{x} For a given number of objects al, . .. ai, it is clear that there is a set consisting of just those objects. This invites the following sort of generalization of the pair set definition:
Definition 7.6 (Enumeration definition schema) VYl,".,Yi({Yl, ... ,Yi} X=Yi)})
= {x:
(x
= Yl v .... V
The following inference rule corresponds to this axiom schema: Enumeration
x E {Yl, ... ,yd x = Yl V X = Y2 V ... V x = Yi
Enum Enum
Note that this is a definition schema, not a definition. That is, we must have in advance some determinate list of constants, variables or function terms in order for the notation above to constitute a proper formula of our set theoretic language. But in practice we will be sloppy and use such expressions as {Yl, .. ·Yi} without specifying what the y's are. We thus can justify our intuitive practice of referring to any finite set by listing its elements, and surrounding the list with braces. We end this section with three theorems:
Theorem 7.7 Any object belongs to its unit set:Vxx E {x}. Theorem 7.8 Any object belongs to any of its pair sets: VxVyx E {x, y}. Theorem 7.9 VXiXi E {Xl, ... ,Xi,'''Xn } (The proofs are left as exercises.)
Problems Prove the following theorems. 1. Everything has a unit set.
2. Say that an object is ultimate iff it is not a member of any set. Then there are no ultimate objects. 3. Prove Theorem 7.7. 4. VxVy{x,y} = {y,x}. (!) 5. VxVy(Sy -- (x E Y ~ {x} ~ y»). (') (Generalize this to sets of more than one element). 6. VxVy(x
= {y}
~
Vz(z E x
~
z = y)).
7. Two objects are identical if their unit sets are identical. (!)
Sets
147
8. Unit sets are unique; that is, given any object x, any two sets having x as their only member are identical.
9. Vx({{x}}
= {x}
-7
X
EX).
10. Prove theorems 7.8, 7.9. 11. 12.
* Assume that a E b * For
-7
b rf- a and that {{a}, a}
= {{b},b}.
Then a
= b.
any four objectsx,Y,z and w, {x,y} = {z,w} only if either (x = z&y = 111) or (x =
w&y = z). 13.
* Two sets are identical just in case they belong to the same sets.
14.
* Prove that {a}
15.
* Show that, for any sets x
7.4
The Null Set
= {y : y =
a}. and y, {x,y} = {z: (z = x V z = y)}.
Sets are groups or collections of things. We determine their identity by looking at the things in them, their members. We know how to tell whether sets are the same or different; given some urelements or sets, we know how to form sets by constructing pairs or subsets whose elements satisfy open sentences. Given the abstraction axiom, we can prove that there is a set without any members. Since we want to construct a set with no elements, we need to specify a condition on membership in this set that nothing could fulfill. One such condition is expressed by the open formula 'z -:F z'. We can require, in other words, that only objects that are not identical to themselves can belong to the set. Since there are no such things, nothing will be in the set.
Theorem 7.10 There is a set without members. 3x(Sx&Vyy
rf- x)
, Prbof: By the axiom of abstraction, there is a set containing just those elements of the universe I of disC0urse that are not self- identical; namely, the set {x : x #- x}. Since everything is selfidentical, nothing belongs to {x : x #- x}. So there is a set with no elements. We'll call the set without members '0'. Definition 7.7 (The Null Set) 0 = {x: x
=1=
x}
Given our definition of 0 and our rules for abstraction, we can show:
Theorem 7.11 S0&Vyy
rf- 0.
In proofs, this is the fact about the null set that will be most useful. That is, rather than adopt an official definition of the null set, we will usually use the symbol '0' in steps of the form:
148
Logic, Sets and Functions
So far we know that there is a set without members. Our use of the constant '0' for the null set, however, needs further explanation; we need to know that there is only one set without members. We need to know, in other words, that the null set is unique. Luckily, we can prove that the null set is unique. Theorem 7.12 0 is the only empty set. \/x((Sx&\/yy (j x) --., x = 0)
The proof uses the axiom of extensionality:
1, ~ 2.
Vx(($x&Vyy¢x)-+x= 0)
~
(($a&Vyy¢a)-+a= 0)
3,
($a&Vyy¢a)
4.
~
5.
6.
ACP
VZ(ZEa~ZE 0)
~
~
(bEa~bE 0)
(bEa-+bE 0)
7.
bEa
ACP
8.
Vyy¢a
&E,3
9.
b¢a
VE,8
10.
-,bEa
E¢,9
11.
bE
0
!,7, 10
12.
~
13.
bE 0
14.
b¢
15.
-,bE
16,
bEa
17.
(bE 0-+bEa)
ACP
o
0 0
(bEa~bE0)
E¢,14
!, 13, 15 ~I,6,
12
18,
$a
&E,3
19.
$0
Th 10
20.
a= 0
Ext,4,l 8, 19
In English: To show that any set without members must be identical to 0, assume that a is set without elements. Everything in a is in 0, since nothing is in a; similarly, everything in 0 is in a. So a and 0 have the same members (namely, none); by the axiom of extensionality, therefore, they are identical. Another very useful theorem about sets that we add without proof (the proof is left as an exercise) is the following: Corollary 7.1 \/x((Sx&x =I- 0) --., 3yy E x)
We have shown, then, that there is one and only one set without any elements. The null set, introduced by George Boole, has proved very useful in mathematics, much as the development of arithmetic with zero made possible tremendous advances in numerical computation.
Sets
149
Problems Prove these theorems, where a and b are sets. 1. The null set is a subset of every set. (!)
2. The null set is a superset only of itself.
3. {x 4.
E a:
x =j:. x} = {y
E
b: y =j:. y}.
0 is not identical to the unit set of any set.
5. For any two sets x and y, if x is a subset of the unit set of y, then either x is the unit set of y, or x is the null set. In symbols, VxVy(x ~ {y} f--> (x = {y}vx = 0)). 6. There are no unicorns. So, {x 7. No squares are round. So, {z
7.5
E E
a : X is a unicorn} = 0.
a: zis a round square} = 0.
Binary Unions, Intersections and Complements
You may be familiar from high school mathematics with the union of two or more sets. Graphically, the union of two circles x and y, construed as sets of points, is the area shaded in this diagram:
y
The union of two sets is the set of all the things in one or the other. The union of several sets is similarly the set of things in at least one of these sets. We'll write the union of x and y as xU y. We can define this set informally in the following way: Definition 7.8 (Binary union) xU y = {z : z E x V z E y} It then follows from the Union axiom that xU y is a set. We can also prove the following fact abouty U z:
Logic, Sets and Functions
150 Theorem 7.13 'r/y'r/z'r/u(u E y U Z
f-4
(u E Y VuE z))
Proof:
1. ~
2. 3. 4. 5.
V'yV'zV'u(uE(yUz)-(uEyvuEz» V'zV'u(uE(aUz)-(uEavuEz)) V'u(uE(aUb)-(uEavuEb» (cE(aUb)-(cEavcEb» ~ (cE(aUb)-(cEavcEb» cE(aUb) [ (cEavcEb) ~ «cEavcEb)-cE(aUb» (cEavcEb) [ cE(aUb) (cE(aUb)-(cEavcEb»
~ ~ ~
6. 7. 8. 9. 10. 11.
ACP
U (Def. of Bin. Union),6 ACP
U (Def. of Bin. Union),9 -1,5,8
Theorem 7.14 justifies us in adding the following simplified inference rule:
Union(U) xEyUz
U
XEyVxEz
U
Theorem 7.14 'r/x'r/ySx U Y
Proof: obvious using the definition of union and Theorem 7.3. We can further generalize the binary function symbol U . Definition 7.9 The union of Yl Yi is the set containing all the members of Yl throughYi' In symbols, Yl U ... U Yi = {z : z E Yl V z E yd.
The corresponding rule of inference is:
Generalized Union(U n ) x E Yl U X E Yl V
U Yi
V X E Yi
un un
We will also define the familiar binary function of intersection. Suppose that we were to diagram two sets, X and Y, as ellipses:
Sets
151
x
y
The intersection of the ellipses is the area shaded above. It consists of the points that are in both regions. Analogously, the intersection of two sets is the set of things that are members of both sets. The shaded area thus represents the intersection of x and y, which we'll write as x n y. If x is the set of red things, and y the set of houses, then x n y is the set of things in both x and y: the set of red houses. Similarly, if Z is the set of integers, and R+ the set of positive real numbers, then Z n R+ is the set of positive integers. We will define the two place function, intersection, in the following way:
Definition 7.10 The intersection of x and y is the set of all members common to x and y. In symbols, VxVyxny = {z:
Z
E x&z E y}
The intersection of two sets x and y is just the set of all members of x that are also in y:{z : Z
E x&z E y}. The abstraction principle justifies the construction of a binary intersection as a
set. We will prove that binary intersections are sets using theorem 7.3. We will do this in several steps. The first will be to show that there is a set of all those objects that x and in y. We can then use abstraction to conclude that the set of all those things in x and y is a set and then we use the definition of binary intersection to conclude that x n y is a set.
Theorem 7.15 VxVySx n y Proof: Use theorem 3 and the definition of binary intersection.
Theorem 7.16 VxVyVu(u Ex n y Proof:
<-+
(u E x&u E y))
Logic, Sets and Functions
152
1.
~
2.
~
3.
~
V xVyV U( UE (Xny )~( UE X&UEy» VyVU( UE (any )~( UE a&UEy» Vu( UE (anb )~( UEa&UE b»
4.
~
5 6. 7. 8.
~
[
10.
[
11.
(cE(anb)~(CEa&CEb))
CE(anb)
ACP
(CEa&CEb)
n (Binary
~
9
(CE (anb )~( CE a&cEb»
Intersection),6
«cEa&CEb)~CE(anb»
(CEa&CEb)
ACP
cE(anb)
n (Binary
(CE (anb)~( CEa&CEb»
Intersection),9
~I,S}8
Using theorem 7.16, we can derive the following rule of inference: Binary Intersection(n) zExny z E x&z E y
n n
The intersection of two sets contains those elements the two sets have in common. It's natural to generalize this idea to more than two sets. The intersection of any number of sets is the set of elements all these sets have in common.
Definition 7.11 The intersection of Yl ... Yi is the set of all members of each of Yl ... Yi' In symbols, Yl
n ... n Yi = {z
E Yl : z E Y2&"'&Z E y;}
Again we have the equivalent rule of inference: Generalized Intersection(nn)
x
n x E Yl & E Yl
Yi
&x E Yi
nn nn
Sometimes it's useful to speak of the objects belonging to one set but not to another. We might, for example, want to talk about the cities with populations over 500,000 that have no National Football League franchises, or the states that have no fewer than 25 electoral votes. In such situations we can speak of the difference between two sets. The difference between sets Y and z (also called the complement of z in or relative to y) is the set of things that are in Y but not in z. The usual symbol for this difference is 'y - z'.
Sets
153
Definition 7.12 The relative complement of z in y is the set of members of y which are not also members of z. In symbols, y - z = {x : x E y&x
F/ z}
Using the definition of complement and Theorem 7.3, we can derive:
Theorem 7.17 VxVySx - y From the definition of relative complement and abstraction, we also derive the following rule of inference: Def. of Complement(- ) xEy-z x E y&x
F/ z
DefDef-
We have now shown that there are many ways of constructing new sets from old ones. At this point it seems appropriate to introduce some properties of sets and relations among sets used throughout mathematics: Two sets that have no members in common are said to be disjoint.
Definition 7.13 x and yare disjoint iff their intersection is empty.
Or, VxVy(x and yare disjoint
<---7
x n y = 0)
djt x and yare disjoint
xny = 0
djt djt
The set R+, for example, is clearly disjoint from the set of all negative reals, R-. The set of red houses is disjoint from the set of tricycles. And the set of symphonies and the set of ducks are disjoint. This notion, too, can be generalized to any number of sets. The obvious generalization, however, turns out to be not very fruitful. A more useful concept is this. We'll say that any set of sets is pairwise disjoint iff any two sets in the set are disjoint in the sense we've already defined.
Definition 7.14 A set x is pairwise disjoint iff any two members of x are disjoint.
Or Vx(xis pairwise disjoint
<---7
VyVz((y E x&z E x&y
i:-
z) -; yandzare disjoint)))
154
Logic, Sets and Functions PD
x is pairwise disjoint VyVz( (y E x&z E x&y
=1=
z)
-+
PD PD
y and z are disjoint)
To take an example, consider a set of several clubs, construing each club as a set of its members. The set is pairwise disjoint if any two of the clubs are pairwise disjoint. This will hold just in case nobody belongs to more than one club in the set. In such a circumstance, the intersection of the set would be empty. It's easy to see that any pairwise disjoint set of sets has an empty intersection, if it contains at least two sets.
Problems Prove the following theorems. Assume throughout that a, b, c and d are all sets. 1. aU
0 = a. (!)
2. a U a = a. (!) 3. aU b = b U a. (!)
4. aU(bUc)=(aUb)Uc. (!) ("".',...\
'. 5. jSuppose that a and b are both subsets of c. Then aU b ~ c. (!) 6. aU b;2 a.
7. Suppose that a
~
b. Then, for any set x, a U x
8. Suppose that a
~
c, and b ~ d. Then aU b ~ cUd.
9. Vx(x
~
a
-+
10. Suppose a
x
=1=
~
b U x. (!)
= a - (a - x)).
0. Then -,3xx
= a-x.
11. The null set and any set are disjoint. (!) 12. a - b
~
a.
13. ana = a. (!) 14.
an b = b n a.
15. an(bnc)=(anb)nc. (!) 16.
(a~c~-+a-b=an(c-b).
17. Suppose that a is a subset of both band c. Thena 18. (a
~ c@
-+
(a
~
(c - b) - a and bare disjoint).
19. a-b=a-(anb). 20. a - 0 = a. (!) 21. a - a
= 0.
~
(!)
.//~.
b n c.
Sets
155
22. Suppose that a
~
b. Then an c ~ b n c.
23. Suppose that a
~
b. Then, for any set x, x - a
~
x-b.
24. For any sets x, y and z, z n (x - y) = (x n z) - y.
25. Suppose that a
~
b n c. Then a and b - c are disjoint.
26. For any sets x, y and z, (x - z) n (y - z) = (x n y) - z.
27. Suppose that no cats are dogs but that all cats, and all dogs, belong to the set of animals.
Then the set of cats and the set of dogs are disjoint. 28. Suppose that a and b are unit sets. Then either a and b are disjoint, or a = b. 29. Suppose a is a unit set, b is a pair set, and an b ibelongs to a.
0. Then at least one of the members of b
30. Suppose that a and b are both subsets of c. Then b ~ c - a iff a
~
c - b.
31. Suppose that a and b are both subsets of c. Then c - b ~ a iff c - a
32.
* For any sets x
33.
* {a,b} n {a,c} is a unit set iff (b = c - b = a).
and y, x
~
y iff xny = x.
34. aU b = aU (b - a).
35. (a
~
b&b
~
c)
<->
aU b = b n c.
36. a - (b - c) = (a - b) U (a n b n c). 37. aU (b n c) = (a U b)
n (a U c). (!)
38. an(bUc) = (anb)U(anc). (!)
39. (aUb)-(anb)=(a-·b)U(b-a). 40. a = b iff aU b ~ an b. 41. a ~ b iff a U b = b.
42. {a, b} = {a}
U
{b}.
43. cU(anb)=an(bUc)iffc~a.
44. Say that a and b are disjoint, and that a U b = c. Then a = c - b.
45. (a-b)U(c-b)=(aUc)-b. 46. (a - b) U b = a iff b
~
a.
47. Suppose that aU b = aU c and that an b = an c. Then b = c.
48. an b ~ (a
n c) U (b n b - c)
49. (a U c) n (b U b - c)
~
50. (a n b) U (a - b) = a.
aU b.
~
b.
156
Logic, Sets and Functions
Definex + Y (the symmetric difference between x and y) as (x - y) u (y - x). Prove each of the following facts (for any sets x, y and z): 51. (x+y)+z=x+(y+z)
52. x
+y
= Y
53. x
+x
=
54. x
+ 0=x
55. x - Y
~
+x
0
x
+Y
56. x
+z
= Y+z
57. x
+ y = 0 +-> x = y
58. xn(y+z)
7.6
--->
x = y
= (xny)+
(xnz)
Unions and Intersections of Single Sets
Often it's useful to combine the members of the members of a set into a single, unified set. Within a corporation, for example, there might be a number of different committees. Many reports and memos will be distributed only to the members of a single committee. Some reports and memos, however, might go to all committee members, no matter what committee they happen to serve on. lf the corporation has a computerized system for internal mailings, the system will need to combine lists for the particular committees to derive a list of committee members in general. In doing this, the system will be constructing the union of the corporation, considered as a set of committees. Given any set, we can in fact form a union. The union of a set of sets is the set of all the things belonging to some member of the set. The union of a set, the members of which are not sets themselves, will be 0. In effect, then, the union of a set is the set of members of members of the set. The union of a set of annual membership lists for a club, for example, is simply the set of all those who have ever belonged to the club. The union of a set of sets of people linked to a series of fires is the set of all those people linked to at least one of the fires. To define unions of single sets within informal set theory (using naive abstraction), we can use the following:
Definition 7.15 (Union of Single Set) Vx(Ux = {z: 3w(w
E
x&z E w)}).
From the definition of union and Th. 7.3, we can derive the following theorem:
Theorem 7.18 VxS U x· We can also derive the following useful rule of inference: Singulary Union(U 1 ) x E Uy 3z(z E y&x E z)
U1
U1
157
Sets
Given that we countenance non-sets or urelements as objects, we must accept the existence of sets like {a, b}, where a and bare urelements. So we must consider what the union of two urelements (or for that matter the union of an urelement) ought to be. Our theory so far allows the union of two urelements to be any set whatever. For instance, the union of two people to be the set of all their children. Perhaps this might not seem so silly. Most mathematicians in practice, however, consider unions only of sets. If you ask them what the union of any two integers is, they'll either stare at you or more likely tell you it's the null set. We'll follow convention and show that the union of any two urelements is 0. Theorem 7.19 VxVy« -.sx&--,Sy) Theorem 7.20 Vy(--,Sy
~ Uy
-+ xU
y = 0)
= 0)
Theorem 7.21 VxVy«.Sx&Sy)
-+
xUy = y)
Here's an example of a proof using the union rule. Prove: if every member of a is a subset of b, then Ua ~ b. (Assume a and b are sets.)
1.
~
(Vx(xEa-xCb)-U(a)Cb)
2.
Vx(xEa-xCb)
3.
~
Vy(yEU(a)-yEb)
4.
~
(CEU(a)-+CEb)
ACP
5.
CEU(a)
ACP
6.
3Z(ZEa&CEZ)
U1,5
7.
(dEa&CEd)
3E,6
8.
dEa
&E,7
9.
(dEa-dCb)
VE,2
10.
dCb
-E,9,8
11. 12.
CE d
&E,7
cEb
C*, 11, 10
13.
Vx$U(x)
Union* (Th. 7.18)
14.
$U(a)
VE, 13
15.
U(a)Cb
Def. of SUbset, 14,3
Just as with union, we can speak of the intersection of any set as the set of objects that all the members of the set have in common. So, suppose that we are members of an organization that wants to celebrate its 25th anniversary by giving awards to all those members who have belonged to the organization for the entire 25 years of its existence. We are asked to compile a list of such members, and are given 25 annual membership lists. We can find out who has belonged to the organization for all 25 years by taking the intersection of the set of lists. We need, in other words, to find the names that all the lists have in common. For another illustration, suppose that we are investigating a series of suspicious fires. For each fire we construct a list of the type of building burned, its individual or corporate owners, officers
158
Logic, Sets and Functions
of owner corporations, insurance companies involved, etc. Thinking of each list as a set, we can then take the intersection of the set of sets to see if all the fires have anything in common. The intersection, in this case, will consist of the people, insurance companies, types of building, etc., linked to every fire. If the intersection contains John Smith, then he is linked to each fire in some way. Perhaps he owned one building, is an officer of a company that owned another, wrote an insurance policy for a third, and so on. To write this in symbols, we must use our abstraction notation. Definition 7.16 The intersection of a set x is the set of those members common to each member ofx. In symbols, nx = {y : Vz(z
E---+
y
E
z)}
The definition of intersection ensures that the intersection of a set (or an urelement) is always a set. Theorem 7.22 VxS n x Proof: use theorem 7.3 and the definition of intersection (details left as an exercise). Abstraction and the definition of intersection allows us to to derive the following rule of inference for intersections: . Singulary Intersection (n 1)
x E ny Vz(zEy---+XEZ)
n1 n1
The notation we've used here is somewhat uncommon. Mathematicians often write the intersection of a set a in these forms instead: nix : X E a}, nxEax. Similarly, instead of writing al n ... nan , mathematicians often write
nai n
i=n
If a is the empty set 0 or an Dr-element, then everything in the universe of discourse belongs to na. If we don't have a limited universe of discourse in mind, na would be a universal set: absolutely everything would belong to na. (Exercise: prove that if Vxx ~ a, then Vxx E na.) In the presence of the principle of abstraction, asserting the existence of such a universal set would lead to Russell's paradox (see section 7.2). In practice, we will be careful to talk about na only if we know that a is a non-empty set, typically, a family of sets.
Problems Prove the following theorems. Assume throughout that a, b, c, and d are sets. 1. VxVyn{x,y}=xny!
2. n{a} = a. 3. Suppose that a is a member of the set of sets b, and that a
~
c. Then nb
~
c.
159
Sets 4. Suppose that a belongs to the set of sets b, and that a and c are disjoint. Then nb n c =
0.
5. (a) If a set of sets is pairwise disjoint and contains at least two elements, its intersection is empty. (b) There is a set of sets that has an empty intersection, but is not pairwise disjoint. 6. Suppose that a is a member of the set of sets b. Then nb
~
a.
7. The intersection of any set of sets having the null set as a member is null. 8. Suppose that a nonempty set a of sets is a subset of a set b. Then na 9. a E b -> a
~
~
nb.
ub.
10. Suppose that every element of a is disjoint with b. Prove that Ua is also disjoint with b. "...-,.
" 1L}v'xxUx = U{x}. ;'.--.... ,..' 12. u0 = 0. 13. u(aUb)=uauUb.
14.
* A transitive set has all members of its members as members. Let a be a nonempty set, all of whose members are transitive. Show that na is transitive.
15.
* Suppose that every
16.
* a is transitive iff u(au {a})
member of a is a subset of some element of b. Show that Ua
~
Ub.
= a.
17. If a is a transitive set of sets, then Ua is a subset of a. 18. If a is a transitive set of sets, then U(a U {a}) is transitive. 19. If a is a transitive set of sets, then Ua is transitive. 20. If a is a set of transitive sets of sets, then Ua is transitive. 21. *anb#0->naunb~n(anb).
22.
* Let a be a set of sets.
23. Assume that a
7.7
~
Show Ua ~ na
b. Then Ua
~
-+
(a
# 0 -+
3ya = {y}).
ub.
Power Sets
In the first section, we defined what it was for one set to be a subset of another. One set is a subset of another just in case all elements of the first also belong to the second. We know how, then, to talk about the subsets of a given set. The powerset of x is the set of all of x's subsets. Suppose, for example, that a legislative body wants to set up committees to study certain kinds of bills. Each committee will be a subset of the body as a whole. So the powerset of the legislative body, taken as a set of its members, is the set of all possible committees that might be formed from it. Definition 7.17 (Powerset) Vx(P(x)
= {z:
z
~
x})
This definition yields the following rule of inference:
Logic, Sets and Functions
160 Powersets(P) x E P(y)
P
xC::y
P
What happens to the powerset function when it is applied to urelements? The following theorem gives us the answer.. Theorem 7.23 't/x(-,Sx
-7
P(x) = 0)
Our Powerset definition (together with Th. 7.3) yields the following theorem, which is often easier to use in proofs than the axiom itself: Theorem 7.24 't/xSP(x))
The formation of power sets increases the power of set theory tremendously. Power sets are generally much larger than the sets from which they are formed. In fact, any set with n members has 2n subsets, so its power set has 2n elements. If the set a has 4 elements, for example, P(a) has 16 elements. If a has 10 elements, P(a) has 1,024 elements. If a has 1,000 members, then P(a) has more elements than there are particles in the universe! More importantly from a mathematical point of view, if a has as many elements as there are natural numbers, then P(a) has as many members as there are real numbers.
Problems Prove the following theorems, assuming that a, band c are sets. 1. P(0) = {0}; P(P(0)) = {0, {0H; P(P(P(0))) = {0, {0}, {{0H, {0, {0} H·
2.
0 is a member of the power set of every set. (!)
3. {0, {0}} E P(P(P(a))).
4. P(a)
=
P(b)
-7
a
=
b.
.••:;<0""',
1 5~>UP(a)=a . ...~••• ,-'l"-:#
6. Vx(x E a
-7
Sx)
-7
c::
a
P(Ua).
7. P(a) n P(b) = P(a n b). 8. P(a) U P(b)
c::
P(a U b).
9. Suppose that a E b. Then P(a) E P(P(Ub)). 10. a 11.
c::
biffP( a)
** 't/x(x
c:: P(b).
E b --4 P(x)
(!)
c:: P(Ub)).
12. VxP({x}) = {0,{x}}. 13. P(a - b)
c::
(P(a) - P(b)) U {0}.
14. For any transitive set x, P(Ux)
c:: P(x).
15. A set is transitive iff its union belongs to its power set.
Sets
161
16. The only set having a power set that is identical to its unit set is the empty set. 17. Suppose that none of the members of set a have any members. Then an P(a)
~
{0}.
18. nP(a) = 0. , 19)p(na) .
... " ..
•
~_r
20. Ua ~ a
~
P(Ua) .
f4
a
~
P(a).
21. * Suppose a E P(b)& n a E p(nb). Then na
7.8
=1=
0 -> na
= nb.
Complex Abstraction
In this section we'll add a new version of abstraction to the simple rule of abstraction presented in 7.2. The new form will be called "complex abstraction", or "c-abstraction" for short. We will add to our language a new kind of term ("c-abstraction terms") and to our proof theory and new inference rule. Any sentence of the old language which is provable in the new system was already provable (in a slightly more roundabout way) in the old system, and any set nameable in the new language (using c- abstraction terms) was already nameable in the old language. Therefore, we are not making any essential changes in the system of section 7.1: the changes will merely make proofs more concise and perspicuous. The following rule extends our language to include the new kind of term:
Complex abstraction terms - formation rule Ift*is a closed term which contains constants Cl, ... A[vl/ ClJ ... , vii Cij} is a closed term.
,Ci,
and A is a formula, then: {t*[vl/Cl,"" vile;] :
Examples: 1. {P(x):xEa}
2. {xuy:xEa&YEb} 3. {Ux: x E a&c (j. x} Term #1 represent the set containing the powersets of any x, where x is any member of a. Term #2 represents the set containing binary unionsx U y, where x is any member of a and y is any member of b. Term #3 signifies the set containing all unions of sets which both belong to a and which do not contain C as a member. We will also introduce a new inference rule which may be applied to exploit and to introduce c-abstraction terms. We will add two forms of the rule, C-Abstr and C-Abstr*. C-Abstraction
x E {t*(Vl .. , v;) : A(Vl ... Vi)} :lxl .,. :lXi(X = t*(Xl ... xi)&A(Xl ... Xi»
C-Abst C-Abst
C-Abstr*
A(Xl ... Xi) t*(Xl,· .. ,Xi)
E
{t*(Vl",Vi) :A(Vl",Vi)}
C-Abst*
162
Logic, Sets and Functions
Note that C-Abstraction is an invertible rule, while C-Abstr* works only in the top-down direction. Here are some examples of C- Abstraction in the top-down direction:
a
E
{< x,y >: x
E
R&y
c E {P(y) : y E b&a n y
E
R}
= 0}
3x3y(a =< x,y > &x E R&y E R) 3y(c = P(y)&y E b&a n y = 0)
Here are some examples of using C-Abstr* in the top-down direction.
aEb a E b&c rt a bE a&b E c
P(a)E{P(x):XEb} na E {nx : x E b&c rt x) < b,b >E {< x,y >: x E a&y E c}
Problems Assume that a and b are sets of sets. 1. Show U{P(z) : z E a} ~ P(Ua).
2. Show {P(z) : z E a} ~ P(P(Ua)). 3. Show {xUa: x E b}
~
P(aUUb).
Chapter 8
Relations The theory of relations lies at the foundation of modern mathematics. Yet the theory emerged only late in the history of logic. Although Aristotle mentioned relations, and medieval logicians talked about them more extensively, only in the late nineteenth century did anyone develop a comprehensive account of relations. Charles Sanders Peirce, the coinventor of quantification theory, devised the first adequate symbolism for relations in the 1880s. A mathematically precise conception of relations had to await the work of Bertrand Russell and Alfred North Whitehead in Principia Mathematica, published in three volumes between 1910 and 1913. Recall that we distinguished monadic predicates from polyadic or relational predicates in quantificationallogic. Monadic predicates are true or false of a single object; polyadic predicates are true or false of sequences of objects, that is, of several objects taken in a particular order. Philosophers have often marked this contrast by speaking of properties as opposed to relations. Properties, on this way of speaking, are exemplified by single objects, while at least two objects must stand in a relation. Binary or dyadic relations correspond to binary or dyadic predicates; ternary or triadic relations, to ternary or triadic predicates, and so on. This chapter will concentrate primarily on binary or dyadic relations, which hold of pairs of objects.
8.1
Sequences and Ordered Pairs
Sets are unordered collections of objects. {Pittsburgh, Philadelphia, Chicago, New York, St. Louis, Montreal} is the set of cities with baseball teams in the National League's Eastern Division. The order in which these cities are listed makes no difference to the identity of the set; {Pittsburgh, Philadelphia, Chicago, New York, St. Louis, Montreal} is the same set as {New York, St. Louis, Montreal, Chicago, Philadelphia, Pittsburgh}. But ordering is crucial for the discussion of relations. Relational predicates are true or false of a number of objects at once, taken in a particular order. Bill's loving Sue, clearly, is not the same as Sue's loving Bill. So the unadorned idea of a set does not suffice for dealing with relations. Sequences, in contrast, have an order. Baseball fans who check the newspaper for the standings want to know who's in first place, who's in second, and so on; they are interested in the order among the teams. From the fan's perspective, the ordering Pittsburgh Philadelphia Montreal St. Louis 163
164
Logic, Sets and Functions Chicago New York
is very different from the listing New York St. Louis Chicago Montreal Philadelphia Pittsburgh. The teams, too, care about a sequence, rather than a set; they strive to be at the top of the standings. Without the concept of an ordering, baseball's annual competition-and, indeed, the very notion of competition-would make no sense. Sequences, then, consist of objects taken in a certain order. The objects in a sequence are its constituents; they constitute the sequence, and it comprises them. To represent a sequence graphically, we can write a list, as with the baseball standings above, or use angle brackets. Although {O, I} is the same set as {I,O}, (0,1) is not the same sequence as (1,0). Similarly, {Pittsburgh, Philadelphia, Chicago, New York, St. Louis, Montreal} is the same set as {New York, St. Louis, Montreal, Chicago, Philadelphia, Pittsburgh}, but (Pittsburgh, Philadelphia, Chicago, New York, St. Louis, Montreal) is not the same sequence as (New York, St. Louis, Montreal, Chicago, Philadelphia, Pittsburgh). A sequence with just two constituents is an ordered pair. A sequence with n constituents is an n-tuple. Throughout most of this chapter, we'll concentrate on ordered pairs, since they are the sequences most useful for analyzing binary relations. It's possible to define sequences as sets in several different ways. Norbert Wiener and Casimir Kuratowski (in 1912 and 1921, respectively) developed a particularly convenient definition. According to this definition, we can think of an ordered pair as a set consisting of two other sets. One is the set of the constituents of the ordered pair; this set captures the idea that sequences can be the same only if they have the same constituents. To capture the distinctive feature of sequences, the element of order, the other set contains just the first element of the ordered pair. The definition thus implies that an ordered pair is identical to another just in case they have the same constituents and the same first constituent. But, of course, if they have the same constituents, and the same first constituent, they must also have the same second constituent. Definition 8.1 The ordered pair (x,y) is the set {{x}, {x,y}}. In symbols, 'r/x'r/y((x,y) = {{x},{x,y}}).
The axiom of extensionality specifies that sets with the same elements are identical. Obviously the analogue of this principle for sequences should not be true. The identity of a sequence consists in both the identity of its constituents and their order. Sequences are identical, not just when they have the same constituents, but when the same constituents occupy the same places of the sequences. A sequence must have the same first constituent as another, and the same second constituent, and the same third constituent, etc., for the sequences to be identical. For ordered pairs, this principle becomes: ordered pairs are identical if and only if they have the same first constituents and the same second constituents.
Relations
165
Sequence Identity. An ordered pair (x, y) is identical with an ordered pair (z, w) if and only if x = z and y = w. In symbols, 'Vx'Vy'Vz'Vw( (x, y) = (z, w) <---+ x = z & y = w).
0= (x, y) = (z, w) x
=z
& y
=w
0=
To show that our definition entails the above rule, assume that (a, b) = (c, d). By the definition, then, {{a}, {a, b}} = {{ c, }{c, d} }. By the axiom of extensionality, these sets have the same members; in particular, {a} belongs to one just in case it belongs to the other. But {a} dearly belongs to {{a}, {a,b}}, so {a} E {{c}, {c, d}}. By the definition of pair sets, this means that {a} = {c} or {a} = {c, d}. In either case, a = c.To see that b = d, observe that, by extensionality, {a, b} E {{c}, {c, d} }. So {a, b} = {c} or {a, b} = {c, d}. If {a, b} = {c}, then b = c, so a = b. But this means that {{a}, {a, a}} = {{a}, {a,d}}, i.e., {{a}} = {{a}, {a,d}}. So {a,d} E {{a}}, which means that {a, d} = {a} by the definition of unit sets. But it follows that d E {a}, so d = a. And, since a = b, it follows that b = d. If, on the other hand, {a,b} = {c,d}, bE {c,d}, so b = c or b = d. If b = c, then a = b, and by the reasoning we've just seen, b = d. So, in any case, b = d. Thus a = c and b = d, which completes the left-to-right direction of the proof. The right-to-Ieft direction is trivial. Note that there is a large amount of arbitrariness in the Wiener-Kuratowski definition of ordered pairs. We can get the effect of requiring that the pairs have the same first constituent and the same second constituent by using a definition that forces them to have the same constituents and the same first constituent. But we could also require that the pairs have the same constituents and the same second constituent: we could define (x, y) as {{x, y}, {y} }. And there are other ways as well. Within set theory, we could assign many different sets the role of (x, y); which set we choose depends chiefly on matters of convenience. What matters is that language about ordered pairs reduces to language about sets. We therefore do not need to introduce ordered pairs as constituting a new kind of object, apart from sets.
8.2
Cartesian Products
The Cartesian product of two sets is a set of ordered pairs. Each pair has a member of the first set as its first constituent, and a member of the second set as its second constituent. Furthermore, every such pair belongs to the Cartesian product. We'll denote the Cartesian product of two sets A and B by writing A x B. Definition 8.2 The Cartesian product of y and z is the the collection of all and only those ordered pairs with elements of y as first constituents and elements of z as second constituents. In symbols, 'VyVz y x z = {(u,v): u E y & v E z}. C-Abstraction and the definition of Cartesian product gives us the following two inference rules: x
xEyxz :Ju:Jv(x = (u, v) & u E Y & v E z)
x
166
Logic, Sets and Functions
x*
(u, v) E Y x
u
E Y &
v
Z
E Z
x*
Theorem 8.1 VxVy Sx x y. Theorem ?? follows from the Axiom of Abstraction. In virtue of the properties of urelements, we can also prove: Theorem 8.2 VxVy«,Sx V ,Sy)
-+
x x y
= 0).
To take some examples: suppose A is a set of variables and B is the set of natural numbers between 0 and 100. The Cartesian product A x B is the set of all pairs matching a variable with a number, having the form (Variable, Number). Thus (v, 57), (u,38), (w,20) will all be elements of A x B. Or, suppose A is the set of Congressmen during the 1985 session and B is {yes, no}. Then A x B will include the pairs (Kemp, yes), (Kemp, no), (Dingell, no), (Pickle, yes), (Gingrich, yes), etc. Notice that, in the case involving variables and numbers, any particular assignment of numbers to variables will be some subset of the Cartesian product A x B. The Cartesian product, in other words, will be a union of all possible assignments of numbers between 0 and 100 to the variables in the set A. Likewise, where A is the set of Congressmen and B the set {yes, no}, A x B will be the union of all possible voting patterns; each way of pairing yes or no with Congressmen will be a subset of the Cartesian product. Though this chapter will generally focus on Cartesian products of two sets, we can generalize the Cartesian production function to a function of any finite number of arguments. To do this, however, we need to generalize the notion of ordered pair. The proper way to do this is by means of a recursive definition, similar to the one that we used in constructing our languages of logic. The base case for our definition here is just the definition of ordered pairs. The recursive clause appeals to it. In many cases such as the present one where we have to show what it is to be an n-tuple for any n, this is the only way we can define the notion. We will repeat the definition of ordered pair here as the base clause in our definition. We will write an n-typle of Xl, ... , X n as (Xl, ... , x n ). Definition 8.3 VxVy (x, y) = {{x}, {x, y}} VXI"",Xn+1 (XI, ... ,Xn+l) = «XI, ... ,Xn),Xn+l) The Cartesian product of n sets YI, ... , Yn (written YI x ... x Yn) will be the set of n-tuples whose ith members all belong to Yi.
Definition 8.4 The Cartesian product of n sets is the set of n-tuples whose ith members belong to the i th set, for each place i. In symbols, VYI···Yi(YIX ... XYi={(ZI",.,Zi): zIEYI& ... &ZiEyd)
167
Relations
x #n. An n-tuple belongs to a Cartesian product of n sets just in case, for each place i, the ith constituent of the n-tuple belongs to the ith set: x#n
x ...
(Zl' ... ,Zi) E Yl Zl E Yl
& ... &
X Yi
Zi E Yi
X #n
Thus, say that A is the set of Congressmen, A the set {yes, no}, C the set of bills to be voted on during the session, and D the set of days during which the session takes place. Members of A x B x C x D will have the form (Congressman, Vote, Bill, Date): so (Garcia, yes, Fair Housing Act, May 8, 1985), (Garcia, no, Fair Housing Act, June 14, 1985), etc., will all belong to the Cartesian product. Once again, any actual record of votes during the session will be a subset of A x B x C x D. Mathematicians often abbreviate the Cartesian product of a set with itself, A x A, as A 2 . More generally, the Cartesian product of a set A with itself n times is An. Examples [Filling in brackets is left as an exercise for the reader.]
[1.] 1. Shew 'v'y(3x(x E a & Y = {x} x b) ~ Y E P(a x b))
2.
Shew 3x(x E a & c = {x} x b) ~ c E P(a x b)
3.
3x(xEa&c={x}xb)
4.
d E a & c = {d} x b
5.
Shew 'v'z(z
6.
Shew e
7.
eE c
8.
e E {d} x b
9.
3u3v(e = (u,v) & U
E
E
c
c
3E
~ Z E
~
e
E
a x b)
a x b)
=E, 4, 7 E
{d} & v
E
b)
x
10.
e=(j,g) &jE{d}&gEb
3E
11.
f=d
Unit, 10
12.
e = (d,g)
=E, 10, 11
13.
(d, g) E a x b
x*, 4,10
14.
eEaxb
=E, 12, 13
15.
c~axb
~,
16.
cEP(axb)
P,15
5
Logic, Sets and Functions
168
[2.]
1. Shew Vx(x E a
2.
Shew c E a
---7
---7
{x} x bE pea x b))
{c} x b E P(a x b)
3.
cEa
4.
Shew Vz (z E {c} x b ---7
Z
E a x b)
Sh€lw e E {e} x b ---7 e E a x b
5. 6.
eE{e}xb
7.
3u3v(e = (u,v) &
8.
e=(j,g) &fE{C}&gEb
3E
9.
f=e
Unit, 8
10.
e = (e,g)
=E, 8, 9
11.
(c,g) E a x b
x*,3,8
12.
eEaxb
=E,lO,ll
13.
{e} x b c;, a x b
c;" 4, Th. ??
14.
{e} x bE pea x b)
P,13
U
E {e} & v E b)
x,6
[3.] 1. axb=axe 2. a
=10
3. Sa & Sb & Se 4. Shew b = c 5.
3y yEa
Cor. 11.1, 2,3
6.
eEa
3E,5
7.
Shew Vx(x E b <-+ x E c)
8.
Sh€lw dEb
9.
Shew
<-+
dEe
---7
10.
dEb
11.
(e, d)
axb
x*, 6, 10
12.
(e, d) E a x c
=E, 1, 11
E
Relations
169
13.
eEa&dEc
14.
dEc
15.
x*,
12
.£hElw .-
Similar 16.
b= c
Ext, 3, 7, 7
Problems
What is A x B, given these sets as A and B?
A
B
l.
{D}
{I}
2.
{D}
{D,l}
3.
{D,l}
{I}
4.
{D,l}
{D,l}
5.
{I}
{D,1,2}
6.
{D,1,2}
{D}
7.
{D,l}
{D,1,2}
8.
{D,1,2}
{1,2}
9.
{O,1,2}
0
0
{I}
10.
Prove the following theorems. 11. For any x, (x,x) = {{x}}.
12. VxVy((Sx & Sy) 13. Vx(x x 0 = 0).
->
(x x y
= 0 -> (x = 0 V Y =
0))).
(!) ((x x y = y x x & x =I- 0 & y =I- 0)
14. VxVy((Sx & Sy)
->
15. VxVyVz(x <;;; y
x x z <;;; y x z).
->
16. VxVyVz((Sx & Sy & Sz)
->
->
x = y)).
->
y <;;; z)).
(!)
((x =I- 0 & x x y <;;; x x z)
17. VxVyVz(x x (y n z) = (x x y) n (x x z)). 18. VxVyVz(x x (y U z) = (x x y) U (x x z)). 19. VxVyVz(x x (y - z) = (x x y) - (x x z)).
170
8.3
Logic, Sets and FUnctions
Relations
Relations, from a mathematical perspective, are simply sets of sequences. An n-ary relation is a set of n-tuples. Binary relations, on which we'll focus most of our attention, are sets of ordered pairs. Intuitively, each pair contains objects that stand in the relation. If Bill loves Sue, for example, then the pair (Bill, Sue) must belong to the "love" relation. Similarly, (0,1), (0,2), (1,3) and (15,213) all belong to the "less than" relation. Identifying relations with sets is very useful mathematically, but it is an oversimplification. As we usually use the term relation, a relation is not a set of n-tuples but that by virtue of which we construct the set of n-tuples. NobO<:lY who asked the question "What is love?" would be satisfied with a list of ordered pairs, including, say, (Antony, Cleopatra), (Napoleon, Josephine), (Abelard, Heloise) and (John, John's mother). Nor would someone trying to understand defeat learn much from a newspaper listing such as Michigan State 31 Ohio State 27 Texas 17 Slippery Rock 35
Michigan 30 Alabama 14 Texas Tech 9 Muhlenberg 23
The distinction is very similar to that between a program and a table of inputs and outputs. We can describe a relation between numbers by listing input and output values, obtaining something like Input Output 4 o 1 4 2 6 3 10 4
16
5
24
But we seem to gain far more understanding of what the relation is if we can say why these numbers are correlated; if we can say, for instance, that, where n is the input, the output is n 2 - n + 4. Similarly, our list of football scores does little to help us to understand defeat. We want to know why the teams are listed together in the way they are. A relation, like a program, from this point of view, is something like a rule. In identifying relations with sets, we are treating them as tables of outcomes of applying rules, not as the rules themselves. It might be more precise, therefore, to speak of sets of sequences as relation tables or diagrams. In this text, however, and in virtually all of mathematics, nothing depends on this distinction. Let's continue, then, to think of relations as sets of sequences. We can say that Bill loves Sue by writing Bill loves Sue (or a corresponding formula, say, Lbs) or by saying (Bill, Sue) E "love". We can say that < 1 by writing 0 < 1 or by writing (0,1) E <. In general, we can use the predicate notation Rxy and the set-theoretic notation (x, y) E R interchangeably. The following is a schema about the ways we talk about relations in set theory and in predicate logic. Let R be any (binary) relation.
°
Notational Interchange. VxVy«x,y)
E
R ...... Rxy).
(NI) (x,y) E R Rxy
(NI)
171
Relations
Suppose that we are interested in the relation corresponding to the predicate Congressman x voted y on z on w, where z is a bill and w is a date. The quadruples in the relation will all have the form (Congressman, Vote, Bill, Date). If A, B, C and D are the sets of Congressmen, votes, bills and dates, respectively, then each quadruple belonging to the relation will also belong to A x B x C x D. But this is just to say that the relation is a subset of this Cartesian product. This is true in general: n-ary relations are subsets of Cartesian products of n sets. Binary relations, specifically, are subsets of Cartesian products of two sets. Definition 8.5 A binary relation is a subset of a Cartesian product of two sets. Symbolically, ~
Vx(x is a binary relation .-. 3y3z(Sy & Sz) & x Theorem 8.3 Vx(x is a binary relation Proof. Use the definitions of
~
-+
y x z)).
Sx)
and of binary relation. _
Definition 8.6
Definition. An n-ary relation is a subset of a Cartesian product of n sets. In symbols,
Vx(x is an n-ary relation .-. 3Yl ... :JYn((SYl & ... & SYn) & x ~ Yl
X ... X
Yn)).
When a relation is a subset of a Cartesian product of a set with itself, the relation is on that set. Definition 8.7 A binary relation x is a relation on a set Y iff x relation x is on a set y just in case x S;; y n .
~
Y x y. In general, an n-ary
Notice that this definition doesn't suffice to let us answer uniquely the question, What set is this relation on? Any relation on a set will also be a relation on any superset of that set. More generally, if R is a subset of Ax B, then R will also be a subset of the Cartesian product of any superset of A with any superset of B. Often, therefore, it's interesting to know just what objects are related by the relation. Say that the first place of a binary relation is the subject place; the second is the object place. The domain of a relation is the set of subjects of the relation. Thus, if the relation is "love", the domain is the set of lovers, that is, the set of all those who love somebody or something. The range of a relation is the set of the relation's objects. The range of "love", therefore, is all those people or things that are loved by anyone. The objects related in some way by the relation constitute the relation's field. The field is simply the union of the range and the domain. In the case of "love", the field is the set consisting of all those who love or are loved. Note that domain, range and field are all junctions of relations. Relations are just objects in our theory. So we can define functions on them, have variables like x, y and z range over them and name them with constants like "a", "b", and "c". So it makes perfect sense to have functions of relations, if it makes sense to have functions of objects like numbers. Stay tuned for our discussion of functions in the next chapter! Definition 8.8 The domain V(R) of R is the collection of all x such that, for some y, Rxy. In symbols, V(R) = {x: 3y (x, y) E R}
172
Logic, Sets and Functions
Theorem 8.4 Vz SV(z) By the Abstraction axiom.
V x E VCR) 3yRxy
V
Definition 8.9 The range R(R) of R is the set of all x such that, for some y, Ryx. In symbols,
R(R) = {x: 3y (y,x) E R} Theorem 8.5 Vx S R(x)
R x E R(R) 3yRyx
R
Definition 8.10 The field F(R) of R is the union of the domain and range of R. In symbols, F(R) = VCR) U R(R). Theorem 8.6 Vx S F(x)
F x E F(R) x E VCR) V x E R(R)
F
A binary relation, then, relates the objects in its domain to those in its range. To take an example, consider the "absolute value" relation on the set Z of positive and negative integers. We can take the absolute value of any integer, positive or negative, so the domain of the relation is Z in its entirety. Absolute values, however, are always positive (if nonzero), so the range is just the subset of Z consisting of the positive integers and zero. For another example, consider the relation "outranks" on the set of Army personnel. The domain of the relation consists of all those who outrank anybody. The domain, then, includes everyone in the Army above the rank of private. The range consists of all those who are outranked by anybody; this includes everyone below the rank of four-star general. Between them, the range and domain exhaust all Army personnel, so the field consists of the entire army. Both these examples illustrate that an object may be in both the domain and the range of a given relation. Any positive integer, for example, occupies both the domain and the range of the "absolute value" relation. And most people in the Army are in both the domain and the range of
Relations
173
the "outranks" relation. Privates outrank nobody, so they occupy only the range; four-star generals occupy only the domain. Everyone in between is in both.
Some examples. Assume Rand S are binary relations. 1. Shew R(R uS) = R(R) U R(S)
2.
Shew Vx(x E R(R U S) .-. x E R(R) U R(S))
3.
Shew a E R(R US) .-. a E R(R) U R(S)
4.
Shew
->
5.
a E R(Ru S)
6.
(b,a) E Ru S
R, 5, :JE
7.
(b, a) E R V (b, a) E S
U,6
8.
a E R(R) Va E R(S)
SL, R, 7
9.
a E R(R) U R(S)
U,8
10.
Shew
~
11.
a E R(R) U R(S)
12.
a E R(R) Va E R(S)
13.
Shew a E R(R)
->
U,11
a E R(R uS)
14.
a E R(R)
15.
(c,a) E R
R,14
16.
(c,a) E R V (c,a) E S
VI, 15
17.
(c,a) E Ru S
U,16
18.
a E R(RU S)
31, R, 17
19.
Shew a E R(S)
20. 21.
->
a E R(RU S)
Similar
a E R(RU S)
1. R is a binary relation
2. Shew R n S is a binary relation 3.
Sa & Sb & R <;; a x b
VE, 12, 13, 19
A
174
Logic, Sets and Functions
4.
ShElW Vx(x ERn S
5.
ShElW C ERn S
---->
---->
x E a x b)
cEaXb
6.
cE RnS
7.
cER&cES
n,6
8.
cEaxb
r;;.*, 3, 7
9.
S Rn S & Sa x b
Th 19, Th ??
RnSr;;.axb
r;;., 4, 9
10.
Def, 3, 1" 31
R n S is a binary relation
11.
Problems Specify the domain, range and field of these relations. 1.
< on the set N of natural numbers (including 0)
2.
~
on N
3. the squaring function on N 4. :::; on Z
5. the squaring function on Z 6.
> on the closed interval [0,1] of the real line
7. ~ on [0,1]
8. < on {O} 9.
< on {0,1}
10. :::; on {0,1} Prove these theorems. 11. Any subset of a binary relation is also a binary relation. 12. The union of any two binary relations is a binary relation.
(!) (!) [Proved above]
13. The intersection of any two binary relations is a binary relation. 14. The difference between any two binary relations is a binary relation. 15. 0 is a binary relation.
(!)
Assume throughout the following that Rand S are binary relations. 16. R(R U S) = R(R) U R(S). (!) [Proved above] 17. R(R n S)
<;;;;
R(R) n R(S).
18. R(R - S)
<;;;;
R(R) - R(S).
(!)
(!) (!)
175
Relations 19. V(R U S) = V(R) U V(S).
(! )
20. V(R n S) <:;; V(R) n V(S).
(! )
21. V(R - S)
2 V(R) - V(S).
22. R <:;; V(R) x R.(R).
23. Let A be a set of binary relations. Then R.(uA) is the union of the ranges of all the relations in A. 24. Let A be a set of binary relations. Then V(uA) is the union of the domains of all the relations in A. 25. If R~ A x B, then V(R) =.4. and R.(R) = B.
8.4
Properties of Relations
Many different kinds of relations merit mathematical interest. This section will present some of the most important properties of relations. Throughout, let's limit ourselves to talking about a binary relation R on a set A. Reflexive relations hold between every object and itself. Every number n, for example, is such that n = n, n s:: nand n 2: n. Identity, greater-than-or-equal-to, and less-than-or-equal-to are all reflexive. Definition 8.10 . R is reflexive on A ijjVx(x E A
--t
Rxx).
On the set of people, "is", "is exactly as tall as", "is at least as tall as" , and "has the same birthday as" are reflexive. Specifying a set A is important because a relation may be reflexive on some sets but not on others. Consider, for example, the relation "analyzes" on the set of people. Presumably, some people analyze themselves, while others don't, so the relation isn't reflexive on that set. But, if all psychoanalysts analyze themselves, then that relation is reflexive on the set of psychoanalysts. Similarly, "absolutevalue" is not reflexive on the set of integers-I-41 =I- -4-but it is reflexive on the set of positive integers. Irreflexive relations never hold between an object and itself. No number, for example, is greater than, or less than, itself. So < and > (as well as =1-) are irreflexive. Definition 8.11 R is irreflexive on A ijjVx(x E A
--t
-.Rxx).
On the set of people, "is taller than", "is shorter than", "is a child of", "is a parent of", and "isn't the same person as" are all irreflexive. Obviously, not all relations are either reflexive or irreflexive. Reflexive relations always hold between an object and itself, while irreflexive relations never do. Many relations, however, sometimes hold between an object and itself and sometimes don't. "Respect", for example, is neither reflexive nor irreflexive: some people respect themselves, and some don't. The relation of being a square of, among numbers, is similarly neither reflexive nor irreflexive: 02 = 0 and 12 = 1, but it's not true in general that n 2 = n. A relation is symmetric if, whenever it holds in one direction, it also holds in the other. Identity is clearly symmetric: if x = y, then y = x. Other relations on numbers that are symmetric are =1-, "have a common divisor", "have the same prime factors" and "have the same absolute value". Definition 8.12 R is symmetric on A ijjVxVy(x, yEA
--t
(Rxy
--t
Ryx)).
Logic, Sets and Functions
176
Among people, "is related to", "is a sibling of", "is a friend of", "has the same birthday as" and "is exactly as tall as" are symmetric. An asymmetric relation, in contrast, holds in at most one direction between any two objects. If it holds in one direction, then it doesn't hold in the other. Both < and > , clearly, are asymmetric relations. Definition 8.13 R is asymmetric on A iflVxVy(x, yEA
(Rxy
->
->
-.Ryx)).
Among people, "is taller than", "is shorter than", "is a child of", and "is a parent of" are all asymmetric. Antisymmetric relations hold in both directions between objects only if those objects are the same. To put it differently, they never hold in both directions between two distinct objects. Among numbers::; and 2: are antisymmetric: there are numbers nand m such that n ::; m and m ::; n, but only when n = m .. Definition 8.14 R is antisymmetric on A iflVxVy(x, yEA
->
((Rxy & Ryx)
->
x = y)).
Again, not all relations are either symmetric, asymmetric or antisymmetric. "Love", for example, isn't symmetric: there are cases of unrequited love. But it's not asymmetric, since some love is not unrequited. "Love" isn't antisymmetric either, for there are cases of mutual love that aren't just cases of self-love. A relation is transitive if, whenever it holds between x and y and also between y and z, it holds between x and z. Among numbers, <, ::;, >, 2:, "divides" and = are all transitive. Definition 8.15 R is transitive on A iflVxVyVz(x,y,z E A
->
((Rxy & Ryz)
->
Rxz)).
Among people, "is at least as tall as", "is exactly as tall as", and "has the same birthday as" are transitive. Whenever an intransitive relation holds between x and y and y and z, it does not hold between x and z. "Is one greater than" is intransitive: if n = m + 1, and m = k + 1, then certainly n i= k + 1. Definition 8.16 R is intransitive on A iflVxVyVz(x, y, z E A
->
((Rxy & Ryz)
->
-.Rxz)).
Among people, "is a year older than", "is ten pounds heavier than", and "is four inches taller than" are all intransitive. Many relations are neither transitive nor intransitive. "Have a common divisor", for example, doesn't always hold between x and z if it does between x and y and y and z. 4 and 12 have a common divisor, for example, and so do 12 and 3. Yet 4 and 3 have no common divisor. But it does hold between x and z sometimes: 4 and 12 have a common divisor, as do 12 and 16. But in this case 4 and 16 also have a divisor in common. A relation is connected on a set if it holds, in one direction or another, between any two distinct objects in the set. Among numbers, <, >, ::;, 2:, and i= are all connected. If nand m are distinct numbers, then one must be larger than the other; either n > m or m > n. Similarly, if nand mare distinct, then n i= m. Definition 8.17 R is connected on A iflVxVy(x, yEA
->
(x
i=
y
->
(Rxy V Ryx))).
This is sometimes called the trichotomy property, because it says that, for any x and y, one of three things must hold: Rxy, Ryx or x = y. A relation is strongly connected on a set if it holds, in one direction or another, between any two objects in the set. This is the same as connectedness, except that the objects no longer have to be distinct. The definition drops the x i= y clause:
Relations
177
Definition 8.18 R is strongly connected on A iffVxVy(x, yEA
--+
(Rxy V Ryx)).
This can be called the dichotomy property, because it says that, for any x and y, either Rxy or Ryx. Although =I, <, >, S;, and::::: are all connected, only S; and::::: are strongly connected. No matter what nand m are, it must be the case that either n S; m or m S; n. But it isn't necessarily true that n < m or m < n; nand m might be equal. Notice that strongly connected relations are reflexive. For any x and y in A, Rxy or Ryx. If we choose the same object for both x and y (say b), then Rbb or Rbb. So strongly connected relations must hold between each object and itself. An example: 1. R is irreflexive on A. 2.
Shew A x A - R is reflexive on A
3.
Vx(x E A
--+
,Rxx)
4.
Shew Vx(x E A
5.
Shew b E A
Irrefl., 1
--+
--+
(x,x) E A x A - R)
(b, b) E A x A - R
6.
bE A
7.
,Rbb
8.
(b, b)
9.
(b, b) E A x A
x,6
(b, b) E A x A - R
Def-, 8, 9
10.
VE, 3, SL, 6 ~
R
A x A - R is reflexive on A
11.
Problems Which of the properties this section has discussed apply to these relations? 1. E (membership) 2. ~ (subset)
3.
c
(proper subset)
4. are disjoint
5.
< (onN)
6. > (on Z) 7. S; (on N) 8. ::::: (on Z) 9. = (on N) 10.
=I
(on Z)
NI,7
Irrefl., 4
178 11. divides (on N) 12. is a factor of (on N) 13. is a prime factor of (on N) 14. is a multiple of (on N) 15. is the power set of 16. is the intersection of 17. is the union of 18. is the complement of 19. is the absolute value of (on Z) 20. is the square of (on N) 21. is a square root of (on N) 22. is the square of (on Z) 23. is a square root of (on Z) 24. is the square of (on the reals) 25. is a square root of (on the reals) 26. is a square root of (on the complex numbers) 27. implies (on the set of formulas) 28. is equivalent to (on the set of formulas) 29. are mutually satisfiable (on the set of formulas) 30. contradicts (on the set of formulas) Which properties do these relations have on the set of people? 31. is the father of 32. is a parent of 33. is a child of 34. is a daughter of 35. is a sibling of 36. is a brother of 37. is a cousin of 38. is a descendant of 39. is related to
Logic, Sets and Functions
179
Relations Give an example of a relation on a set that is: 40. both reflexive and irreflexive 41. symmetric and antisymmetric 42. reflexive, symmetric andtransitive 43. reflexive, antisymmetric and transitive 44. irreflexive, asymmetric and transitive 45. reflexive and symmetric, but not transitive 46. reflexive and antisymmetric, but not transitive. 47. irreflexive, symmetric, but not transitive 48. symmetric and intransitive 49. antisymmetric and intransitive 50. asymmetric and intransitive Prove these theorems. Let Rand S be relations on sets A and B, respectively. 51. If R is intransitive on A, then R is irreflexive on A. 52. If R is symmetric and transitive on A, then R is reflexive on F(R). 53. If R is strongly connected on A, then R is reflexive on A. 54. If R is strongly connected on A, then R is dense on A. 55. If R is asymmetric on A, then R is antisymmetric on A. 56. If R is asymmetric on A, then R is irreflexive on A. 57. If R is transitive and irreflexive on A, then R is asymmetric on A. 58. If Rand S are reflexive on A and B, respectively, the R U S is reflexive on A U B. 59. If R is irreflexive on A and T
~
R, then T is also irreflexive on A.
60. If R is reflexive on A, then (A x A) - R is irreflexive on A. 61. If R is irreflexive on A, then (A x A) - R is reflexive on A. 62. If R is connected on A and T 2 R, then T is also connected on A. 63. If R is irreflexive on A, and S on B, then R
n Sis
irreflexive on
An B.
64. If R is irreflexive on A, and Son B, then R - Sis irreflexive on A-B. 65. If R is reflexive on A, then, if T 2 R, T is reflexive on A. 66. If R is symmetric on A, and Son B, then R
nS
is symmetric on
An B.
67. If R is symmetric on A, and S on A, then R - S is symmetric on A.
Logic, Sets and Functions
180
On the set {a, I} there are 16 possible binary relations, listed below. For each, list the properties above that hold of it. 68. 0 69.
{(O, O)}
70.
{(O, I)}
71.
{(I, I)}
72. {(1,0)} 73.
{(O,O) ,(0,1)}
74.
{(O,O), (1,1)}
75.
{(O,O),(I,O)}
76. {(0,1), (1,
I)}
77. {(a, 1), (1,0)} 78. {(1,1),(1,0)} 79. {(0, 0) , (0, 1) , (1, 1) } 80.
{(O,O) ,(0,1), (1,0)}
81.
{(O,O), (1, 1), (1,0)}
82. {(a, 1), (1, 1), (1,0)} 83. {(0, 0) , (0, 1) , (1, 1) , (1, 0) }
8.5
Ordering Relations
This chapter began by discussing the concept of order. Order, recall, is what distinguishes sequences from sets. One way of introducing order into a set, however, is to talk about a relation on that set. Given the set of teams in the National League's Eastern Division, for example, we can introduce an ordering simply by relating the teams according to their won-lost records. "Has a better winning percentage than" is a relation on the set of teams. Furthermore, it orders the teams. Of course, it may not decide all questions of ordering; there may be ties. By thinking of a relation ordering the set, rather than simply of a sequence as given, we gain some insight into the principle behind the ordering. We can also consider other orderings of the same set. For instance, we might be interested in the relations "draws larger attendance than", "resides in a larger metropolitan area than" and "has more television revenue than". The orderings these relations induce might themselves have some important relations to the won-lost ordering. There are many different kinds of orderings. Some orderings allow ties; some don't. Some orderings rank everything in the set against everything else; others may allow some objects to go unranked viz-a-viz others. Some ordering relations are reflexive, while others are irreflexive. Let's begin by examining some reflexive ordering relations. For example:
Relations
181
o A O~O~O-O
c 0"-0'-'0.-.0
+/+/
0"-0
B
0"-0"-0--0
o O..-O~O..-O
0+-0
(A) has no ties or unranked elements. (B) has no ties, but does have an unranked element. (C) has no unranked elements, but does have several ties. And (D) consists of two orderings that are unrelated to each other. If these diagrams seem too abstract, imagine four golf tournaments set up for members of the PGA. In tournament (A), all PGA members compete; all finish, and no ties, for any place, are permitted. In tournament (B), no ties are permitted for any place, but one PGA member does not compete (or, at least, doesn't finish). In tournament (C), everyone in the PGA competes and finishes, but ties are allowed. Finally, in tournament (D), each PGA member enters exactly one of two competitions (say, on two different courses). Within each each competition, no ties are allowed, but there's no attempt to rank those in competition 1 (playing at Pebble Beach, say) against those in competition 2 (playing, perhaps, at St. Andrews). If we rule out ties, we obtain a partial ordering. A relation partially orders a set iff, on that set, it is reflexive, transitive and antisymmetric. Definition 8.19 Relations R is a partial ordering of A iff R is reflexive, transitive and antisymmetric on A.
The antisymmetry requirement has the effect of precluding ties. Implication, for example, is not a partial ordering of the set of quantificational formulas. Formulas can imply each other without being identical. Equivalence, for the same reason, isn't a partial ordering either. Identity, however, is, and trivially so. If x = y and y = x, then certainly x = y. Both ~ and 2: partially order the set of numbers: n 2: m and m 2: n only if n = m, so 2: (and, with it, ~) is antisymmetric. Partial orderings may allow unranked elements, so, of the four orderings diagrammed above, only (C) fails to be a partial ordering. Tournament (A) has no ties and no unranked elements, so the ordering is not only reflexive and transitive but also antisymmetric. Tournament (B) has an unranked element, but still permits no ties. Tournament (D) consists of two separate competitions, but within them no ties are allowed. (A), (B) and (D) are all partial orderings. If we banish both ties and unranked elements, we get a linear ordering or chain. A relation linearly orders a set iff it's reflexive, transitive, antisymmetric and connected on that set. Definition 8.20 A relation R is a linear ordering of A iff R is reflexive, transitive, antisymmetric, and connected on A.
The requirement of connectedness eliminates unranked elements, because it demands that any two distinct elements be ranked against each other. Both ~ and 2: rank every number against every other, and allow no ties, so both are linear orders. Of our diagrams above, only (A) is a linear order. Every other diagram shows ties or unranked elements. The diagram suggests, incidentally,
182
Logic, Sets and Functions
why linear orderings are called linear, and also called chains: every linear ordering looks like (A) in having the form
o..-o..-o-a.-o..-o..-o-o So far, all the orderings we've discussed have been reflexive. But < and> , for example, are also ordering relations. When dealing with irreflexive orderings, the cases of ties and unranked elements really become the same. If Sam and Gary tie in the tournament, then, using the reflexive ordering "has at least as Iowa score as" , we can say that Sam and Gary stand in that relation to themselves and each other. But, using the irreflexive ordering "has a lower score than" , we have to say that they are unrelated to each other. In dealing with irreflexive orderings, therefore, ties and unranked elements come to the same thing. A relation is a strict partial ordering of a set A iff it is irreflexive, transitive and asymmetric on A. The asymmetry requirement adds no weight to this definition; any irreflexive and transitive relation is asymmetric. So we can simplify the definition to speak only of irreflexivity and transitivity. Definition 8.21 A relation R is a strict partial ordering of A iff R is irrefiexive and transitive
on A. (We could also speak of just asymmetry and transitivity, since asymmetry entails irreflexivity.) Both
are strict partial orderings on the numbers. "Is an ancestor of" and "is a descendant of" are strict partial orderings of the set of people. Strict partial orderings are partial in that they allow unranked elements, or, equivalently, ties. Each of these diagrams, then, represents a strict partial ordering. (No circle is related to itself; to indicate this, the circles here are filled rather than open.)
. ...-.--. ././..--.. ..--.. E
.--.
G
.--. .--.
•
..--...--..-. ..--...--...-. ..-. F
H
Notice that there are no vertical lines in these diagrams; asymmetry requires that. The circles are black, signalling that the elements are not self-related, because of irreflexivity. Otherwise, however, they are basically the same as diagrams (A)-(D) above. They can represent the same golf tournaments we mentioned earlier. If we rule out ties and unranked elements, we get a strict linear ordering. A relation is a strict linear ordering of a set just in case it is irreflexive, asymmetric, transitive and connected on the set. Definition 8.22 A relation R is a strict linear ordering of A iff R is irrefiexive, asymmetric,
transitive and connected on A.
Relations
183
< and>, because they rank every number viz-a-viz every other, satisfy the additional requirement of connectedness. Hence they strictly linearly order the numbers. "Is an ancestor of", however, is not a strict linear ordering of the set of people; it's very easy to pick two people neither of which is an ancestor of the other. The connectedness requirement rules out both ties and unranked elements, so, of our four diagrams (E)-(H), only (E) represents a strict linear order. We can devise an even stronger irreflexive ordering relation. A relation R is well-founded on a set A iff every nonempty subset of A has an R-minimal element, i.e., an element to which nothing in the subset stands in the relation R. Definition 8.23 R is well-founded on A iff V'x((x --.Rzy))).
~
A & x
i-
0) ~ 3y(y E x & V'z(z E x
->
The relation < is clearly well-founded on the set of natural numbers. In any subset of the natural numbers there will be a smallest number. The same relation, however, isn't well-founded on the set of integers. The subset consisting of the negative numbers, for example, has no least element. Similarly, < is not well-founded on the reals, or even on the real interval (0,1); there is no real number closest to O. R well-orders A just in case R is well-founded and connected on A. The less-than relation, for example, well-orders the natural numbers. Every well-ordering is a strict linear ordering.
Theorem 8.7 Every well-ordering of a set is a strict linear ordering of that set. Well-orderings are connected. There are well-founded relations, however, that are not connectedviz., partial well-orderings. R is a partial well-ordering of A iff R is both well-founded and transitive on A. It's easy to show that every partial well-ordering of A is a strict partial ordering of A. (Transitivity is trivial, and the proof of irreflexivity is exactly as above.) Every well-ordering is a partial well-ordering, but the reverse is not true. On the set of people, "is an ancestor of" and "is a descendant of" are partial well-orderings, but they aren't full well-orderings, since neither is connected: not any two people are linked by the "ancestor" or "descendant" relation.
Problems Classify these orderings. l.
< on N
2. :::; on N
3.
~
onN
4. ~ on N
5. = on N 6.
< on Z
7. ~ on Z
8. :::; on Z 9. 10.
~
on Z
< on the reals
Logic, Sets and Functions
184 11.
:s on the reals
12.
:s on [0,1] :s on [0,1] :s on (0,1)
13. 14.
15. < on (0,1) 16. divides, on N 17. divides, on Z 18. is a factor of, on N 19. is a multiple of, on Z 20. is a power of, on N 21. E, on the transitive sets 22.
~,on
sets
23. C, on sets 24. is a subformula of, on formulas of Q 25. follows from, on formulas of Q 26. are equidistant from point a, on 2-dimensional plane 27. is at least as far from point a, on 2-dimensional plane 28. is farther from point a, on 2-dimensional plane 29. is closer to point a, on 2-dimensional plane 30. finishes higher than, on set of baseball teams Prove these theorems. Suppose that, for any x and y in A, Rxy iff Sxy for problems 31-34 31. If R partially orders A, so does S. 32. If R orders A, so does S. 33. If R is a strict partial ordering of A, so is S. 34. If R is a strict linear ordering of A, so is S. 35. Let Rand S be strict linear orders on A and B, respectively. Define a relation Q on A x Bas follows: Ifxlfylfzlfw(Q (x, y) (z, w)) R (Rxz V (x = z & Syw))). Show that Q is a strict linear order on A x B. 36. If Rand S are strict linear orderings on A and S
~
A x A, then R
~
S iff R = S.
Relations
8.6
185
Relations between Relations
In this section we'll introduce a new kind of syntactic entity, a complex predicate, to reveal relations among relations. Obviously, there is a close relationship between < and >, or ::; and 2::, or "parent" and "child". If n < m, then m > n. Similarly, if Hank is Alice's child, then Alice is Hank's parent. In short, whenever one relation holds in one direction, the other relation holds in the other. Such relations are converses. Definition 8.24 The converse of R, R- 1 , is the collection of all ordered pairs (y, x) such that (x,y) E R. Symbolically,
R- 1 = {(y,x): (x,y
E
R)}
Conv
Conv Every ordered pair in a relation is in its converse, but with the constituents switched. That is, all first constituents become second constituents, and vice versa. Thus, if (John, Bill) E "kick", (Bill, John) E "kick" -1, i.e., "be kicked" . Theorem 8.8 For all R, 1. V(R) = R(R- 1 ), 2. R(R) = V(R- 1 ), and 3. if R ~ A x B, then R- 1 ~ B x A.
Proof. Suppose a E V( R). Then Rab, for some b. So, R-l ba. Thus, a E R(R- 1 ). (The rest of the proof proceeds similarly.) • . A more complex relationship is exemplified by "parent" and "grandparent" or by "uncle', "brother" and "parent". A grandparent is a parent of a parent. An uncle is a brother of a parent (or a husband of a sister of a parent). We can analyze many relations such as "grandparent" and "uncle" as compositions. We'll write the composition of Sand T as "S 0 T". Definition 8.25 SoT = {(x, y): 3z( (x, z) E T & (z, y) E S)}.
Comp SoTxy 3z(Txz & Szy)
Comp
To say that Ned is Jed's grandparent, then, is to say that Ned is the parent of somebody who is the parent of Jed. "Grandparent" is the composition of parent with itself. "Uncle" is actually the union
186
Logic, Sets and Functions
of two different compositions. If Lou is Hanna's uncle, it may be that Lou is the brother of one of Hanna's parents. "Uncle" in this sense is the composition of "parent" and "brother". Of course, Lou may be the husband of a sister of a parent of Hanna; in this sense "uncle" is the composition of the composition of "parent" and "sister" with the relation of "husband". These senses are roughly those of "uncle by blood" and "uncle by marriage"; "uncle" per se is the union of the two. Since binary relations are sets of ordered pairs, usual set-theoretic operations apply to them. So we can also form new relations from old by taking unions, intersections, complements, differences, etc.
Some examples: 1. Shew (RoS)-l = S-l oR- 1
[by Ext]
2.
Shew
~
3.
a
E
(R
4.
a
E
{(x,y): RoSyx}
5.
a = (b,c) & RoScb
CAbstraction, 4, 3E
6.
Sed & Rdb
Camp, 5, 3E
7.
S-ldc & R-1bd
Conv, 6
8.
(b,c)
9.
a
10.
ShE>w
11.
etc.
E
S-l
S-1
E
S)-l
0
0
0
R- 1
R- 1
Def. Conv, 3
31,7, Comp
=E, 5, 8
+-
1. Assume R is irreflexi've on A. 2. Shew R- 1 is irreflexive on A.
3. Vx(x 4.
E
A
~
-,Rxx)
ShE>w Vx(x E A ~ -,R- 1 xx)
5.
bE a
6.
-,Rbb
7.
Shew -,R- 1bb
8.
R- 1bb
9.
Rbb
Conv, 8
-,Rbb
R,6
10.
Relations
11.
187
R- 1 is irreflexive on A.
Irrefl., 4
1. R is transitive on A
Assumption
2. Shew R- 1 is transitive on A 3.
'Vx'Vy'Vz(x,y,z E A ---.. ((Rxy & Ryz) ~ Rxz)
4.
Shew 'Vx'Vy'Vz(x, y, z E A ~ ((R-1xy & R-1yz) ~ R-1xz))
5.
Trans., 1
ShElw b, C, dE A ~ ((R-1bc & R-1cd) ~ R-1bd)
6.
b,c,d E A
7.
R-1bc & R-1cd
8.
Reb & Rdc
Conv, 7
9.
Rdb
Trans., 1, 'VE, 8, SL
R-1bd
Conv, 9
10. 11.
R- 1 is transitive on A
1. Shew R
0
Trans., 2
S is reflexive on A if both Rand S are reflexive on A
2.
R is reflexive on A & S is reflexive on A
3.
Shew 'Vx(x E A
4.
Shew b EA
~
~
(x, x) E R
0
S)
(b, b) E R 0 S
5.
bE A
6.
Rbb
Refl, 2, 5
7.
Sbb
Refl, 2, 5
8.
3x(Sbx & Rxb)
&I, 6, 7, 3I
9.
(b,b) E Ro S
Camp, 8
1. Rand S are strongly connected on A 2. Shew R 3. 4.
0
Assumption
S is strongly connected on A
ShElw 'Vx'Vy(x, yEA Shew b,c E A
~
~
(R
0
Sxy V R
0
(RoSbcV RoScb)
Syx))
188
Logic, Sets and Functions
5.
b,eE A
6.
Sbb
St. Conn, 1, 5, SL
7.
Sec
St. Conn, 1, 5, 8L
8.
Rbev Reb
St. Conn, 1, 5
9.
Case 1: Rbe
10.
RaCbe
&1, 6, 9, ::II, Comp
11.
RaSbeV RoSeb
vI, 10
12.
Case 2: Reb
13.
RaSeb
&1, 7, 12, ::II, Comp
14.
R a Sbe V R a Seb
vI,13
R a Sbe
15.
V
R
0
Seb
Problems What is the converse of each of these relations? 1. < 2.
<
3. > 4. >
5. 6.
i-
7. E 8. C
9. C 10. divides 11. is a multiple of 12. is absolute value of (on Z) 13. is the square of (on N) 14. is the square of (on Z) 15. is taller than 16. is at least as tall as
VE, 8, 11, 14
Relations
189
17. is exactly as tall as 18. has the same birthday as 19. is parent of 20. is child of 21. is grandparent of 22. is uncle of 23. is cousin of 24. is related to 25. hates
Prove the following theorems. Assume that R, 5, T and V are binary relations on the set A. 26. The converse of R is reflexive on A if R is.
(!)
27. The converse of R is irrefiexive on A if R is. 28. The converse of R is symmetric on A if A is. 29. The converse of R is asymmetric on A if R is. 30. The converse of R is antisymmetric on A if R is. 31. The converse of R is transitive on A if R is. 32. The converse of R is connected on A if R is.
33. The converse of R is strongly connected on A if R is. 34. The converse of R is intransitive on A if R is.
35. R 05 is reflexive on A if both Rand 5 are. 36. R
0
5 is strongly connected if both Rand S are.
37. The converse of the converse of R is R. 38. (R U 5)-1
=
R-1 U 5- 1.
39. (R n 8)-1
=
R- 1 n 5- 1.
40. (R - 5)-1
= R- 1 -
5- 1.
41. Ro 0 = 0.
42. D(8
0
R)
~
D(R).
43. (80 R)-l = R- 1 08- 1. 44. To (80 R) = (T 0 8) 0 R.
45. (R
~
8 & T
~
V)
~
ToR
~
V
0
8.
Logic, Sets and Functions
190 46. (8 U T) 0 R
= (80 R) U (T 0 R).
47. (8 n T) 0 R
~
(80 R)
n (T 0
R).
48. (8-T)oR2(8oR)-(ToR).
49. (A x B)-l = B x A. 50. (A x B)
0
(A x B)
~
A x B.
51. \fx(x E D(R) ~ (x, x) E R- 1 0 R). 52. If R is reflexive on A, then D( R)
53. R is symmetric on A iff R
=
= D( R- 1 ).
R-l .
54. R is asymmetric on A iff Rand R- 1 are disjoint. 55. R is antisymmetric on A iff R n R- 1 is a subset of the identity relation.
56. R is transitive on A iff R
0
R
~
R.
57. R is strongly connected on A iff R
U
R- 1 = F(R)2.
58. R is reflexive, symmetric and transitive on A iff R- 1
59. R is symmetric and transitive on A iff R = R
0
0
R = R.
R- 1 .
Given these relations for Rand 5, what is 80 R? (Assume that the relations are defined on the integers, for 64-77; on sets, for 78-81; and on the set of people, for 82-92.)
R
8
60.
<
<
61.
<
<
62.
<
63.
<
>
64.
<
>
65.
<
=/:
66.
>
>
67.
>
>
68.
>
69.
>
=/:
70.
>
<
=/:
=/:
71.
72.
'\' '111'."
~
c' ~~
Relations
191
73.
>
74.
E
C
75.
E
E
76.
C
C
77.
C
E
78.
parent
brother
79.
brother
parent
80.
child
parent
8l.
child
child
82.
brother
child
83.
child
brother
84.
mother
spouse
85.
spouse
mother
86.
spouse
brother
87.
brother
spouse
88.
spouse
spouse
8.7
Restrictions and Images
The < relation is a strict linear ordering of the real numbers. Nevertheless, we've often spoken in this chapter of the properties of < on the natural numbers or the integers. These properties, as we've observed, can differ; < is dense on the real line but not on the integers. Nevertheless, < on the integers is simply a subset of < on the reals. To obtain < on Z from < on the real numbers, in fact, we have only to eliminate all the ordered pairs comprising any nonintegral reals. In short, we restrict both the domain and the range of the relation to the set of integers. We can, and often do, restrict other kinds of relations. The "parent" and "child" examples in this chapter have dealt, implicitly or explicitly, with the domain of people. But procreation far outreaches humanity. The relation makes perfectly good sense when applied to cats, dogs, and rabbits. The parent-child relation among humans, furthermore, is a subset of the broader parent-child relation linking any animals biologically connected in the appropriate way. Formally, we can say that the relation among humans is the restriction of the broader relationship to humanity. In general, we restrict a relation to a set by omitting all ordered pairs in the relation whose first constituents fall outside the set. The restriction of 5 to a set A is written "5 ! A". Definition 8.26 5!A = {(x,y): x E A & (x,y) E S}.
192
Logic, Sets and Functions
1 (Restriction) SlAxy Sxy & x E A
1
The restriction of the parent-child relation to the set of humans has the effect we want, since people never have other kinds of animals as children. In the case of <, however, restricting the domain to the integers is not enough. Pairs such as (1r,5) will be dropped, but pairs such as (3,1r) will remain. If we want to eliminate them as well, we must also restrict the converse relation to Z. An important related concept to restriction is what we might call "super restriction." The super restriction of a relation to a set omits all those pairs in the relation whose first or second constituents fall outside the set. The super restriction of S to a set A is written "SIIA". •
Definition 8.27 SIIA = {(x, y):
X
E A & yEA & (x, y) E S}.
II
(Super-restriction) SIIAxy Sxy & x E A & yEA
Suppose now that we want to consider the relation "x contributed to y's Presidential campaign", where the domain is the set of Political Action Committees and the range is the set of Presidential candidates. We can examine the activities of, say, labor-union PACs by restricting this relation to the set of labor-union Political Action Committees. Presumably, these committees will have made contributions to some campaigns and not others. Those candidates receiving funds from labor PACs will appear as second constituents in the ordered pairs in the restricted relation. The set of all such candidates-that is, all such second constituents-is the image of the set of labor-union PACs under the relation. More generally, the image of a set A under a relation R is the set of things that R relates to any members of A. It is, in other words, the range of the relation restricted to A. We'll write the image of A under R as "R[A]".
Definition 8.28 R[A] = {x: 3y(y E A & Ryx)}
[I x
E R[A]
3y(y E A & Ryx)
[]
To return to our example, R[A] is the set of candidates receiving campaign funds from any Political Action Committee in A. The square brackets here indicate that we are concerned with a set rather than an object in the field of the relation.
Relations
193
What if we want to know which Political Action Committees gave money to a particular set of candidates, say, Republicans? In this case we can imagine ourselves restricting the range of the relation to Republican politicians, and seeing which PACs appear as first constituents of the remaining ordered pairs. The set of PACs giving money to Republicans is the converse or inverse image of the set of Republicans under the relation. The converse image or inverse image R- 1 [A] of a set A under a relation R is the collection of objects R-relating to something in A. Definition 8.29 . R-1[A] = {x: 3y(y E A & Rxy)}
INV I
3y(y E A & Rxy)
Inv I
If A is the set of Republicans, then R-1[A] is the set of PACs giving money to Republicans. The notation here points out that the inverse image of A under R is the same as the image of A under R- 1 : both are written R- 1 [A]. To see that they are the same, recall that the converse of a relation "flip-flops" the ordered pairs of the original relation. The domain of the converse is the range of the original, and vice versa. The inverse image of A under R is the set of all x such that, for some y in A, Rxy. The image of A under R- 1 is the set of all x such that, for some y in A, R-1yx. But, according to the definition of converse relations, these come to the same thing.
Some examples. 1. ShffiV (R n S)[A] ~ R[A] n S[A]
2.
bE (Rn S)[A]
3.
c E A & (R n S)cb
[ J, 2
4.
(c,b) E RnS
NI,3
5.
(c,b) E R & (c,b) E S
n,4
6.
:3x(x E A & Rxb)
31,3,5
7.
bE R[A]
[ J,
8.
bE S[A]
:31, 3, 5, [ )
9.
b E R[A] n S[A]
7,8, n
1. A ~ V(R) 2. Shew A ~ R- 1[R[A]] 3.
ShffiV \/x(x E A
->
x
E
R-1[R[AJ])
6
Logic, Sets and Functions
194 4.
bE A
5.
bE V(R)
<;:;*,1,4
6.
Rbc
V, 5, 3E
7.
3y(y E A & Rye)
&1,4,6,31
8.
eE R[A]
[ ], 7
9.
3y(y E R[A] & Rby)
&1,6,8,31
bE R-l [R[AJ]
Inv 1,9
10.
1. Shew R[A] = R(R!A)
2.
(by ExtJ
Shew 'rfx(x E R[AJ R R(R!A))
3.
Shew ---.
4.
bE R[A]
5.
cEA&Reb
[ ], 4, 3E
6.
R!Aeb
1,5
7.
bE R(R1A)
31,6, R
8.
Shew
9.
etc.
~
Problems Prove the following theorems. Assume that A, B, C, D, E and F are sets, R <;:; 1. R1A
= R n (A
x R(R)).
2. If A <;:; B, then R1A <;:; R1B.
3. Rl(A n B) 4. Rl(A
u B)
= R!A n R!B. = R!A
u R!B.
5. Rl(A - B) = R1A - R1B. 6. (50 R)lA = 50 (R1A).
7. R[AJ = R(R!A). 8. R[A U B] = R[AJ U R[BJ. 9. R(A n BJ <;:; R[A] n R[BJ. 10. R[A - B] ~ R[A] - R[B].
ex D and 5
<;:; Ex F.
Relations
195
11. If A ~ B, then R[A] ~ R[B].
12. (R U S)[A]
= R[A] U SIAl.
13. (R n S)[A] ~ R[A] 14.
R[0]
=
n SIAl.
0.
15. R- 1 [0] = 0.
16. R[A] = 0 iff A and D(R) are disjoint. 17. V(R)nA~R-l[R[AlJ. 18. R[A]
n B = R[A n R-1[BlJ.
19. R[V(R)] = R(R).
20. R-1[R(R)] = V(R).
21. R1V(R) = R.
22. R-11R(R) = R- 1 .
8.8
Equivalence Relations and Partitions
Equivalence relations are reflexive, symmetric and transitive. It follows that they are Euclidean as well. Identity, for example, is an equivalence relation. So are equivalence among logical formulas or sentences of English; equidistance from a point in a three-dimensional space; and "is exactly as tall as" or "has the same birthday as" on the set of people. Definition 8.30 R is an equivalence relation on A iff R and transitive on A.
~
A x A and R is reflexive, symmetric
Equivalence relations on a given set divide the set into various groups. Within each group, everything bears the equivalence relation to everything else. But no links extend across groups; nothing bears the equivalence relation to something in another group. Consider, for example, the relation "has the same birthday as" on the set of people. This equivalence relation splits the set of people into 366 groups, one for each day of a leap year. Group #1, let's say, contains all and only those people whose birthday is January 1. Everyone in this group has the same birthday as everyone else in the group. But, clearly, nobody in group #1 has the same birthday as anyone in any other group. To take another example, think of the relation of equidistance from a point in a three-dimensional space. This relation splits the space into an infinite number of spheres centered on the given point. All the points on any particular sphere are equidistant from that point. But no point is the same distance from the given point as a point on another sphere. Such a collection of spheres partitions a space. One set A partitions another, B, just in case it meets three conditions. First, A must be a collection of mutually exhaustive subsets of B; the union of A (the set of members of members of A) should be B. Second, every member of A should be nonempty. Third, any two distinct members of A should be disjoint. A, therefore, must be a set of nonempty, disjoint and mutually exhaustive subsets of B. Definition 8.31 A set A partitions a set B (or is a partition of B) iff 1. the union of A is B (UA = B);
Logic, Sets and Functions
196 2. every member of A is nonempty: (Vx(x E A
-->
:Jy y
3. A is pairwise disjoint: (VxVy((x E A & yEA & x
-=I-
E
x)); and
y)
-->
x n y = 0)).
Clause (1) guarantees that the subsets must be mutually exhaustive; it specifies that every member of B must belong to at least one subset in A. Clause (2) guarantees that the empty set is not parading as a member of the partition. And clause (3) requires that the subsets in A be nonoverlapping. To repeat, then: a partition is a set of nonempty, mutually disjoint and exhaustive subsets. As our examples have indicated, partitions and equivalence relations have intimate connections. The equivalence relation "has the same birthday as" partitions the set of people; the relation of equidistance from a point partitions a plane. In general, any equivalence relation on a set induces a partition on that set. Notice that all those people in group #1, having a birthday on January 1, have the same birthday as everyone else in the group, but don't have the same birthday as anyone outside the group. The group, in other words, is precisely the image, under the equivalence relation, of each member of the group. (More precisely, it's the image of the unit set of each member of the group; images are defined for sets.) Every group, therefore, is its own image under the equivalence relation. Such groups are called equivalence classes. In what follows we will use the following shorthand formulas to expedite matters. The partition induced by R on a set A will be written as "IIA(R)" and this will stand for the set of equivalence classes: IIA(R) = {R[{y}]: YEA}. Theorem 8.9 If R is an equivalence relation on A and A is a set, then II(R) is a partition of A. Any equivalence relation on a set, therefore, induces a partition on that set. A partition of a set also induces an equivalence relation. Say that we divide a set of people into various groups. No matter how arbitrary the division seems, there is an equivalence relation corresponding to the partition: "belongs to the same group as". We'll write this relation corresponding to the partition II as "R A (II)". "R(II)", in other words, is simply the set of ordered pairs, . The equivalence relation induced by II on A will be written "R A (II)" and is the set of ordered pairs:
RA(II) = {(x,y) E uII x uIT: 3z(z E IT & x E z & y E z)} Theorem 8.10 If II is a partition of A, then R(IT) is an equivalence relation on A. Problems 1. How many possible partitions are there of a unit set? What are they?
2. How many possible partitions of 0 are there? 3. How many possible partitions are there of a set of just two elements? List them. 4. How many possible partitions of a set of three elements are there? List them. 5. How many possible partitions of a set of n elements are there? Prove the following theorems. IIA(R) and RA(II) are the same sets mentioned above in the chapternamely {R[{z}]: Z E A} and {(x,y): :Jz(z E II & x E z & y E z)} respectively. 6. Suppose that a relation R is symmetric and transitive on the set A. Then R is an equivalence relation on F(R).
Relations
197
7. If a set A is nonempty, then {A} is a partition of A. 8. The converse of an equivalence relation on A is also an equivalen~e relation on A. 9. Let a, bE A, and let R be an equivalence relation on A. Then R[{a}] = R[{b}] iff Rab.
10. If R is an equivalence relation on a set A, then, for any x, YEA, R[ {x} 1 and R[ {y}] are either
identical or disjoint. *11. Let B partition the set A, and let R be an equivalence relation on A. Then the partition uniquely determined by R is B-i.e., B = {R[{z}]: z E A} iff R is the equivalence relation uniquely determined by B-i.e., R = {(x, y): 3z(z E B & X E z & y E z)}.
12. If B partitions the set A and C E B, then C
~
A.
13. Where x, y, Z, WEN, define R (x, y) (z, w) as x +W = Y + z. Then R is an equivalence relation
on N x N. 14. Let Rand S be equivalence relations on a nonempty set A. Then R n S is also an equivalence relation on A. 15. If R is an equivalence relation on a set A, then IIA(R) = IIA(R-l). 16. Suppose that R is an equivalence relation on set A and B E IIA(R). Then B = R[B].
Chapter 9
Functions We've been dealing with the concept of a function informally throughout much of this text. This chapter, however, will try to make the notion precise. Functions are vitally important in many different branches of mathematics and computer science. In spite of their centrality to mathematics, only in the nineteenth century did mathematicians begin to examine functions carefully and replace intuition with logic in treating them. Gottlob Frege, Richard Dedekind, Giuseppe Peano and Russell and Whitehead's Principia Mathematica did much to clarify the notion of a function.
9.1
Functions and Relations
A function is a relation of a special kind. A function takes a certain number of inputs or arguments and yields an output or value. Thus the addition function relates the inputs 3 and 4 to the output 7. Given the arguments 8 and 14, it yields the value 22. The squaring function takes 2 as input, for instance, and yields 4; takes 4 and yields 16; takes 16 and yields 256; and so on. An n-ary function relates n arguments or inputs to a value. Addition, as often construed, is a binary function; the squaring function is singulary. This chapter will focus almost completely on singulary functions: those which take a single input and yield a single output. To see what makes a relation a function, we must consider some additional properties of relations. A relation f is a function from a set A to a set B iff 1) f relates each object in A to at least one object in B, 2) f relates at most one object in B to each object in A, and 3) f relates no object in A to an object outside of B. A relation that relates everything in A to exactly one thing in the set B is thus a function from A into B. This is captured in the following definition: Definition 9.1 R is a function from a set A to a set B (symbolically, R: A
-->
B iff
1. V(R) = A;
2. VxVyVz((Rxy & Rxz)
3. R
<;;;;
~
y
= z);
and
A x B.
The relation of a person to his or her mother, excluding Adam and Eve, is a function from the set of persons into the set of persons. Everyone has exactly one mother. Presumably every U.S. citizen over a certain age has a Social Security number, and nobody has two; the relation of a person to his or her Social Security number is also a function, from the set of U.S. citizens over some age to the set of 9-digit numbers. 199
Logic, Sets and Functions
200
A function from A into B is also called a map, or mapping, of A into B. Though functions are just relations that satisfy a requirement of functionality, the notation for functions differs significantly from the usual notation for relations. Recall that a relation is a set of ordered pairs: Rab and (a, b) E R are equivalent expressions. Because a function is a relation, it too is a set of ordered pairs. The first constituents are the arguments or inputs of the function; the second constituents are the outputs or values. It's natural, then, to write as equivalent fab and (a, b) E f. But the most common notation for functions writes the same idea as f(a) = b. These notations are not strictly interchangeable, however. While our rules for relations say it is perfectly licit to move from fab to (a, b) E f, we must be careful when we use the notation f(a) = b. This is because it is not merely the fact that f is a relation which makes this notation permissable, but that f is a function. Suppose, for example, that we know that fab and fac and b i= c. If we allow ourselves to move freely between these three types of notation, we can show that b = c, which is a contradiction. If we allow such free movement only when we know f is a function, however, the contradiction cannot be reached. This is because when f is a function and fab and fae, then b must in fact be identical to c. We therefore incorporate this restriction into our rules for function notation interchange.
Func
f:
C ------; D
fab f(a) = b
f:
C ------; D
C f(a) = b fab
aE
Func
There is another difficulty with the notation for functions-viz., with function symbols. As we have seen, functions are sets (of ordered pairs). As sets functions don't and in fact can't contain a pair (x, y) for every object x that our theory countenances. This accords perfectly well with standard usage. A mathematician who was asked "What is the positive square root of your favorite niece?" would take the speaker to be making some sort of joke. But the language of function symbols has its problems in this respect. It is a theorem of our first order logic with function symbols and identity that 't/x3y f(x) = y. So the singular term "f(a)" denotes an object even when f is the function, positive square root, and a is your favorite niece! As with many problems that surface when we try to systematize our informal way of talking into rigorous logic, there are many possible solutions. One very drastic solution to our current problem is to change the basic logical rules for quantification when dealing with function symbols. We won't do that here. Instead, we will simply leave the denotation of "f(x)" undefined if either f is not a function or x ~ R(f). In any interpretation of our set theory by means of a model, we can follow the policy of the logician Frege and assign any term "f(x)" whose value is undefined a special Dr-element * , which is outside of the intended domain of our theory. This ploy ensures that any function term "f", when combined with any singular term r, is a proper singular term. We will never mention this peculiar object * within our set theory.
Functions
201
The following are useful derivable rules for functions:
FnV
fcd
f: A fed
eEA
dEB
f: A
->
Func*
FnR
B
----->
B
f: A eE
----->
A
fef(e)
Func = B
f: A
----->
B
fed & fee d=e
The following new formation rules will have to be added to our formal language, reflecting the fact that we are allowing names of functions to have a dual life, one as the name of a set (a kind of binary relation) and another as a function symbol (as in Chapter 6 ).
If t and t' are closed terms, then t(t') is a closed term (a function application term). If to, t 1 and t2 are closed terms, then tot 1t2 is a sentence (abbreviates (to, tl) E to). Problems Which of these relations R <;;; A x B are functions from A to B, for the indicated sets A and B? Explain.
A
B
1. x
N
N
2. x+1
N
N
3. x-I
N
N
4. x-I
Z
Z
Ixl
z
6. Ixl
z z
N
7. x 2
N
N
8. x 2
Z
N
9. x 2
R
R
Rxy +---+ y =
5.
10.
ft
N
N
11.
ft ft ft
N
Z
Z
Z
R
R
N
N
12. 13.
14. 2x
202
Logic, Sets and Functions
15. 2x
Z
16.
XX
N - {O} N - {O}
17.
XX
Z - {O} Z - {O}
18.
XX
R - {O} R - {O}
Z
19. y!X
N - {O} N - {O}
20. yIX
Z - {O} Z - {O}
21. yIX
R-{O} R-{O}
22. x 2
+ 2x + 1
N
N
23. x 2
+ 2x + 1
R
R
24. x!
N
N
25. x!
Z
Z
26. x!
R
R
27. In(x)
R
R
28. x/2
N
N
29. x/2
R
R
30. -x
N
N
31. -x
Z
Z
32. -x
N
Z
33. -x
Z
N
34. l/x
N
R
35. l/x
R
R
36. 1/1n(x)
R
R
37. sinx
R
R
38. cos x
R
R
39. tan x
R
R
40. sinh x
R
R
Prove the following theorems. Assume that f: A A, B, C, D are all sets. 41. Do not do this problem. 42. go f: A ---. C. 43. Vx(x E A
-+
(g
0
f)(x) = g(f(x))).
(!)
--->
B, g: B
--->
C and h: C
--->
D, and that
Functions
203
,.,.
f) = (h
44.'1 h
0
(g
45. f
U
g: Au B
46. f
~
47. If f
0
--->
9 iff (D(f)
~
0
g) 0 f.
B u C iff Vx(x
~
D(g) and Vx(x
9 and D(g)
48. If D(f)
~
~
A Ii B
~
f(x) = g(x»).
D(f)
~
f(x) = g(x))).
E
D(f), then f = g.
= D(g) and Vx(x
*49. If Vf\ig(f, 9 E A
E
E
D(f)
(3x3y(f: x
~
--->
(!)
f(x) = g(x», then f = g.
y) & (f
~
9V9
*50. Let R be an equivalence relation on B, and Q Q is an equivalence relation on A.
*51. Suppose that VxVy(x, yEA A.
9.2
-+
~
I))), then 3z3w(uA: z
= {(x,y):
x,y
E
---7
A & (f(x),g(x»)
w).
E
R}. Then
(Rxy 3 f(x) = f(y))). Then R is an equivalence relation on
Properties of Functions
Because functions are relations, all the properties of relations we discussed in the last chapter also apply to functions. In general, however, these properties are not very interesting when so applied. Reflexivity, for example, is an important property for classifying relations. Yet the only reflexive function on any set is the identity function on that set. (The identity function is the function mapping every object into itself.) It's easy to see why. If f is reflexive, then every pair (a, a) E f, where a E D(f). So f maps every object into itself. Furthermore, by clause (2) of the definition of function-hereafter "clause (2)", f maps every object into at most one thing-namely, itself. So f is just the identity function. More interesting properties of functions stem from thinking about a simple picture. Functions, being relations, relate objects in one set to those in another set. We might imagine that some objects go unrelated, while others relate to more than one object. Diagram 1 shows several possible cases. In (1), some object in the set A relates to more than one object in B. In (2), some object in A relates to nothing in B. In (3), some object in B relates to more than one object in A. Finally, in (4), some object in B relates to nothing in A.
204
Logic, Sets and Functions
1
2
~
-
'-
'-
A
B
A
B
4
3
Diagram 1: If we construe R as a subset of Ax B, it's clear that in neither (1) nor (2) can R be a function. In (1), R does not even meet clause (2). It relates some object in A to two different entities in B. In (2), R may meet clause (2), but A isn't its domain; there is something in A that it fails to relate to anything in B. In (3) and (4), however, R can be a function from A to B. In speaking of functions, then, we restrict ourselves to speaking about two of these four possible cases. Suppose we also decided to rule out circumstances such as (3). Suppose, that is, that we don't want to talk about functions that take two different inputs in A to the same output in B. We can do this by limiting ourselves to speaking of injective functions.
Definition 9.2 A function f: A
--+
B is injective (or one-one) iff it maps distinct arguments
into distinct values:
Inj
f : A --+ B is injective VxVy(x, yEA --> (x =f. y --> f(x) =f. f(y))) & f : A
--+
B
Inj
The squaring function (f(x) = x 2 ) is injective on the natural numbers: distinct natural numbers have distinct squares. It is not injective, however, on the set of integers: both nand -n, for any n, have the square n 2 . The relation is still a function on the integers, because every integer has exactly one integer as its square. Each input still yields one and only one output. But some outputs result from more than one input, so the function on the integers isn't injective. The relation between a person and his or her mother is a function, since everyone has exactly one mother, but it isn't injective. Several different people may have the. same mother. A functil;m mapping people into their eye colors also fails to be injective, assuming that there are people who have the same color eyes. The function relating people to their Social Security numbers, in contrast, ought to be injective; no two people should have the same number.
Functions
205
Suppose now that we're willing to accept cases in which two distinct objects ~ap into the same output. We're concerned, however, to rule out cases in which objects in the set B go unrelated. That is, say we want to rule out cases such as that depicted in diagram (4) above. We can require that f: A ---. B relate something or other in A to everysingle object in B by requiring that B be the range of f. In so doing, we're limiting ourselves to talking about surjective functions.
Definition 9.3 A function f: A for some input.
--t
B is surjective (or onto) iff everything in B is a value of f
Surj
f : A ---. B is surjective B ~ R(J) & f : A ---+ B
Surj
If f is a function from A to B, then R(J) ~ B by definition. If f is surjective, then B ~ R(J), soB=R(Jk The squaring function from N to N is not surjective, because not every natural number is the square of a natural number. 5, for example, is not the square of any natural number. The squaring function is surjective, however, from the nonnegative reals to the nonnegative reals. The doubling function f (x) = 2x is surjective on the reals; every real is equal to two times some other real. The function mapping people into their birthdays is undoubtedly surjective; every day is somebody's birthday. To return to our diagrams for a moment: (1) and (2) depict relations that aren't functions, violating clause (2) and clause (1) of the definition of function, respectively. (3) depicts a function that fails to be injective. (4) depicts a function that fails to be surjective. No diagram depicts a function that relates everything in A to one and only one object in B, and every object in B to one and only one object in A.
Functions that do this- that are both injective and surjective--are bijective functions. They map distinct inputs into distinct outputs. They furthermore leave no member of B unrelated. Definition. A function f: A ---+ B is bijective (or a one-to-one correspondence) iff f is both injective and surjective (onto B).
Bijective functions are called one-to-one correspondences because they establish a correspondence between members of A and members of B. Each element of A is mapped into one and only one object in B; every object in B is a value of the function for one and only one object in A. The doubling function on the reals is bijective; it correlates each number n with one and only one number, 2n. Each number n is also the value of the function for exactly one input, namely, n/2. This is a special case, since the set being mapped and the set being mapped into are identical. A bijection f: A - - t A is called a permutation. Equivalently, a function f: A ---. B a permutation iff it is bijective and A = B. The doubling function is a permutation of the real numbers.
Logic, Sets and Functions
206
The following are useful derivable rules.
Surj*
Inj*
f: A
f: A --+ B is injective c,d E A c=/=d f (c) =/= f (d)
~
B is surjective
cE B
3x(f(x) = c & x
E
A)
Func Camp Biject Conv f:A~B
f: A f-I:
B is bijective B --+ A is bijective
g:B~C
--+
xEA go f(x)
=
g(f(x))
Inj Conv f: A cE
~
B is injective
A
f l(f(c))
= c&
f
1:
RCn -- A
is bijective
Proof of Bijec Conv:
1. f: A
~
B is bijective
2. Show f-l: B 3. f: A
~
4. f: R(f)
--+
A is bijective
B is injective ~
A is bijective
Func Conv
5. f: A - - B is surjective 6. B C;;; R(f) 7. B
= R(f)
DeL func, Th1, 6 =E, 4,7
8. f-l: A - - B is bijective
Problems Say whether these functions f: A for the indicated values of A and B.
f(x) =
A
B
N
N
1
N
N
3. x-I
Z
Z
Ixl
Z
Z
1. x 2.
4.
x+
--+
B are injective, surjective, bijective, or none of the above,
207
Functions
Z
N
6. x-
N
N
7. x 2
Z
N
8. x 2
R
R
N
R
R+
R
11. 2x
N
N
12. 2x
Z
Z
13. 2x
R
R
14. XX
Z+
N
15. XX
Z+
Z
N
R
17. x 2 + 2x + 1
N
N
18. x 2 + 2x + 1
R
R
19. xl
N
N
20. In(x)
Z+
R
21. x/2
R
R
22. -x
Z
Z
23. -x
N
Z
24. l/x
R - {O} R
25. sinx
R
R
26. x, rounded to R nearest integer
Z
5. Ixl .)
9. 10.
16.
ft ft
\IX
ft 2': 0) (assuming ft 2': 0)
(assuming
(assuming even roots 2': 0)
Prove the following theorems. 27. If j: A
---+
B is surjective, then 1)(1-1) = B.
28. If j: A
---+
B is injective, then j-l is a function from R(1) to A.
29. Let
f: A
30. If f: A
---+
---+
B. Then f- 1 is a function from B to A iff f is bijective.
B is a bijection, then f-l: B
---+
A is also a bijection.
31. If f: A ---+ Band g: C ---+ D are injective functions and A n C ---+ B n D is also injective.
f n g: An C
---+
B n D, then f n g:
208
Logic, Sets and Functions
32.' Let
R be a strict linear order on A. Let f: A f is injective.
----->
A be such that \::/xVy(x, yEA - (Rxy
->
~jRf(x)f(Y))). Then
33. Say that a function f: R -----> R is strictly increasing iff VxVy(x strictly increasing functions on the reals are injective.
-+
f(x) < f(y)). Then all
34. Where < is a strict linear order on A, let f: A -----> A be such that VxVy(x, yEA - (x < y f(x) < f(y)))· Then Vx\::/y(x, yEA - (f(x) < f(y) - x < y)).
->
35. Let f: A -----> A, f-l: A -----> A be bijections. Suppose that, for some x E A, f(x) -=I- f-l(X). Then, for some x E A, f(f(x)) -=I- x. 36. If f: A
----->
A and g: A
*12. Let f: A -----> B, g: B injection.
----->
----->
*13. If f: A -----> Band g: C ranges, then f U g: A U C
A are both injective, then fog is injective.
P(A), and g(x) -----> ----->
= {y E
A: f(y)
= x}.
Then, if f is bijective, 9 is an
D are injective functions with disjoint domains and disjoint BuD is also an injective function.
***14. Let f: P(A) -----> P(A) be such that, for any x and y, x s:;; y s:;; A - f(x) s:;; f(y). Let B = n(x E P(A): f(x) s:;; x}, and C = U{x E P(A): x s:;; f(x)}. Then f(B) = Band f(C) = C. ***15. Let f: P(A) -----> P(A) be such that, for any x and y, x s:;; y s:;; A - f(x) s:;; f(y). Let B = n{x E P(A): f(x) s:;; x}, and C = u{x E P(A): x s:;; f(x)}. Then \::/x(x = f(x) - B s:;; x s:;; C). Prove the following theorems. Throughout, assume that f: C -----> D and g: E functions. (Note that this doesn't guarantee that f- 1 and g-1 are functions.)
----->
Fare
41. D(f-l)=R(f). 42. R(f-l) = D(f).
43. (f
0
g)-1 = g-1
0
f-l.
44. If f is a bijection, then for all xED, f 45. If f-l: D
----->
0
f-l(X) = x.
C, then, for all x E R(f), f(r 1 (x)) = x.
46. Suppose that \::/x\::/y(x, y E C -+ (Rxy f-+ f(x) = f(y))). Let Cj f be the set of all images of unit sets of members of C ({ R[{x}] : X E C}), and let g(R[{ x}]) = f(x). Then 9 is an injective function from C j f to D.
Chapter 10
Induction Mathematics adds to logic a series of axioms: the axioms of set theory. From set theory virtually all of classical mathematics develops. In practice, however, mathematics also adds a rule of inference. The rule-- mathematical induction - is not really new; it's derivable from logic and set theory. Mathematical induction has proved very useful in a wide range of areas, including number theory, algebra, set theory, logic, and the theory of computation. Despite its name, mathematical induction is a deductive process, not at all like the empirical method of generalizing laws from particular cases that is usually called induction. The French mathematicians Blaise Pascal and Pierre Fermat and the Swiss mathematician Jakob Bernoulli first constructed proofs by mathematical induction in the seventeenth century. Ideas relating closely to mathematical induction have a longer history. The sixteenth-century Italian mathematicction in the seventeenth century. Ideas relating closely to mathematical induction have a longer history. The sixteenth-century Italian mathematician Francesco Maurolycus and the seventeenth-century English mathematician John Wallis used arguments having roughly the form: "A(x) is true of the first case; A(x) is true of the second case; A(x) is true of the third case; and so on for all other cases." The "and so on" part Maurolycus and Wallis left undeveloped: Maurolycus, for example, concludes a proof in which he has established a conclusion for two particular cases, "And in like manner to infinity as the proposition states." These methods have much in common with the practice of ancient Hindu mathematicians, who argued similarly, deriving solutions to series of equations by solving the first, showing how to obtain the second's solution from the first's solution, and then simply saying that the same tactic would work in every case. This technique of "Hindu induction" has similarities to modern uses of induction in mathematics. But the first explicit argument from the nth case to the n+1st case to establish something for all cases occurs in Bernoulli in 1686. Wallis, in 1656, was the first to introduce the name induction for this argument form, thinking of the analogy between generalizing theorems about numbers in mathematics and generalizing laws about matters of fact in natural science. The name mathematical induction didn't appear until Augustus De Morgan introduced it in an encyclopedia article "Induction (Mathematics", in the Penny Cyclopedia) in 1838. Isaac Todhunter popularized the name in his 1866 textbook, Algebra. Frege first showed that the principle of mathematical induction is valid, given the principles of logic and, in effect, set theory.
10.1
The Natural Numbers and Definition by Recursion
Induction is a way of proving that every member of an infinite set of a certain kind has some property. Suppose, for instance, that we want to prove that something is true for every natural number. We might be able to use universal proof. So, we could prove our theorem for an arbitrary
209
210
Logic, Sets and Functions
natural number, and conclude that the theorem holds for all natural numbers. In many cases, universal proof suffices for proving theorems about the numbers. But the set of natural numbers has a structure. The numbers, after all, come in a particular order. The method of universal proof takes no account of this order. It's not surprising, then, that one can prove more things about the numbers by taking their ordering into account than one can by using just universal proof. Induction is a method of proof that relies on the ordering of sets like the natural numbers. First, we need to define the natural numbers within set theory. To begin with, we'll define the number a as the empty set 0. Next, we'll define the successor operation. The successor operation enables us define each of the natural numbers in terms of O. Thus, 1 is defined as the successor of 0, 2 as the successor of the successor of 0, etc. Traditionally, mathematicians have identified the successor of x with x U {x}. For example, the successor of 0 , that is, of 0, is 0 U {0}, which isthe same as {0}. Since the number 1 is the successor of 0, we have defined 1 as {0} or {a}, which comes to the same thing. 2 is the successor of 1, so 2 = {0} U {{0}}, which is identical to the set {0, {0}}, or, in other words, {a, I}. In general, the number n can be identified with the set {a, 1, 2, ... , n -I}. Notice that the set identified with n has exactly n members. Definition 10.1 0 = 0. Definition 10.2 The successor of x = xU {x}.
A natural number is, intuitively, something which is either zero or the successor of a natural number. This won't do as a simple or 'explicit' definition of the natural numbers, since it contains an element of circularity: 'natural number' is being defined in terms of natural numbers. This circularity can be evaded in the following way. First, we will define the notion of an 'inductive set'. A set is inductive just in case it contains the number zero, and, for all y, if y belongs to the set, then the successor of y must also belong to the set. Definition 10.3 A set x is inductive iff 0 E x and, for all y, ify E x, then succ(y) E x. Formally: Vx(x is inductive f-+ Sx&O E x&Vy(y Ex ........ succ(y) EX))
This definition guarantees that every inductive set contains every natural number. We can't define a natural number as anything which belongs to any inductive set, however. An inductive set can contain lots of things besides the natural numbers: the definition only requires that certain things (namely, the natural numbers) be included; it doesn't exclude anything. We can, however, define the natural numbers as those which are in common among all the inductive sets. In other words, we can define a 'natural number' as something which belongs to all inductive sets. Definition 10.4 A set x is a natural number iff x belongs to every inductive set.
Unfortunately, this definition alone is still not adequate. The axioms of set theory we introduced in chapter one are not sufficient to guarantee that there is even one inductive set. In fact, they are not sufficient to guarantee that there is any set with infinitely many members. If there are no inductive sets, then everything belongs to every inductive set, and everything (including you and I) are natural numbers. To provide an adequate basis for the theory of numbers, we must add an additional axiom, traditionally known as the axiom of infinity. Axiom 10.1 (Axiom of Infinity) There exists at least one inductive set. Theorem 10.1 The set N = {x : X is a natural number} is inductive.
Corollary 10.1 If x is a natural number, then so is succ(x).
'-
Induction
211
Corollary 10.2 If x is a natural number and x :j:. 0, then there is a natural number y such that x = succ(y). We can also define the familiar less-than relation, <, in set theory. Given our definitions of the natural numbers, it so happens that n < m iff n E m.
Definition 10.5
Definitions. m < n
+->
mEn. m > n
+->
n < m.
Each natural number m is just the set of numbers less than m, so it's not surprising that '<' can be defined as 'E'. Although we won't do so here, it's possible to prove, in our set theory, all of the usual facts about < on N. For example, we could prove that < is irreflexive, asymmetric, transitive, connected, and well-founded. In other words, we could prove that < is a well-ordering, and therefore a strict linear ordering, on N. In addition, we can prove that every number is less than (a member of) its successor.
Theorem 10.2 < is a well-ordering of N. Corollary 10.3 The law of trichotomy. For all natural numbers nand m, exactly one of the following holds: n < m, n = m, or m < n. Theorem 10.3 Vxx < succ(x) The only thing that remains to be done is to define the familiar arithmetic functions, like addition and multiplication. By exploiting the structure of the natural numbers, we can employ a novel kind of definition, known as 'definition by recursion'. We define a function f by first defining its value for 0 and then specifying how to compute the value of f(succ(n», given the value of f(n). Let's consider for example a typical definition by recursion of addition. Let n, m be natural numbers, and let 'succ' be the successor function for the natural numbers. Then,
n+O=n n + succ(m) = succ(n + m) Had we but world enough and time, we might think of using this recursive definition as building up all the values of the binary function + "from below." By starting with the value of the function at 0, and then continuing on for each number one could, one would think, get the value of + for all numbers. But this is itself not a justification or proof that the function has been explicitly defined for all elements in the set. The proof of the adequacy of definition by recursion is again proved within set theory, and we rely again on the properties of N to prove it. What is really going on when we give a definition by recursion? If you look closely, typically we are defining one function (+ in the example above) by means of another function (successor or 'succ' in the example above). This is what the statement of the recursion theorem takes into account. Let succ be the successor function on N as before.
Theorem 10.4 (Recursion Theorem) Suppose SA, b E A and f : A unique function g: N - 4 A, such that: (i) g(O) = b& (ii) Vx(x E N
-4
g(succ(x» = f(g(x»).
-4
A. Then there exists a
212
Logic, Sets and Functions
The proof of the recursion theorem is somewhat lengthy, and we will also omit it here. l But we can easily correlate the precise statement of the recursion theorem with our example above to explain what is going on. 9 is the function being defined by recursion, f is the function used to define g. To define addition precisely in set theory, we would take A to be the set N, f to be the function succ, and a to be the element O. Notice too that the recursion theorem is stated in terms of functions of one argument. So to define addition as a two place function, we first construct by recursion a function n + for each number n. The true function + would then be the union of all these functions constructed by recursion. Therefore, we define, for each natural number n, a unique function An: N - - t N defined by the conditions: An(O) = n An(succ(k)) = succ(An(k)).
Thus, we've defined the value of An for 0 and we've defined the value for succ(k) in terms of the value for k, as Theorem 10.4 requires. Addition is a function from ordered pairs of natural numbers into N. Our definition of An guarantees that n + m = An(m), for each nand m.
Definition 10.6 (Addition) n
+m =
An(m)
In the case of multiplication, we use Theorem 10.4 to define, for each number m, a unique function M m : N ~ N.
Mm(O)
=0
Mm(succ(n))
= Mm(n) + m.
Multiplication is a function from ordered pairs of natural numbers into N, and we've defined M m in such a way that M m (n) = m . n, for all natural numbers m and n.
Definition 10.7 (Multiplication) m· n = Mm(n) In the case of exponentiation, we define, for each number m, a unique function Em: N
~
Em(O) = 1 Em(succ(n)) = Em(n) . m.
Then, the exponent function Exp (where 'Exp(m, n)' means the same as m n ) is defined as:
Definition 10.8 (Exponentiation) Exp(m, n) = Em(n) The factorial function is defined as follows:
leO) = 1 (succ(n))! = (n
+ 1) . nL
Or, more conventionally: O! = 1
(succ(n))! = (n 1 For
+ 1) . nL
a proof see Halmos (1970).
N.
Induction
213
Finally, we can define the finite sum and finite product operations. Suppose that we are given a function j: N - 7 N. Then we can define the finite sum operation, E~o j(i) by stipulating:
o ~j(i)
I
j(O)
i=O
succ(n)
n
~ j(i)
~ j(i)
i=O
i=O
+ j(succ(n»
Similarly, the finite product, TIf=oj(i) can be defined by stipulating:
= =
j(O) (TIf=oj(i»· j(succ(n»
Many properties of these recursively defined operations are proved by the method of mathematical induction. As was mentioned in the introduction, a proof by induction has the following two-part structure. First, there is a base case, in which the theorem is proved to hold for the single case of the number zero. Secondly, there is an inductive case, in which it is assumed that the theorem holds for an arbitrary number n and the task is to prove that it must also hold for the successor of n. A proof of this form is called a simple weak induction on the natural numbers. For example, suppose that we want to prove a theorem of the form, Vx(x is a natural number - 7 A(x) ), where 'A(x)' is some formula with a free variable 'x'. The stucture of a simple weak inductive proof of this formula would be:
Simple Weak Induction on N Show Vx(x
E
N
-7
A(x»
(i) Base case. Show A(O).
(ii) Inductive case. Let k be an arbitrary natural number. Assume A(k). Show A(succ(k».
Recursive definitions justify inductive proofs in the following way. Whether a particular object satisfies an inductively defined term depends on whether previously defined objects satisfy that term. This dependence of the properties of objects on the properties of previously characterized objects enables one to argue from n to its successor. The corollaries to theorems Th. 10.5 and Th. 10.6 demonstrate how one uses induction to prove facts about functions defined recursively. Theorem 10.5 For natural numbers m, n:
(a) m+O=m
(b) m Proof:
+ succ(n) = succ(m + n)
214
Logic, Sets and Functions
(a) m
+ 0 = Am(O),
(b) m + succ(n) Am(n) = m
=
'+'.
by def. of
Am(O)
= m,
by def. of Am.
Am(succ(n)), by def. of '+'. Am(succ(n)) = succ(Am(n)), by def. of Am. by def. of '+'. So, m + succ(n) = succ(m + n), by =E.
+ n,
Corollary 10.4 : 0 + m = m Proof. By induction. (i) Base case. 0
+0
= O. By Th lO.5(a).
(ii) Inductive case. Let k be an arbitrary number. Assume 0 + k = k. Show
0 + succ(k)
=
succ(k).
0+ succ(k)
=
succ(O + k).
By Th 10.5(b).
=
succ(k).
By indo assumption and =E.
Corollary 10.5 : For natural numbers m, n: succ(m)
+n
= succ(m
+ n).
Proof. By induction on n. (i) Base case. Show succ(m) + 0 = succ(m + 0) By Th 1O.5(a), succ(m) + 0 = succ(m) , and m + 0 = m. Thus, succ(m + 0) = succ(m) also. So, by =E, succ(m) + 0 = succ(m + 0).
(ii) Inductive case. Let k be an arbitrary number. Assume succ(m) + k Show
Theorem 10.6
succ(m)
+ succ( k)
succ(m)
+ succ(k)
+ succ( k )) succ(succ(m) + k), succ(succ(m + k)), succ(m + succ(k)) ,
= succ(m
+ k).
succ( m
by Th lO.5(b). by assumption. by Th 1O.5(b).
For natural numbers m, n:
(a)m·O=O (b) m· (succ( n)) = m . n
+m
Corollary 10.6 : O· m = 0 Corollary 10.7 : succ(m) . n = (m . n)
+n
Corollary 10.8 : m· succ(O) = m Proofs: (left as an exercise). We need the following theorem to show that our common use of "+1" as an abbreviation of "successor of" is harmless. Theorem 10.7 . For natural number m: m
+ succ(O)
= succ(m)
Proof: m
+ succ(O) =
[by Th 10.5(b)] succ(m + 0) succ(m) [since m + 0 = m, by Th 1O.5(a)]
Now we can adopt the usual arabic numerals as abbreviations:
Induction
215
Notational equivalents 1
succ(O)
2 3
succ(l) succ(2)
etc. Therefore, we can restate Th 10.7 as: m + 1 = succ(m). We can also use induction to verify some familiar facts about the relationship between addition, multiplication, exponentiation and the i relation. Theorem 10.8 . For all m and all n, if 0 < m, then n < n
+ m.
Proof: By induction (on the variable m). (i) Base case. Show: if 0 < 0, then n < n means that 0 E 0, which is impossible.
+ O.
Assume 0 < O. By definition of '0' and '<', this
+ k. There are two subcases: (a) 0 < k, and (b)
(ii) Inductive case. Assume that for all n, if 0 < k, then n < n Show: for all n, if 0 < k + 1, then n < n + (k -,0 < k.
+ 1).
+
(a) Assume 0 < k. Then, for all n, n < n + k. By theorem 10.3, n k < (n + k) + 1. By theorem 10.2, < is transitive. Consequently, n < (n + k) + 1. Finally, by theorem 1O.5(b), (n + k) + 1 = n + (k + 1). Thus, by =E, n < n + (k + 1). (b) Assume -,0 < k. By Corollary 10.3 (Trichotomy), we must have either 0 = k or k < O. The latter is impossible, so 0 = k. Hence, k + 1 = 1. By Theorems 10.3 and 10.7, n < n + 1. Therefore, n < n + (k + 1). Theorem 10.9 . For all numbers m and n, if 1 < m and 0 < n, then n < n· m. Proof: left as an exercise. Theorem 10.10 . If 0
< a, then for every n, 0 < an.
Proof: By induction on n. Assume that 0 < a. (i) Base case. Show 0 < aD. By def. of exponent, aD = 1. By theorem 10.30.3,0< 1.
(ii) Inductive case. Assume 0 <
ak .
Show 0 < ak+l. [by def. of exponent] Since a> 0, for some b, a = b + 1. a k . (b + 1) = a k . b + a k [ by the distributive law]. 0 < a k , so 0 < a k . b + a k ( by theorems Th 10.2, Th. 10.8 and Th. 10.9). Therefore, 0 < ak+l.
ak+l = a k . a
Given the general laws of arithmetic we have verified, and given our scheme for translating all numerals into our language of '0' and 'succ', we can verify any computation involving numerals. If· n is a natural number ,let #n stand for the numeral which represents n, that is, #n stands for the result of applying the symbol 'succ' n times to the numeral '0'. For any natural numbers n, m and k, if n + m = k, then wecan verify the theorem which has the form 'in + #m = #k'. The same result holds for multiplication, exponents, and the factorial. For example, the formula '#2 + #2 = #4' can be proved as follows:
Logic, Sets and Functions
216 Show: succ( succ( 0)) Proof:
+ succ( succ( 0)) = succ( succ(succ( succ( 0)))))
succ(succ(O))
+ succ(succ(O))
=
succ[succ(succ(O))
+ succ(O)],
by theorem 1O.5b. =
succ(succ[succ(succ(O))
+ 0]'
by theorem 10.5b. succ( succ( succ( succ( 0) ))) ,
by theorem 10.5a. Thus, we can assert the following theorem schema:
Theorem 10.11 . For all natural numbers n, m and k: (a) if n
+m
= k, then I#n
+ #m =
#k' is a theorem.
(b) if n· m = k, then I#n· #m = #k' is a theorem. (c) if n m
=
k, then lexp(#n, #m) = #k' is a theorem.
(d) if n! = m, then I#n! = #m' is a theorem.
Exercises 1. Prove by induction that + is associative; i.e., Show VxVyVz(x, y, zEN ~ (x + y) + z = x + (y
+ z)).
2. Prove by induction that + is commutative; Le., Show VxVy(x, yEN ~ (x + y) = (y + x)) .
.~"\
3,j Prove by induction the distributive law ".'
Show VxVyVz(x, y, zEN
~
(x
+ y)
. z= x· z
+ y. z).
4. Prove by induction that· is associative; i.e., Show VxVyVz(x,y,z E N ~ (x· y). z = x· (y. z)). 5. Prove by induction that· is commutative; i.e., Show VxVy(x, yEN ~ (x· y) = (y. x)). 6. Prove the distributive law for exponentiation; Le., Show VxVyVz(x, y, zEN ~ (x. y)Z = X Z . yZ). (7-.) Show VxVyVz(x,y,z E N ~ x Y . X Z = x(Y+z)) 8. Show VxVyVz(x, y, zEN
~
(xyY
= x Y' Z)
9. Prove Theorem 10.6. 10. Prove Corollary 10.7. 11. Prove Corollary 10.8. 12. Prove Theorem 10.9. 13. Prove 1
+ 3 = 4.
*14. Prove that < is a well-ordering of N. **15. Prove Theorem 10.4.
Induction
10.2
217
Weak Induction on the Natural Numbers
As we have said, mathematical induction originally arose as a method of proof in arithmetic. As introduced by Bernoulli, induction uses the order of the natural numbers to prove theorems about them. A proof by mathematical induction, in Bernoulli's sense, consists of two parts: (1) a basis case, showing that 0 (or 1, if the theorem is to be proved only for n > 0, or, more generally, for m, if the theorem is to be proved only for n ~ m) has some property P, and (2) an inductive step, showing that, if n has P, then so does n + 1. The proof establishes the conclusion that every number has property P. The proof method, now often called weak induction, thus has the structure:
Weak Induction on N, for n
~
m
Show Vx(x ~ m --. A) Show A[m/x] Assume n ~ m Show A[n/x] --. A[n + I/x]
To take a simple example of such a proof, consider the successive sums of the natural numbers:
o o
1 1
23456 3 6 10 15 21
7
8
n
28
36
n(n+I)/2
The numbers 1, 3, 6, 10, 15, etc. Maurolycus called triangular numbers. We can state the theorem that this table suggests, then, as either For any n, the sum of the first n+ 1 natural numbers (i.e., of the first n positive integers) is n(n+l)/2 or as For any n, the nth triangular number is n(n+l}/2. In modern notation, we can also express the theorem as ~. ~t
n(n+ 1)
= --'---2---'- .
i=O
To construct a proof by mathematical induction, we can begin by showing that the theorem holds of o. Basis: let n = o. The sum of the first natural number, 0, is just 0; furthermore, 0(0+1)/2 = O. So the sum is 0(0+1)/2. Inductive step: assume for conditional proof that the theorem holds for n. We must show that the theorem holds of n + 1, i.e., that
I:
i = (n +
1)((~ +
1) + 1) .
i=O
That is, we have to show that the n + 1st triangular number is (n + l)(n + 2)/2. Now the n + 1st triangular number is just the nth triangular number plus n + 1: n+l
n
Li=Li+(n+I). i=O
i=O
But, by the assumption that the theorem holds for n (called the inductive hypothesis), we can conclude that the n + 1st triangular number is n(n + 1)/2 + (n + 1). By algebra, this is (n 2 + n +
Logic, Sets and Functions
218
2n+2)j2, i.e., (n 2 +3n+2)j2. But this is just (n+1)(n+2)j2, which is what we need to show. For another example, consider this theorem of Maurolycus: "Every integer plus the preceding integer equals the collateral odd number." In other words, the nth odd number is equal to n + (n - 1). It's easy to see that this holds of the first several odd numbers:
° °
1 1
2
3
456
7
8
n
3
5
7911
13
15
n+(n-1)
Notice that each odd number is two greater than the preceding odd number. To prove the theorem, we can use mathematical induction. Say that On is the nth odd number. Basis: we need to show that the first odd number is equal to n+(n-1). The first odd number, OIl is 1; furthermore, 1 + (1- 1) = 1. So the theorem holds of the first odd number. Inductive step: assume that the theorem holds of n. Assume, in other words, that the nth odd number is equal to n + (n - 1). We need to show that the theorem holds of n + 1, so we need to prove that the n + 1st odd number is equal to (n + 1) + «(n + 1) -1), i.e., 2n + 1. In general, On+l = On + 2. By the inductive hypothesis, On = n + (n - 1), i.e., 2n - 1. Thus On+l = 2n - 1 + 2, so On+l = 2n + 1. As another example, I will prove the so-called quotient theorem: Quotient Theorem. Let x, YEN, Y ::; x, x =I 0, y =I 0. Then there exist q, r E N such that: q =I 0, x = q. y + r, andr <= y. Proof: By induction on x (starting with 1). Base case. Show for all yEN such that y ::; 1 and y -# 0, there exist q, r E N such that: q =I 0, 1 = q . y + r, and r < y. Assume bEN, b ::; 1, and b =I 0. So, b = 1. 1 = 1 ·1 + 0, and < 1. So, 1 = 1· b + 0, and 0< b. Use 3 I to get: 3q,r EN (1 = q. b + r&r < b).
°
Inductive step. Assume for all yEN such that y ::; k and y =I 0, there exist q, r E N such that: q =I 0, k = q . y + r, and r < y. [Inductive hypothesis] Show: for all yEN such that y ::; k + 1 and y -# 0, there exist q, r E N such that: q =I 0, k + 1 = q . y + r, andr < y. Let c E Nand c ::; k + 1, and c -# 0. Either c ::; k or c > k.
Case 1. c::; k. c::; k, and c =I 0. By inductive hypothesis, there exist m and n such that m =I 0, k = m . c + n, and n < c. k + 1 = m . c + (n + 1). There are two subcases: n + 1 < c and n + 1 ~ c. Case 1a. n + 1 < c. Then m and n + 1 are the q and r we need for k + 1. Case lb. n+ 1 ~ c. Since n < c, we have that n+ 1 = c. So, k+ 1 = m·c+c = (m+ 1) ·c+O. So, let q = m + 1, and let r = 0. Case 2. c > k. Since c::; k
+ 1, c = k + 1.
So, k
+ 1 = c· 1 + 0.
Let q = 1 and r = 0.
This completes the proof.
COUNTING PRINCIPLES If a set is finite, then its members can be numbered. There will be a unique natural number which corresponds to the size of the set. Using the theory of functions developed in Chapter Three, we can give a precise definition to the notion of two' sets being equinumerous.
Definition 10.9 Sets A and B are equinumerous, A ;:::; B, iff there is a function f such that f : A -> B is a bijection.
Induction
219
A natural number n corresponds to the size (or "cardinality") of a set A iff A and n are equinumerous. Let I A I represent the size of A. Definition 10.10 If A is finite, then I A
1= n
iff n is a natural number and A
~
n.
As you might expect, no two distinct natural numbers are equinumerous. First, we need to prove the following lemma: Lemma 10.1 Ifm + 1 ~ n + 1, then m
~
n.
Proof. Suppose that m + 1 ~ n + 1. Then there is a bijection f : m + 1 --> n + 1. Since f is a bijection, and n E n + 1, there is a k E m + 1 such that f (k) = n. There are two cases to consider: k = m, and k f: m. If k = m, then k r m is a bijection from m to n, and so m ~ n. (k r m is still I-to-l, and it must be onto n, since f was onto n + 1, and the only value which k r m is missing does not belong to n.) Consequently, we can suppose that k f: m. Consider the following function 9 : 9 = (f - {( k, n) }) U {( k, f (m»)). The function 9 is the result of removing from f the ordered pair which assigns a value of n and replaces it with the value which f assigns to m. The function 9 r m is a bijection from m to n, as can be easily verified. Thus, m ~ n. Theorem 10.12 If m and n are natural numbers and m
~
n, then m = n.
Proof. By induction on n. (i) Base case. Assume m
~
O. Then m must be empty, and so m = O.
(ii) Inductive case. Assume for all m, if m
~
k, then m = k. Show, for all m, if m ~ k
+ 1, then
m=k+1.
Assume that m ~ k + 1, for m an arbitrary number. k + 1 is not empty, and so m is not empty. m f: 0, so there is a number p such that m = p + 1. Hence, p
+ 1 ~ k + 1.
By the lemma, it follows that p ~ k. By the inductive hypothesis, it follQws that p = k Hence, + 1 = k + 1. Consequently, m = k + 1.
p
On the basis of Theorem 10.12, we can show that every finite set has a unique size. Consequently, our function I A I is well- defined. Corollary 10.9 If m is a natural number, then 1 m
1= m.
The following are some of the fundamental principles for computing the sizes of complex sets. Theorem 10.13 (The Sum Rule) If A and B are disjoint, finite sets, then I AuB
Corollary 10.10 If A and B are finite sets, then I Au B
1=1
A
1+ I B I
1=1 A I + I B I - 1An B I .
Corollary 10.11 If A = {A o,'" ,An}, a finite family of finite sets, and A is pairwise disjoint, then n
IAI=LIAI. i=O
Theorem 10.14 (The Product Rule) If A and B are finite sets, then I A x B
1=1
A
I . IB I .
220
Logic, Sets and Functions
= {f := f
Theorem 10.15 (The Exponent Rule) If A and B are finite sets, and C from A into B}, then 1 C I = I B IIAI . Corollary 10.12 (The Powerset Rule) If A is a finite set, then I peA)
1= 21A1
is a function
.
Proofs. Left as exercises. Problems Prove each of the following theorems using mathematical induction. 1. The sum of the first n odd numbers is n 2 . 2. The sum of the first n even numbers (i~cludingO) is n 2 - n. 3,.'The sum of the first n even numbers (excluding 0) is n 2 + n.
'··'~.4''''
4. The sum of the squares of the first n positive integers is n(n + 1)(2n + 1)/6. 5.
.
IIf_1
I n+l
i~l =
6. The sum of the squares of the first n even numbers is 2n(n
+ 1)(2n + 1)/3.
7. The sum of the cubes of the first n positive integers is (n 2 + n)2/4. 8. The sum of the cubes of the first n even numbers is 2(n 2 + n)2. 9. The product of the first n powers of 2 is n
10.
IIi_I(I -
11.
' "n
I
(HI)'
)=~ 2(n+l)
1 _-.!!:..WHI i(i+l) - n+l
12. 2 2n
13. 22n -
2
2(n +n)/2.
.
1 is divisible by 3, for each positive inte\er n.
1
+I
is divisible by 3 for every positive integer n.
14. fen) = 3n 5 15. 3 2n +l
+ 5n 3 + 7n is divisible by 15 for
+ 2n + 2
every integer n.
is divisible by 7 for every n ;::: O.
16. Where a is an integer, a > 0, and n ;::: 0, (a 17. I:~=l i(i!) = (n
+
+ l)n
- 1 is divisible by a.
I)! - 1.
18. Suppose that f: N -.... N is such that 'Vx'Vy(J(x 'Vx(J(x) = xf(I)).
+ y)
= f(x)
+ fey)).
19. Let a > I and k be natural numbers. Define f: N -.... N: (1) f(O) f(k) + ak+l. Then 'Vx(J(x) = (ax+! -1)/(a - 1)).
Then f(O) = 0 and
=
1; (2) f(k
+ 1) =
20. Let V be an initial segment of the positive integers, from 1 to n. Then there is no strict partial ordering n on V such that Rnl and, for all i such that 1 ::; i < n, Ri(i + 1). 21. Show: for all n E N, n
+ 3 < 5n 2.
Induction
221
22. For all natural numbers x, y and n, x-y divides x n _yn without remainder. (Do the induction on n.) 23. A map is properly colored iff, whenever two regions share a common border, they are of different colors. Show that any map which is formed by n straight lines can be properely colored using exactly two colors. 24. Show: 2:~=1 (8i - 5) = 4n 2 25. Show: .
(n
+ I)! -
n. 1.
26. Show that, for all natural numbers n, n 3 /3
..-~-.;,.,
.:
2:Z:1 i· i! =
-
....
27. Show that, for all natural numbers n, n 3
+ n 5 /5
= 7n/15 is an integer.
+ 5n + 6 is evenly divisible by 3.
28. The Tower of Hanoi game consists of3 pegs and n disks. Each of the n disks is of a unique diameter. The game begins with all the disks on one peg in order of size, with the largest disk on the bottom. A move in the game consists of removing the uppermost disk from one of the pegs and placing it on top of the stack of disks on another peg. A move is legal only if a larger disk is not placed on top of a smaller one. The puzzle is solved when all of the disks are located on a peg other than the one they were on originally. Show that a Tower of Hanoi puzzle with n disks can be solved in 2n - 1 legal moves. 29. Prove that if 5 is a set with k members and x rj. 5, then 5 U {x} has k
+ 1 members.
30. Prove that if 5 is a set with k members and A is a proper subset of 5, then 31. Let m and n be natural numbers, with m < n. Show that if j : n 32. Let j be a function defined recursively as follows: j(O) j(n) = 2n - 1 , for n ~ 1.
= 1;
-+
1A 1<
k.
m, then j is not injective.
j(k + 1)
= 2::=0 j(i).
Prove that
33. Prove Theorem 10.13. 34. Prove Corollary 10.10 35. Prove Corollary 10.11. 36. Prove Theorem 10.14. 37. Prove Theorem 10.15. 38. Prove Corollary 10.12.
10.3
Strong Induction on Natural Numbers
Bernoulli's principle of induction contains, as its crucial step, the move from n to n + 1. Not all indu~tive proofs, however, use this technique. The weak induction principle is equivalent to another principle, called the principle of strong induction. A proof by strong induction also begins with a basis case, proving the result for 0, 1 or some other minimal value. But the inductive step of the proof uses a different inductive hypothesis. The strong inductive step establishes a conditional: if the theorem holds of all natural numbers less than n, it also holds of n. Rather than moving from n to n + 1, therefore, strong inductive proofs move from all m < n to n. Sometimes, for this reason, strong induction is called course-oj-values induction.
Logic, Sets and Functions
222 It's easy to see that this, too, is valid. The basis case holds, then, for all numbers less than 1; so it holds for 1. number less than 2, so it must hold for 2. And so on. number. Both weak and strong induction, then, amount exploitation.
establishes the theorem for o. The theorem Since it holds for 0 and 1, it holds for every The theorem thus holds for every natural to an infinite series of proofs by conditional
Theorem 10.16 (Strong Induction) (Restrict all quantifiers to N) [A(O)&Vx(Vy(y < x
-->
A(y))
-->
A(x))]
-->
VxA(x)
Proof. Assume A(O)&Vx[Vy(y < x
---->
A(y))
-->
A(x)]
Show VxA(x)
Show A(b) Assume -,A(b) Let c = {x EN: -,A(x)} bEe (by Abstr.) c =J; 0, and c
dE c&Vx(x
N, so, since < is well-founded on N, c has a least member d:
<:::;;
E
c
-,x < d).
---->
-,A(d) (by Abstr.).
Shew Vy(y < d
---->
A(y))
Assume e < d
Show A(e) Assume -,A(e) eEc -,e
e
-->
A(y))
-->
A(d), by assumption.
A(d) -,A(d), by Repetition. QED.
Induction
223
Strong Induction on N Show V'xA Show A[O/x] Assume V'y(y Show A[n/x]
--+
A[y/x])
This rule can be generalized to be used in proofs where the base case is some number, like 1 or 2, which is greater than 0:
Strong Induction on N, for n Show V'x(x ~ m Show A[m/x]
--+
~
m
--+
A[y/x])
A)
Assume n > m&V'y(m Show A[n/x]
~
y
Notice that the inductive hypothesis is (n > m&V'y(m ~ y < n --+ A [y/x))) , rather than the V'y(m ~ y < n --+ A[y/x]) that the above discussion suggests. This is simply to avoid duplication. If we were forced to take account of the case where n = m in the inductive step, we would have to repeat the proof of the basis case. Again, m represents a minimal value; if we want to prove the theorem for all positive integers, we should start with 1. If we want to prove a theorem for all positive integers greater than k, we should start with k + 1. To take an example of a proof by strong induction, let's prove the theorem that for every x ~ 2, there is a prime y and a number z > 0 such that x = y. z. Base case: If n
= 2,
2
= 2 . 1,
2 is a prime and 1 > O.
Inductive hypothesis: suppose for every x such that 2 z > 0 such that x = y. z.
~
x < n, there is a prime y and number
Show: there is a prime y and a number z such that n = y . z. Either n is prime or not. Case 1. If n is prime, then n = n· 1, n is prime, and 1 > O. Let n be y and 1 be z. Case 2. If n is not prime, then for some a and b greater than 1, n = a· b, and a and b are less than n (show this by induction). By ind hyp, a = m· q, for some prime m and q > O. By associativity of multiplication, n = m . (q . b). Since q > 0 and b > 1, (q . b) > 1. Let m be y and let (q. b) be z.
Logic, Sets and Functions
224 Problems
Leonardo of Pisa (c. 1170 -1250), an Italian monk who was also called Fibonacci, devised a sequence of numbers, (1,1,2,3,5,8,13,21,34,55,89,144,233,377,610, ... ), that have many interesting properties. Corresponding to these sequence is a function f: N -; N, where f(O) is the first Fibonacci number, f(l) is the second, etc. This function can be defined recursively as follows: f(O) = 1; f(l) = 1; f(k + 2) = f(k + 1) + f(k). For problems 1 and 2, prove the corresponding theorems about the Fibonacci numbers. 1. L~o f(i) = fen
+ 2) -
2. Z=~=o f(i)2 = fen) . fen
1.
+ 1).
i=O
3. Let the function b be defined recursively as follows: b(O) 2· b(k + 1) + b(k). Prove that for every n, ben) is odd.
10.4
1; b(l)
1; and b(k
+ 2)
Induction on Sets other than the Natural Numbers
Induction relies essentially on an ordering of the objects in a particular domain. The natural numbers, of course, come with this ordering built in. But any set can serve as a ground for inductive proof if it's ordered in a similar way. The ordering must allow for an answer to the question, What is the next member of the set?, or, at least, the question, What are the next members of the set? This requires a well-ordering of the set or of some partition of it. The set, in other words, must be partially well-ordered. A strict linear ordering doesn't suffice. The reals, for instance, are strictly linearly ordered by <. But there is no answer to the question, What is the next real after I?, if we interpret next in terms of the < ordering. Given any real greater than 1, we can find another between it and 1, because < is dense on the real line. The less-than relation, then, doesn't order the reals in the way that induction requires. It doesn't well-order the real line. Whether the reals can be well-ordered at all is somewhat controversial; the axioms of set theory presented in Chapter 1 don't decide the question. In any case, to performan inductive proof, we need to work with at least a partial well-ordering. That a set can be partially well-ordered isn't enough; we need an actual partial well-ordering to do the proof.
Strong Induction on Sets other than N Let R be any partial well-ordering of set B (and R
~
B x B)
Show 'Vx(x E B -+ A(x)) Assume e E B Show 'Vy(Rye -+ A(y)) -+ A(e) Assume 'Vy(Rye -; A(y)) Show A(e)
Suppose, for example, that we are dealing with formulas of sentential logic. We can order these formulas by associating, with each formula, a number- say, that number of connective occurrences
Induction
225
in the formula, or the number of binary connective occurrences, or the number of occurrences of sentence letters, etc. This divides the set of formulas into various groups. If we choose the function mapping each formula into the number of connective occurrences in it, the groups are: those with no connective occurrences; those with one connective occurrence; those with two connective occurrences; etc. Alternatively, suppose we are dealing with finite mathematical trees. We can associate, with each tree, a number specifying the number of nodes on the longest branch of the tree. This also orders the set of finite trees into classes: those whose longest branch has only one node; those whose longest branches have two nodes; etc. As these examples suggest, any function from a set into N partitions the set. It splits the class into the set of those things mapped into 0; those mapped into 1; those mapped into 2; etc. Where f is a function from a set A into N, the partition is the set of the inverse images of unit sets under f. The partition, that is, is the set of all nonempty f-l[{ n}] for n E N. We can view these inverse images as ordered as follows: Ifxlfy(:Jn:Jm(n,m EN &x = f- 1 [{n}] & y = f-l[{m}]) ~ (x < Y f--> n < m)). This is a well-ordering of the partition. The ordering on the set A itself is also easy to construct from the function: Ifxlfy(x, yEA ~ (x < y ...... f(x) < f(y))). This is not a well-ordering but a partial well-ordering. To meet the well-ordering requirement, then, we normally construct a function, which we'll call the index function, mapping each element of the set in question into the natural numbers. This function partitions the set and introduces on it the appropriate ordering.
Strong Induction on Sets (Using Index function) Let f: B
~
N
Show Ifx(x E B ~ A(x)) Assume c E B Show f(c) = 0 ~ A(c) Assume f(c) = 0 Show A(c) Assume f(c) = n, n > 0 Show Ify(f(y) < n ~ A(y)) --+ A(c) Assume Ify(f(y) < n --+ A(y)) Show A(c)
The second requirement pertains to the structure of the set forming the ground of the inductive proof. If this set lacks the right sort of structure, then even a function such as the above will do little good. Suppose, for example, that we want to prove something about all U.S. citizens. We construct a function mapping citizens into numbers by associating, let's say, each citizen with the number of characters in his or her full name. This function maps James Earl Carter into 15, Sandra Day O'Connor into 17, and Willard van Orman Quine, John Cougar Mellenkamp and Nicholas Michael Asher all into 20. Now, perhaps we want to prove that no citizen can eat just one potato chip. The basis case is easy: we need to show that no citizens with names of 0 characters can eat just one chip. Assuming that every citizen has a name with at least one character, there are no such citizens, so the basis case is trivial. The inductive step, however, is impossible, and not because of the oddity of what we want to prove. Let's say we try strong induction. Then we assume that all
226
Logic, Sets and Functions
citizens with names of less than n characters are unable to eat just one potato chip. We must show that those with names of n characters can't eat just one. To accomplish this, however, we must be able to establish some relationship between the potato-chip-eating abilities of those with names of n characters and the same abilities of those with shorter names. And, in all probability, no such relationship exists. Bluntly, the length of a person's name has nothing to do with how many potato chips that person can eat at one sitting. The ordering we introduce by mapping the members of the set into N, therefore, must correspond in some way to structure in that set relevant to what we want to prove. This is an extremely vague requirement. But we can make it somewhat more precise by considering our examples of inductive proofs on N. To prove our theorem about triangular numbers, we needed to be able to relate the nth triangular number to the n - 1st such number. In other words, we needed to be able to express a relationship between each triangular number and the one preceding it. We relied on the fact that the n + 1st triangular number is just the nth triangular number plus n + 1: n+l
n
i=O
i=O
I>= I>+(n+l)
To prove our theorem about odd numbers, we had to do the same: we needed to express the relationship between each odd number and the odd number preceding it (On+! = On + 2). In general, in order to use weak induction, we must establish a relationship between the value of some function for n and that function's value for n + 1. If we are dealing with a set other than N, we need to establish a link between objects the index function maps into n and those it maps into n + 1. The success of weak induction depends on a relationship between the objects in any equivalence class and those in the next equivalence class. The success of strong induction requires a relation between the members of any equivalence class and those in preceding equivalence classes. Consider, for instance, our proof that every positive integer is a product of primes. We relied on the fact that any composite number n can be factored into two smaller numbers m and k such that km = n. We can view positive integers not only as successors of other integers but as sums or products of other, smaller integers. The sum and product operations need not carry us from n to n + 1; typically, they carry us from k and m to an n larger than both k and m. When these operations are relevant to the proof, therefore, there is a dependence of facts about any number on facts about smaller numbers. A similar point holds for sets other than N. We can use strong induction to prove facts about members of A with an index function f just in case the relevant properties of A's members in a certain equivalence class under f depend on properties of members of preceding equivalence classes. Normally, as we have already said, inductive proof methods apply to problems involving recursively defined functions. The method of recursive definition, however, also applies to sets. We have already encountered an example of a recursive definition of a set, the recursive definition of the set of formulas. Such definitions specify what a term means by specifying how to construct instances to which it applies. Thus, we've explained formula by saying that all sentence letters are formulas, all negations of formulas are formulas, etc. Any recursive definition of a.- set F like the set of formulas consists of three parts: basic clauses, which postulate that entities of certain kinds fall under F; inductive clauses, which postulate that if entities of certain kinds fall under F, then so do entities of other kinds; and an extremal clause, postulating that those are the only objects falling under F. Recall that our definition of formula, in sentential logic, had the form: (1) Any sentence letter is a formula. (2) If A is a formula, then -.A is a formula.
Induction
227
(3) If A and B are formulas, then (A&B), (A V B), (A
~
B), and (A
<-4
B) are formulas.
(4) Every formula can be constructed from a finite number of application of (1)-(3). Here, (1) is the basic clause, (2) and (3) are recursive clauses, and (4) is the extremal clause. A recursive definition without any recursive clauses is simply an explicit definition, which could be paraphrased into a universalized biconditional. The general framework for a recursive definition of a set can be stated thus:
Recursive Definitions of Sets: To define A, a subset of a given set U: (1) Basic clause: specify minimal elements. Vx E U(¢(x) ~ x E A) (2) Inductive clauses
VxVy
E
Af(x, y)
E
A&
Vx E Ag(x) E A & (3) Extremal clause A is the intersection of all the subsets of U satisfying conditions (1) and (2).
A very important special case of recursive definition occurs whenever A is freely generated from the set {x: ¢(x)} by functions f and g. Set A is freely generated by this recursive definition just in case the set of basic elements is disjoint from the range of each of the functions gi, and each of these ranges is disjoint from all the others. Definition. A set A is freely generated from set {x : ¢(x)} by functions f and g iff
(1) A is defined recursively in terms of ¢,
f
and g, as above,
(2) The sets {x: ¢(x)}, R(f), and R(g) are pairwise disjoint, (3)
f and 9 are injective.
Let's put our definition of SL into this form. Let U be the set of n-ary sequences of symbols from the vocabulary V, Le., U = U{V n : n EN}, where v n is the set of functions from n into V. We define L as a subset of U by the clauses: (1) Vx E U(x : 1
(2)
VXI E
~
V&x(O) is a sentence letter
LNeg(xI) E L E LConj(Xl,X2) E L E LDisj(Xl,X2) E L E LCond(Xl,X2) E L E LBicond(xl,x2) E L
VX I VX2 VXIVX2 VXIVX2 VX I VX2
(3) L = n{A: A
~
U&(1)[AjL]&(2)[AjL]}
~
x
E
L)
228
Logic, Sets and Functions
We will define the function Neg, Conj, Disj, Cond and Bicond in the section 10.6, but there intended meaning should already be clear here: N eg(x) is the negation of x, C onj (x, y) is the conjunction of x and y, in that order, etc. In this definition, the language £. is freely generated from the set of sentence letters by the functions Neg, Conj, Disj, Cond, and Bicond, since the range of each of these functions is disjoint from the set of sentence letters and from the ranges of each of the other functions, and since each of these functions is one-to-one. Given any recursive definition, we can define an appropriate index function f in the following way. Let's suppose that we're defining a subset A of a domain U by recursion. First, we can define a function approx: N -> P(U) by recursion:
approx(O) approx(k + 1)
{XEU:¢>(x)} approx(k) U {z E U : 3x3y(x, y E approx(k) &(x = f(x,y) V x = g(x))}
= =
approx(O) consists of the set of minimal elements of A, i.e., elements of U which got into A by one of the basic clauses of the recursive definition. In the case of the language £. of SL, approx(O) is the set of all 1-ary sequences [¢>], where ¢> is a sentence letter (i.e., ¢> is p, q, r, .. .). approx(k + 1) consists of union of approx(k) with the set of elements which got into A by applying one of the inductive clauses of the definition to elements of approx(k). Thus, approx(k + 1) consists of all the elements of A which have been added to A by the time we've reached the k + 1st stage of the recursive process. In the case of SL, approx(k + 1) will include all sentences whose grammatical tree contains no branch longer than k + 1. Now, define the index function rnk : UR(approx) -> N as: rnk(x) = the least n E N such that x E approx(n) If x E UR(approx) , then x E approx(n) , for some n. By the fact that <, N is well-founded, it follows that there is a least such n. The connectedness of < on N guarantees that there is a unique such number, so our definition of rnk(x) is always proper. In the case of SL, rnk(A) will be the length of the longest branch of the grammatical tree of A. For rank to be an appropriate index function for A, we must know that V(rnk) = A, that is, that A = UR( approx). We can show by strong induction on N that UR( approx) <; A. To show that A ~ UR(approx), suppose that x E A. Then x belongs to every set b such that {x E U : ¢>(x)} <; b and b is closed under f and 9 (by the extremal clause of the recursive definition). UR(approx) is such a set, so x E UR(appox). Thus, A <; UR(approx). The index function rnk is appropriate to use in constructing inductive proofs about A because of the following two facts:
(1) Suppose rnk(z) = O. Then
Z
E
approx(O). So, ¢>(z).
(2) Suppose rnk(z) = k + 1. Then z E approx(k
+ 1).
So, 3x3y(x, y E approx(k)&(z = f(x, y) V z = g(x))) So, 3x3y(rnk(x) < rnk(z)&rnk(y) < rnk(z)&(z = f(x, y) V z = g(x))) For every z E A, either rnk(z) = 0 or rnk(z) = k+1 for some k. So, every z E A either satisfies a basic clause of the definition or it satisfies some inductive clause with respect to some x and y whose ranks are less than that of z. All of the minimal elements of A satisfy some basic clause, and all of the other elements of A entered it by means of applying some inductive clause to elements already in A. Thus, if we can prove that some formula A is true of everything satisfying a basic clause of the definition, and that, for every inductive clause, A is true of the output on the assumption that A is true of every input, then we have proved that A is true of every element of A.
I
Induction
229
Proof by Strong Induction (using recursive definition) Show \fx(x E B ---7 A(x)) Assume c E B [Assume rnk(c) = 0] Show ¢(c) ---7 A(c)
Assume Vy(rnk(y) < rnk(c) -+ A(y)) [Ind. Hyp.] Assume c = f(d 1 , d2) V c = g(d 1 ) Assume rnk(d 1 ) < rnk(c)&rnk(d2) < rnk(c) Therefore A(dt}&A(d2) Show A(c) [use separate cases] Case 1. Assume c = f(d 1 , d2) Case 2. Assume c = g(dd
The basic index function rnk is not the only acceptable function to use in connection with a recursive definition. Any function g: N ---7 A is all right so long as whenever g(x) < g(y), y's being in A does not depend on x's being in A. In order to define precisely which functions are all right, we must specify a relation dep which expresses the dependency relation: depxy iff y's being in A depends on x's being in A. Let dep be defined on A as follows:
Vx\fy(depxyR(rnk(x) < rnk(y)&:Jz
E
.e[y
= f(x,z) Vy = f(z,x)
Vy
= g(x)]))
Finally, we can define an acceptable index function for A, given the recursive definition above: Definition 10.11 h: A
---7
N is an acceptable index function for A if and only if:
Vx E
In the case of the language index functions:
AVy
.e of SL,
E
A(depxy
---7
h(x) < h(y))
it is easy to verify that the following would be acceptable
- number of characters in a string - number of sentence letters and negations - number of parentheses and negations - number of connective symbols The following would be unacceptable as index functions: - number of parentheses [Compare p and -,-,p]
I
230
Logic, Sets and FUnctions
- number of right parentheses - number of sentence letters Since relations are simply sets of ordered pairs, these can be defined recursively in exactly the way described above. For example, suppose we wanted to define the relation of 'x is an ancestor of y', given the relation Par = {(x,y) : xis a parent ofy}. The definition would proceed as follows: (1) Base clause. VxVy(Par xy (2) Inductive clause.
--+
(x, y)
E
Anc)
VxVyVzVw( (x, y)
E
Anc & (z, w)
E
Anc & y = z
--+
(x, w)
E
Anc)
The base clause guarantees that if x is a parent of y, then x is an ancestor of y. The inductive clause stipulates that if x is an ancestor of y, and y is an ancestor of w, then x is an ancestor of w. The presence of this clause ensures that the relation Anc is transitive. In fact, the ancestor relationis what is called the transitive closure of the parent relation. The definition of the ancestor relation does not fit perfectly the pattern for recursive definitions of sets we gave above, since the inductive clause does not mention a function. We could achieve the same result by restating the inductive clause thus: (2) Inductive clause:
Vx,y
E
Ancg*(x,y)
E
Anc
We can then define the function g* as a function from (H x H) x (H x H) into H x H (where H is the set of humans, alive and dead):
V(x, y)V(z, w) E H x H
[if y
= z,
then g( (x, y), (z, w))
= (x, w), and
ify =f- z, then g«(x,y), (z,w)) = (x,y)] Since functions are simply binary relations, we could define a function recursively in the same way. However, we would then have to verify that the binary relation we have defined is in fact a function. In some cases, we can use a special kind of recursive definition which always defines a function, eliminating the need for this additional proof. This special kind of recursive definition depends on having a set which has been freely generated (see above).
Recursive Definition of a Function
Let A be a subset of U which has been freely generated from B by j and g. Assume that V is a set, and that b : B --+ V, j: V x V --+ V, and G : V --+ V. To define j : A
--+
V :
(i) Basic clause. Vx
E
Bj(x) = b(x);
(ii) Recursive clause. VxVy E Aj(f(x, y)) = j(j(x),j(y)) Vx E Aj(g(x)) = G(j(x))
Induction
231
For example, we could, given an assignment of truth-values to the sentence letters of 1:-, define recursively a function which assigns appropriate truth-values to all of the sentences of 1:-. For simplicity's sake, let's pretend that I:- has only two connectives: negation and conjunction. Let B be the set of sentence letters, let V be the set {T, F} of truth-values, and let b be a function from B into V, assigning truth-values to all of the sentence letters of 1:-. First, we need to introduce the functions which represent the truth tables for the negation sign and the conjunction sign. The first function, N, is a function from V into V. N = {(T, F), (F, Tn. That is, for input T, N yields F, and for input F, N yields T. Thesecond function, C, is a function from V x V into V.
c
= {((T, T), T), ((T,
F), F), «F, T), F), (F, F), F)}
We can now define a function j which assigns appropriate truth- values to every sentence of 1:-, since I:- has been freely generated from B by Neg and Conj.
(i) Base clause.
'Ix E Bj(x)
= b(x).
(ii) Inductive clauses.
VxVy
E
I:-j(Conj(x, y») = C(j(x),j(y» 'Ix E I:-j(Neg(x» = N(j(x»
To sentence letters, j assigns the same value as does b. To determine the value of j for a conjunction, first apply j to each of the conjuncts and then apply the function C (which represents the truth-table for conjunction) to the result. To determine the value of j for a negation, first apply j to the negated formula and then apply the function N.
Problems 1. Consider the following recursive definition of the set S
~
N x N:
(i) (0,0) E S. (ii) if (m, n) E S, then (m + 2, n
+ 3)
E S.
(a) List four members of S. (b) Prove that if (m, n) E S, then 5 evenly divides m 2. Consider the following recursive definition of the set T
~
+ n.
N x N.
(i) (0,0) E T. (ii) If (m, n)
E
T, then(m
+ 1, n), (m + 1, n + 1)
, and(m + 1, n
(a) List six members of T. (b) Prove that 2m 2: n for all (m, n) E T. 3. Consider the following definition of the set A
~
N x N:
(i) (0,0) E A.
(ii) If (m, n) E A, then (m + 1, n) and (m, n Show that A = N x N.
+ 1)
4. Let AI, A 2 , . .. , An be a sequence of subsets of S. (a) Give a recursive definition of: Uj=IA j
.
belong to A.
+ 2)
all belong to T.
Logic, Sets and Functions
232 (b) Prove that for every n, Uj=1 A j <:;; S.
5. Assume that for all x, if x had blue eyes, then both of x's parents had blue eyes. Show that, for all x, if x had blue eyes, then all of x's ancestors had blue eyes. 6. Define the property of ''whitish'' as follows: Base. If surface x is visually indistinguishable from the Standard White Patch in the Smithsonian, then x is whitish. Recursive. If surface x is visually indistinguishable from y, and y is whitish, then x is whitish. Assume that the relation of visual indistinguishability is reflexive, symmetric and transitive. (a) Prove that for all x, if x is whitish, then x is visually indistinguishable from the Standard White Patch. (b) Is the result proved in part (a) plausible? If not, which assumption was faulty? 7. Define a mutual admiration society as follows: (B) If x =I- y, x admires y, and y admires x, then {x,y} is a MAS.
(R) If S is a MAS, x admires everyone in S, and someone in S admires x, then S U {x} is a MAS. (a) Prove that if S is a MAS, then there is someone in S who is admired by everyone else in S.
-
(b) Prove that if S is a MAS, then the relation of admiration is connected on S. 8. Let b be a truth-function defined on all sentence letters, and let j be a truth-function for SL defined recursively in terms of b (as above). Assume that 4> is a formula that contains no instances of -', -., or R. Assume that the range of b is {T}. Prove that the range of j is also {T}. 9. Assume that bi and b2 are truth-functions defined on all sentence letters, and that bl(p) = b2(p) and bdq) = b2(q). Let ji and j2 be truth functions for SL defined in terms of bi and b2, respectively. Assume that 4> is a formula of SL that contains only the sentence letters p and q. Prove that jl(4)) = j2(4)). 10. Define the set of plans recursively as follows. (B) If A is the name of a simple action, then A is a plan.
(R)(l) If 4> is a formula of SL, and Yis a plan, then "while 4>, Y" is a plan. (2) If YI and Y 2 are plans, then "YI ; Y2 " (i.e., "first do YI , then Y 2 "') is a plan. (3) If Y is a plan, then "y*" (i.e., do Y any number of times in succession, and then stop) is a plan. (a) Construct in this formal language a plan corresponding to the following recipe: "Break the egg. Heat stove. Stir egg until solid. Eat to your heart's delight." (b) Prove that any plan contains at least one simple action.
Induction
10.5
233
Graphs, Thees and Lists
In this section, we will introduce several kinds of structures which are wideiy used in mathematics and other formal sciences: graphs, trees, and lists. Structures of these kinds can be defined recursively, and theorems about them are typically proved by induction. The first type of structure is that of a graph. A graph is a kind of network of points, called 'vertices', connected by line segments, called 'edges'. Formally speaking, a graph is a triple langleA, B, J) , where A is the set of edges, B is the set of vertices, and f is a function from A into the set of pair sets of objects from B, that is, {{x, y} : X E [3&y E B}. The function f assigns two endpoints from B to each edge in A. The following diagrams represent graphs.
1 d
9
3
d
A
3
b B
The graph represented by this diagram has four vertices and seven edges (labelled a through g). There are two edges (a and b) connecting vertex 1 and vertex 2: these are called parallel edges. There are also two parallel edges between 3 and 4. There is also an edge, d, connecting vertex 4 to itself: such an edge is called a 'loop'. In diagram B, the graph has three vertices and five edges, including a loop and one pair of parallel edges. The degree of a vertex in a graph is the number of ends of edges which meet at that vertex. Thus, in the above diagram, vertex 1 has a degree of three, and vertex 4 has a degree of five (note that a loop increases the degree of a vertex by two, not one). Definition 10.12 IfG is a graph consisting of (A, B, f) , and G' is a graph consisting of (A', B', 1'), then G' is a subgraph of G if and only if A' ~ A, B' ~ B, and f i A' = 1'.
Suppose that graph G consists of the triple (A, B, I). A path of length n through G consists of two sequences: an n- ary sequence of edges from set A,langleall a2, ... , an), and an n + l-ary sequence of vertices from set B, (b I , b2, ... , bn , bn + I) , which have the following property: for each edge ai, f (ai) = {b i , bi+ d. This means that the edges of the path are connected, in order, end to end. For example, in diagram A, the pair of sequences (a, c, d, c) and (1,2,4,4,2) is a path of length four. In contrast, the pair of sequences (a, b, g, b) and (1,2,1,3,2) is not a path through the graph A, since f(b) = {I, 2}, not {3,2}. Definition 10.13 A pathp, {(al, ... ,an ) , (bI, ... ,bn+l)}, is closed iffb i =bn+ I . Definition 10.14 A path is simple iff all of its edges are distinct. Definition 10.15 A path p, {(aI, . .. , an) , (b I , ... , bn+l)} , is a cycle iff p is closed and simple and the vertices b2 through bn+l are distinct.
234
Logic, Sets and Functions
Definition 10.16 'A graph is acyclic if it contains no cycles. One of the oldest problems involving graphs is the Konigsberg bridge problem, which was solved by the Swiss mathematician Leonhard Euler. The puzzle consisted of the question of whether it was possible to take a walk in Konigsberg, crossing each bridge only once, and returning to one's starting position. The arrangement of bridges and land masses in Konigsberg corresponded to the diagram on the left.
9
0-
(
D
In the language of graphs, the Konigsberg puzzle can be stated thus: is there a closed path in the graph on the left which uses each edge exactly once? Such a path is called an Euler circuit. Euler proved that no such path exists for this graph. He proved a more general result, Theorem 10.17. Theorem 10.17 If graph G has an Euler circuit, then every vertex of G has an even degree. Proof Assume that p = {(al,"" an), (b l , ... , bn+d} is an Euler circuit in G of length n. For k < n, let Pk be the first k edges of p. Lemma 10.2 For every k such that 1 :::; k :::; n, either (1) Pk is a circuit, and every vertex in Pk has even degree in the subgraph consisting of the edges of Pk, or (2) Pk is not a circuit, and every vertex in the path except the first and the last has even degree in the subgraph consisting of the edges of Pk, and the vertices bl and bk+l have odd degree in that subgraph. [Proof of Lemma] By induction on k. (i) Base case. k = 1. If Pk is a circuit, then bl has a degree of two, which is even. If Pk is not a circuit, then the vertices bl and b2 each have a degree of 1, which is odd. There are no other vertices.
(ii) Inductive case.
Assume that either (1) Pm is a circuit, and every vertex in Pm has even degree in the subgraph, or (2) Pm is not a circuit, and every vertex in the path except the first and the last has even degree in the subgraph, and the vertices bl and bk+ I have odd degree in that subgraph.
Case (1). Pm is a circuit. If Pm+! is also a circuit, then the degree of bi is increased by two. Therefore, all the vertices have even degree. If Pm+! is not a circuit, then the degrees of bi and of bm + 2 are increased by one, making them odd. All other degrees remain even.
Induction
235
Case (2). Pm is not a circuit. If Pm+l is a circuit, then the degrees of bl and of bm are each increased by one, making them both even. If Pm+! is not a circuit, then the degree of bl is unchanged, so it remains odd. The degrees of bm+l and bm + 2 are each increased by one, making the first even and the second odd.
[Proof of the theorem] Since p = Pn , and p is a circuit, every vertex in the subgraph corresponding to p has an even degree in that subgraph. Since P contains all the edges of G, this means that every vertex in G haseven degree. We can also prove something which is close to the converse of Theorem 10.17. First, we must introduce another definition. Definition 10.17 A graph is connected iff for every vertex in the graph, there is a path in the graph from one to the other. Theorem 10.18 A finite connected graph G in which every vertex has an even degree has an Euler circuit. Lemma 10.3 If G is a finite connected graph in which every vertex has an even degree, then there is at least one cycle in G. [Proof of Lemma] Let p be one of the longest paths in G all of whose vertices are distinct. Suppose the vertex sequence of pis (b l , ... , bn ). Vertex bn has an even degree, so its degree must be at least two. Therefore, there must be another edge terminating at bn other than the one which links bn - l to bn in path p. This other edge must connect bn to some vertex in p, since if it connected bn with some distinct vertex, say bn + I, then the path corresponding to (b I , ... , bn , bn + I) would be a path longer than p, all of whose vertices are distinct, contrary to our assumption. Therefore, G contains at least one cycle. [Proof of the theorem] By strong induction on the number of edges. (i) Base case. Suppose G has only one edge. Then, it must be a loop, else bl would not have an even degree. This loop is an Euler circuit for G.
(ii) Inductive case. Suppose G has n edges, n > 1. Inductive hypothesis: Every finite connected graph with fewer than n edges in which every vertex has an even degree has an Euler circuit. By Lemma 10.3, we know that G has at least one circuit C. Let G\C be the sub-graph which is obtained by removing from G all the edges of C and all the vertices on C of degree 2. Removing the edges of C removes an even number from the degree of each vertex of G. Consequently, all the vertices in G\C have even degree in G\C. G\C consists of a finite number of connected subgraphs, HI,' .. , H k . Each of these connected subgraphs Hi has fewer than n edges, so, by the inductive hypothesis, each of them has an Euler circuit, Pl. Since G was connected, each of these subgraphs has at least one vertex on the cycle C. For each Hi, choose such a vertex hi on C. Therefore, G has an Euler circuit, which can be constructed in the following manner. Start with the initial vertex bi on C. Proceed along C until you come to the first vertex chosen as one of the hi'S. At this point, leave the cycle C and follow instead the Euler circuit Pl. Upon returning to hi, proceed again along C until the next chosen vertex is reached. Eventually, all of the edges will be traversed exactly once, and one will return again tob l . Another important type of graph is a directed graph or digraph. If the edges on an ordinary graph can be though of as two- way streets, all the edges on a digraph must be thought of as one-way streets. A digraph is a triple (A, B, f) , where A is a set of edges, B is a set of vertices, and f is a
Logic, Sets and Functions
236
function which assigns an ordered pair (not a pair set) of vertices to each edge. A digraph is usually diagrammed by representing edges as one-way arrows. There are two kinds of paths through digraphs. An undirected path through a digraph consists of a sequence (al,"" an) of edges and a sequence (b l , ... , bn + l ) of vertices such that for every i :::; n, either f(ai) = (bi,b HI ) or f(ai) = (bi+l,b i ). A directed path through a digraph consists of a sequence (al,'" ,an) of edges and a sequence (b l , ... , bn+l) of vertices such that for every i :::; n, f(ai) = (b i , bi+l) . This means that the direction of a directed path must respect the direction of its edges. Let's say that a proper digraph is one with no loops and no parallel edges. There is a very simple way of defining proper digraphs in set theory. Each proper digraph whose vertices belong to set A can be thought of as simply a binary relation on A. Conversely, every binary relation on a set A corresponds to a proper digraph whose vertices belong to A. When there is a directed edge from vertex b to vertex b' in digraph G, this corresponds to the presence of the ordered pair (b, b') in the binary relation R(G) which corresponds to G. We can in fact define a digraph as an ordered pair (A, R) , where A is a set of vertices and R is a binary relation on A, Le., R ~ A x A. The next type of structure we will consider is that of a tree. A tree is simply a connected, acyclic graph. Since trees are acyclic, they have no loops or parallel edges. Definition 10.18 A graph is a tree iff it is connected and acyclic. Definition 10.19 x is a leaf of tree T iff x is a vertex of degree one in T.
Theorem 10.19 A finite tree with at least one edge has at least two leaves. Proof. Consider a longest simple path in the tree, with vertex sequence (b l , ... , bn ) . Since the path is acyclic, bl -1= bn , and since the path is one of the longest, both blandb n must be leaves. Theorem 10.20 For every n ? 2, a tree T with n vertices has exactly n - 1 edges. Proof By weak induction. (i) Base case. T has 2 vertices. If T contained more than one edge, it would have to contain either a loop or a pair of parallel edges, which is impossible since it is a tree.
(ii) Inductive case. Assume that any tree with k vertices has k - 1 edges. Assume that T has k + 1 vertices. By Theorem 10.19, T has a leaf, boo Let To be the tree which results from removing bo and the single edge connected to bo . To is a tree with k vertices. S6, by IH, To has k - 1 edges. Therefore, T has kedges. Corresponding to directed graphs, there are also structures called directed trees. A directed tree is simply a connected digraph with no undirected paths which are cycles. If all of the arrows are pointing away from a single vertex, the tree is called a rooted tree, and that vertex is called its root. The class of rooted trees can be defined recursively.
237
Induction
Recursive Definition of Rooted Trees Let A be a set of possible vertices. (i) Basic clause. If b E B, then (I b 1,0) is a rooted tree, and b is its root. [This is a trivial tree, consisting of a single vertex and no edges.] (ii) Recursive clause. If (AI, R 1 ) , ... ,(Ak, Rk) are rooted trees with roots rI,"" rk, u{ Ai : 1 ::; i ::; k} is pairwise disjoint, rEA, and r is not an element of U{ Ai : 1 ::; i ::; k}, then T is a rooted tree and r is its root, where the set of vertices of T is {r} U U{ Ai : 1 ::; i ::; k , and the set of directed edges of T is ({r} x {rl' ... ,rd) U U{ R i : 1 ::; i ::; k}.
A single vertex is a trivial rooted tree. If one starts with a finite collection of rooted trees, none of which have any vertices in common, and one adds a new vertex which is connected by outwardly directed edges to the roots of the rooted trees, the resulting digraph will also be a rooted tree. If vertex b is connected by an outwardly directed edges to vertices bI , ... ,bn in tree T, then b is called the parent of bl , ... ,bn , and bl , ... ,bn are called the children of b. The depth of a rooted tree can also be defined recursively: Recursive Definition of the Depth of Rooted Trees Let T = (A, R) be a rooted tree, with root r. (i) Base clause. If A is a unit set (i.e., T is a trivial, one-vertex tree), then the depth of T is zero. (ii) Inductive clause. Suppose T is not a trivial tree. Then r has children in T, say rl, ... , r n . Each of these r i 's is the root of a subgraph T i , which is itself a rooted tree. Thus, each of the Ti's has a depth: d 1 , •.• , d n . The depth of Tis 1 +Max{d I , ••• , d n }. A binary tree is a rooted tree all of whose parent vertices have at most two children. A regular binary tree is a binary tree all of whose parent vertices have exactly two children. The level number of a vertex in a rooted tree is the length of the unique directed path from the root to that vertex. (The level number of the root itself is 0.) A full binary tree is a regular binary tree all of whose leaves have the same level number. The depth of a rooted tree (defined recursively above) is also the maximum level number of any vertex in the tree. Definition 10.20 (A, R, S) is an ordered rooted tree iff (A, R) is a rooted tree and S is a binary relation on A which linearly orders the children of every vertex in (A, R) . The final type of structure which we will introduce in this section is that of a list. A list is a kind of finite sequence. We have already defined n-ary sequences in Chapter 8. The n-ary sequence
238
Logic, Sets and Functions
(aI, ... ,an) is defined in terms of nested ordered pairs: (( ... ((aI, az), a3), . .. an-I), an) . For many purposes, it is convenient to define n-ary sequences in a different way. Given the notion of a natural number, it is useful to think of an n-ary sequence or list as a function from n into some set. Suppose f is a function from some natural number n into A, i.e., f : n -+ A. Remember that n is defined as the set {O, 1,2, ... , n - I}. This set is f's domain, and it has n members. The length of a list f is always V(f). The first member of the list which f represents is f(O), the second member is f(l), and so on, until we reach the nth member, which is f(n - 1). In general, we can define the n-ary list [ao,aI, ... ,an-l] as that function fsuch that f(O) = ao,f(l) = al, ... ,f(n-1) = an-I. From this definition, we can derive the following rules:
List Def
f: n -+ A f = [1(0), f(I), . .. , f(n - 1)]
[ao,al,az, ... ,cin-d: n
-+
{ao,al,az, ... ,an-I}
i E N&i < n [ao,al,az, ... ,an-l](i) = ai
It is also useful to have names for certain collections of these lists, analogous to the n-ary Cartesian products of Chapter 8.
Definitions. An
{x: x: n
-+
A}
A
=
{x: 3y(y E N&y
A
=
{x: 3y(y E N&x: y
NIL
=
o
< n&x : y ---T
---T
A}
A}
An is the collection of all n-ary lists of elements of A. A
'Ix
E
V(f)(f(+n)(n
+ x) = f(x))
Next, we need to define operations that add an element onto the end of a list , and that turn a single element into a list. The first operation shall be called 'Cons', and the second 'List'.
Induction
239
Definition. VxVy(Cons(x, y) = xU {(V(x), y)} Vx(List(x) = Cons(NIL,x))
Using the function Cons, it is possible to define the set A
E
A
(ii) Inductive clause. Vx E A
f" NIL = f
(ii) Recursive clause: VfVgVxf"Cons(g,x) = Cons(f"g,x) Alternative (non-recursive) Definition. VfVg[f"g = f U g(+'D(f))] Exercise: prove that the two definitions are equivalent in extension. We need to verify that concatenation is associative, and that concatenating a list with the empty list leaves the list unchanged:
Proof. (left as exercise) Theorem 10.22 Vx(xANIL = NILAx = x) Proof. (left as exercise) Another useful operation is that of the reversal of a list, which we will symbolize by placing a backwards arrow, +- , before the name of a list. The reversal operation can also be defined recursively. Recursive Definition of List Reversal: (i) Basic clause.
+-
NIL
= NIL.
(ii) Inductive clause. Vx E A < NVy E A+- (Cons(x, y)) = List(y)A( +- x) Theorem 10.23 For all x, +- (List(x))
= List(x).
Theorem 10.24 For all lists x and y, +- (xl\y) = (+- y)A(+- x). Theorem 10.25 For all lists x, +- (+- (x)) = x. Theorem 10.26 For all lists x, V(+- x) = V(x).
240
Logic, Sets and Functions
Proofs left as exercises. One class of problems which is quite common in computer science applications is that of extracting lists of vertices from ordered rooted trees. There are essentially threepossible ordering schemes: preorder, postorder, and inorder. In preorder, a parent-vertex is listed first, followed by the lists for its children. In postorder, the lists for the children are listed first and the parent-vertex follows them. Inorder applies only to binary trees: in a binary tree, every vertex has at most two children. In the case of inorder, the list for the first child (conventionally called the left sub-tree) is listed first, followed immediately by the parent, which is finally followed by the list for the second child (the right sub-tree), if there is one. Suppose T is an ordered rooted tree with root r, and in which r's children are, in order, rl, ... , r n . Call the sub-trees corresponding to these children T 1, ... , Tn . Recursive Definitions of Preorder:
(i) Basic clause. If T is a trivial tree, then Preorder(T) = List(r).
(ii) Inductive clause. If T is not a trivial tree"then Preorder(T)
= List(r)/\ Preorder(T1)f\ .. .f\ Preorder(Tn ).
Recursive Definition of Postorder:
(i) Basic clause. If T is a trivial tree, then Postorder(T) = List(r). (ii) Inductive clause. If T is not a trivial tree, then Postorder(T) = Postorder(Tdf\ ... f\ Postorder (Tn)f\ List(r). Recursive Definition of Inorder:
(i) Basic clause. If T is a trivial tree, then Inorder(T) = List(r). (ii) Inductive clause. If T is not a trivial tree, then Inorder(T) = Inorder(T1)f\List(r)/\ Inorder(T2 ). Problems
1. Prove that a binary tree of depth n has a maximum of 2n + 1
-
1 vertices.
2. Prove that if a binary tree has a depth of n, then it has at most 2n leaves. 3. Create a recursive definition for tree (not rooted tree) and use strong induction to prove Theorem 10.19. 4. Let S = {a, b}. Let B be the subset of S N defined recursively as follows:
(B) [a] and [b] are in B. (R) If w is in B, then [ajA[b]f\w E B, and [b]A[a]/\w E B. (a) List six members of B. (b) Prove that if wEB, then length(w) is odd. 5. Prove Theorem 10.21. 6. Prove Theorem 10.22. 7. Prove Theorem 10.23.
Induction
241
8. Prove Theorem 10.24. 9. Prove Theorem 10.25.
10. Prove Theorem 10.26. 11. Construct the grammatical tree for the following wffs:
(i) (p
~
(q&r))
(ii) ((p&q) V (r (iii) ((p
~
p)
V
~
q))
p.
Label each vertex with the symbol (sentence letter or connective) that it contributes to the formula. Then, list the symbols of the wff (ignoring parentheses), using (a) Prearder, (b) Postorder (c) I norder. Which of the resulting lists is ambiguous? I
12. Show that a connected graph with n vertices has at least n - 1 edges. 13. Show that the sum of the degrees of the vertices of a tree with n vertices is 2n - 2. 14. Consider a full binary tree T of depth d. (a) How many leaves does it have? (b) How many vertices does it have? 15. Prove that the depth of a rooted tree is equal to the maximum level number of any vertex of the tree. "16 . Define an algorithm which will, given the list of the symbols of a wff of the language of SL, construct the grammatical tree for the wff.
10.6
Formal Languages
Formal languages are defined by recursion. Proving theorems about such languages almost always involves resorting to strong induction. Once again, we'll turn to the language of sentential logic as an example. First of all, I would like to define the various syntactic functions: negation, disjunction, etc. Let V be the basic vocabulary of the language; V ={'P', 'Q', 'R', ... , '.', '&', ... }. Now we can define the syntactic functions of SL:
"ix
E
V< N Neg(x)
List('. ')"x
"ix,y E V< NConj(x,y)
List('( ')"x" List('& ')"y" List(') ')
"ix, y E V< N Disj(x, y)
List('( ')"x" List('
V
"ix, y E V< N Cond(x, y)
List('( ')"x"List('
~
"ix, y E V< N Bicond(x, y)
')"y" List(') ')
')"y" List(') ')
List('( ')"x" List(' ~ ')"y" List(') ')
Let's demonstrate a theorem regarding formulas of sentential logic. Recall that a sentence is syntactically ambiguous iff it can be generated by way of two distinct phrase structure trees. Chapter Two asserted that formulas of sentential logic are immune from syntactic ambiguity: every formula has a unique phrase structure tree. Let's prove that this is the case, using strong induction.
242
Logic, Sets and Functions
First, the set of formulas within sentential logic, is defined recursively. The set of formulas thus constitutes an acceptable ground for inductive proof methods. Many index functions introduce orderings on the set of formulas that conform to the inductive definition. We can order sentential logic's formulas by counting their connective occurrences or by using the depth of the rooted grammatical tree; these would result in acceptable index functions. In order to prove that every formula has a unique grammatical tree, we must first show that every formula has a unique main connective. First, we will define what it is for something to be a main connective of a formula, and then claim that every formula has a unique main connective. In order to do this, we need to introduce two new functions, If : V< N --> Nand rt : V< N --> N. The function l gives the number of left parentheses in a finite string of symbols, and the function r gives the number of right parentheses. Each can be defined recursively, using the Cons function. For example, If can be defined thus:
If(NIL) =0
(bc)
\Ix E V< Nlf(Cons(x, 'C)) = If(x)
(ic)
\Ix
E
V< N\ly
E
V[y
-I- 'C
Theorem 10.27 . \lx\ly E V< Nlf(x/\y) = If(x)
-->
+1
If(Cons(x,y) = If(x)]
+ If(y)
Theorem 10.28 . If(List('(') = 1. These same results can also be established for the function rt : V< N --> N. Before proceeding with the proof of the main theorem, we need to establish several lemmas (preliminary results). Lemma 10.4 : If a E
.c,
then If(a) = rt(a)
Proof: by strong induction. Let f : .c --> N be an acceptable index function for might give the number of sentential connectives in a formula of .c. (i) Basis case: f(a)
= O. So, a = List(b) whereb is a sentence letter. If(a)
.c,
.e.g.,
f
< f(a), If(b)
=
= 0 = rt(a).
(ii)Inductive step: f(a) > O. Inductive hypothesis: for every b such that f(b) rt(b). There are two cases:
= List('-' ')/\b, where b E .c. If(a) = If(b) = rt(b) = rt(a). : a = B(c, e) = List('( ')/\c/\List(b)/\e/\ List(') '), where b is a binary connective ('& " 'V', '-->', or ' ...... ') and c,e E .c. By inductive hypothesis, If(c) = rt(c) and If(e) = rt(e). If(a) = If(c) + If(e) + 1, and rt(a) = rt(c) + rt(e) + 1, so If(a) = rt(a).
Case 1 . a = Neg(b) Case 2
Definition 10.21 If a E Lemma 10.5 : If a E
.c
.c, a =
b/\c, and V(c)
-I- 0, then b is a proper initial segment of a.
and b is a proper initial segment of a, then If(b) 2: rt(b).
Proof: by strong induction on a. Basis case: f(a) = 0, so a proper initial segment.
= List(c), and c is a sentence letter. Since V(a) = 1, a cannot have a
Inductive step: f(a) > 0, and for every b E .c such that f(b) < f(a), and every c which is a proper initial segment of b, If(c) 2: rt(c). There are two cases.
Induction
243
= Neg(c) = List('.. ')I"C,C E L. Since b is a proper initial segment of a,b = List(' .. ')d, where d is a proper initial segment of c. By inductive hypothesis, If(d) 2: r(td). Since If(b) = If(d) and rt(b) = rt(d), If(b) 2: rt(b).
Case 1 : a
Case 2 : a = List('( 'Y'd' List(dY'e ll List(' '), where d is a binary connective and c, eEL. There are four sub-cases: Case 2a . b = List('( ')lI g, and 9 is a proper initial segment of c. By inductive hypothesis, If(g) 2: rt(g). Since If(b) = If(g) + 1, and rt(b) = rt(g), If(b) 2: rt(b). Case 2b . b = List('( ')II C or b = List('( ')IICII List(d). By Lemma 10.4, If(c) = rt(c). Since If(b) = If(c) + 1, and rt(b) = rt(c), If(b) 2: rt(b). Case 2c . b = List('( ')IICII List(d)lIg, and 9 is a proper initial segment of e. By inductive hypothesis, If(g) 2: rt(g). By Lemma 10.4, If(c) = rt(c). Since If(b) = If(c) + If(g) + 1, and rt(b) = rt(c) + rt(g), If(b) 2: rt(b). Case 2d . b = List('( ')IICII List(d)lI e. By Lemma 10.4, If(c) = rt(c) and If(e) = rt(e). Since
If(b) = If(c)
+ If(e) + l,andrt(b) = rt(c) + rt(e),lf(b) 2: rt(b).
Lemma 10.6 : If dEL and a is a proper initial segment of d, then a rf. L. Proof of lemma - by strong induction. Basis case. f(d) = o. So, d = List(x) and x is a sentence letter. Since d is a I-place list, d cannot have a proper initial segment. This contradicts the assumption that a is such a proper initial segment of d. Inductive step: f(d) > 0, and for every b such that f(b) initial segment of b, c rf. L. There are two cases.
< f(d),
and every c which is a proper
Case 1 : d = List(' .. ')II X and x E L. Since a is a proper initial segment of d, a = List(' .. ')lI y , where y is a proper initial segment of x. By inductive hypothesis, y rf. L. The only way List(' .. ')lI y could belong to 12 would be for y to belong to 12 (this can be proved by a simple strong induction). Since y rf. 12 and a = List(' .. 'y" y , a rf. L. Case 2 . d = List('( ')/lXIl List(y)1I Zll List(') '), where y is a binary connective, and x and z belong to L. There are four subcases: Case 2a . a = List('( ')lIh and h is a proper initial segment of x. By Lemma 10.1, If(h) 2: rt(h). Since If(a) = If(h) + 1 and rt(a) = rt(h), If(a) > rt(a). So, by Lemma 10.1, a rf. L. Case 2b . a = List('( ')/l x or a = List('( ')/lXIl List(y). By Lemma 10.1, If(x) = rt(x). Since If(a) = If(x) + 1, and rt(a) = rt(x), If(a) > rt(a) By Lemma 10.1, a rf. L. Case 2c . a = List('( ')/lXIl List(y)llh, where h is a proper initial segment of z. By Lemma 10.2, If(h) 2: rt(h). By Lemma 10.1, If(x) = rt(x). Since If(a) = If(x) + If(h) + 1, and rt(a) = rt(x) + rt(h), If(a) > rt(a). By Lemma 10.1, a rf. L. Case 2d . a = List('( ')IIXII List(y)II Z . By Lemma 10.1, If(x) = rt(x) and If(z) = rt(z). Since If(a) = If(x) + l(z) + 1, and rt(a) = rt(x) + rt(z), If(a) > rt(a). So, by Lemma 10.1, a rf. L. Now we can define what it is to be a "main connective" of a formula. "Main connective" is to be a binary relation on 12 x N defined as follows: Definition 10.22 m is a main connective of f iff
(1) m = 0, and
244
Logic, Sets and Functions (i) (ii)
1 = [xl and x is a sentence letter, or 1 = ['...., 'l"x and x E .c.
(2) m = n + 1, and 1 = ['( ']"x"[y]"z"[') x and z belong to L, and D(x) = n.
'J, y is a binary connective (either '&, 'V', '-+' or 'i--;'),
Theorem 10.29 : every wff 1 in L has a unique main connective. Proof. First, we can prove by induction that every sentence in £, has at least one main connective. We proceed by cases. If 1 ELand 1 = List(x), where x is a sentence letter, then 0 is a main connective of 1. If 1 = Neg(g)andg E L, then again 0 is a main connective of 1. If 1 = Conj(g, h) or Disj(g, h) or Cond(g, h) or Bicon(g, h), and 9 and h belong to L, then a main connective of 1 is D(g) + 1. Now, we must show that no sentence in L can have two distinct main connectives. Suppose for contradiction that 1 ELand that m and n are main connective off and m '" n. There are two cases to consider: either one of m or n is equal to zero, or neither are equal to zero. Case 1 : suppose one is equal to z. Without loss of generality, we can suppose that m = 0, n '" O. If o is a main connective of 1, then 1(0) is either a sentence letter or '...., '. But if n is a main connective of f and n '" 0, then f(O) must be a left-parenthesis, 'C. But f(O) cannot be both a left parenthesis and either a negation or a sentence letter. So, case 1 is disposed of. Case 2
suppose neither m nor n is equal to O. So,
f
= List('( ')"a" List(b)"c" List(') ')
where b is a binary connective, a and c belong to L, and D(a) = m. Also, f = List('( ')"d"List(e)"g"List(')) " where e is a binary connective, d and 9 belong to L, and D(d) = n. Since m '" n, a :f d. By the connectedness of j on N, either m < n or n < m. Let's suppose that m < n. This means that d is longer than a, and both agree with f for the numbers from 0 to m - 1, so d = a"h, for some list h such that D(h) :f O. In other words, a is a proper initial segment of d. Since dEL, by lemma 10.3, a rt L, contradicting our assumption. Theorem 10.30 : every wff in L has a unique phrase structure tree. Proof. Basis case : C is a sentence letter. And it's obvious that sentence letters have a unique phrase structure tree; trees for sentence letters have only one node. Inductive step: We assume that every sub-formula of C has a unique tree. Case 1 : C has a negation as main connective. Then 0 is C's unique main connective. Therefore, for some unique formula D, C = ....,D. By inductive hypothesis, D has a unique phrase structure tree. But it follows that C has a unique phrase structure tree, namely,
c
o
Induction
245
Case 2 : C has a conjunction as main connective. Then, for some unique formulas D and E, C = (D&E). D and E have unique phrase structure trees, by the inductive hypothesis. Thus C has the unique tree
c
/\
o
E
Case 3 : C has a disjunction as main connective. Then, for some unique formulas D and E, C = (D V E). D and E have unique phrase structure trees, by the inductive hypothesis. Thus C has the unique tree
c
/\
o
E
Case 4 : C has a conditional as main connective. Then, for some unique formulas D and E, C = (D --> E). D and E have unique phrase structure trees; thus C has the unique tree
c
/\
o
E
Case 5 : C has a biconditional as main connective. Then, for some formulas D and E, C = (D ....... E). Again D and E have unique phrase structure trees by the inductive hypothesis, and, hence, C has the unique tree
c
/\
o
E
Logic, Sets and Functions
246
In every case, then, C has a unique phrase structure tree. This completes the inductive step, and the entire proof. In order to put the recursive definition of QL from Chapter 4 into the form required by this chapter, we need first to define the substitution function. We want the substitution function Sub to be a function from V< N x V x V, that is, the set of ordered triples whose first constituent is a finite string of symbols and whose second and third constituents are symbols, into V< N. Sub(A, v, c) shall be A(v j c], that is, the result of substituting every occurrence of 'c' in A with an occurrence of 'v'. Since Sub is a 3-ary function, it will be a subset of a 4-ary relation. We can define Sub recursively: For all y,z
E
VSub(NIL,z,y) = NIL.
For all x E V < N and all y, z VSub(Cons(x, y), z, y) = Cons(Sub(x, z, y), z). For all x E V < N and all y, z, w E V[y Sub(Cons(x, y), z, w) = Cons(Sub(x, z, w), y)J.
-I
w
E
---->
We can then verify the following facts: V'y,z E VSub(List(y),z,y) = List(z) V'y, z, wE V[y
-I-
z
---->
Sub(List(y) , w, z)
V'x,y E V < NV'z,w E VSub(xl\y,z,w)
= List(y)] = Sub(x, z,w)I\Sub(y, z,w)
Now we can define QL as follows: Base clauses: 1. If x is a sentence letter, then List(x) E QL.
2. If x is an n-ary predicate, and y is an n-ary list of constants, then List(x)l\y E QL. Inductive clauses: 3. If x E QL, then Neg(x) E QL. 4. If x, Y E QL, then Conj(x, y), Disj(x, y), Cond(x, y), Bicond(x, y) E QL. 5. If x
E QL, y is a constant, z is a variable, xCi) = y for some i, and xCi) i E Vex), then List('V' ')I\List(z)I\Sub(x,y,z) E QL.
-I-
z for every
Extremal clause: 6. QL
= n{x E V<
N : (1) - (5)(xj QLJ)
Problems Working within the confines of sentential logic, prove the following theorems. Let if(A) be the number of left parenthesis occurrences in the formula Aj rt(A) be the number of right parenthesis occurrences in A; peA) the number of parenthesis occurrences in A; b(A) the number of binary connective occurrences in A; and seA) the number of sentence letter occurrences in A.
Induction 1. For any formula A, If(A)
247
= rt(A).
2. For any formula A, p(A) = 2b(A). 3. For any formula A, s(A) = b(A)
+ 1.
4. For any formula A, s(A) = 1 + rt(A). 5. For any formula A, p(A) = 2s(A) - 2 6. For any formula A, any interpretation of A (Le., assignment of truth values to the sentence letters in A) assigns at least one truth value to A. 7. For any formula A, any interpretation of A (Le., assignment of truth values to the sentence letters in A) assigns at most one truth value to A. Say that two pairs of parentheses separate each other iff they occur in the order (k(m)k)m, and that a pairing of n left parentheses with n right parentheses is proper iff a left parenthesis is always paired with a right parenthesis to the right of it, and no two pairs separate each other. Prove these theorems (from Kleene 1974). 8. If 2n parentheses are properly paired, and one pair is removed, then remaining 2n - 2 parentheses are properly paired. 9. A proper pairing of 2n parentheses, where n > 0, contains an innermost pair, Le., a pair which includes no other parentheses between them. 10. A set of 2n parentheses admits at most one proper pairing. 11. If a set A of 2n parentheses and a consecutive subset B ~ A of 2m of them both admit proper pairings, the the proper pairing in B forms a part of the proper pairing of A. (That is, each parenthesis of the pairing has the same mate in both subsets.) Consider the language consisting of the two vocabulary items 'a ' and 'b'. The term string is defined inductively: (1) ['a '] is a string. (2) If A is a string, then ['(']"A"'['b']"'[')'] is a string. (3) There are no other strings. Prove these theorems about an arbitrary string S of this language. 12. S contains exactly one a .
13. The number of parentheses in S is twice the number of b s in S. 14. In S, all left parentheses are to the left of a, and all right parentheses are to the right of a. Consider the language consisting of vocabulary items '0 ' and '1' and defining strings inductively as follows. (1) 1'0 '] is a string. (2) 1'1 '] is a string. (3) If A is a string, then ['0 ']'" A and A"'I'I'] are strings. (4) There are no other strings. Prove these theorems about this language. 15. In any string T, all 0 s appear to the left of all 1 s. 16. Some strings are syntactically ambiguous. 17. If a string T contains both 0 sand 1 s, it is syntactically ambiguous. 18. If a string contains only 0 s or only 1 s, it has a unique phrase structure tree.
Logic, Sets and Functions
248
Consider the '''droid-and-doctor'' language consisting of vocabulary items' . , and' It defines strings inductively as follows: (1) ['
~
is a string, then [' ']" A" [' • theoremsabout this language.
21. A string contains no
(2) ['
~ '] is a string.
(3) If A
'] is a string. (4) There are no other strings. Prove these
19. Any string contains at most one 20. A string contains at least one
1Ir '] is a string.
II '.
II
m
if it contains an even number of
~ s.
II s if it contains an odd number of ~ s.
22. Every string has a unique phrase structure tree. Consider the "burger-and-brew" language consisting of vocabulary items' defines strings inductively as follows: (1) ['
~
~
'] is a string. (2) ['
~
~
also begins with a
~
24. A string contains at least one
~
~
'. It
'] is a string. (3) If A is a
string, then A" [' '] is a string. (4) If A is a string, then [' ']"A" [' There are no other strings. Prove these theorems about this language. 23. Any string ending in a
~ , and ' ~
~
'] is a string. (5)
.
if it contains an even number of
~
s.
25. Every string has a unique phrase structure tree. Consider the "Star-Wars" language consisting of vocabulary items ' strings inductively as follows: (1) ['
~
'] is a string. (2) ['
~ , and'
(t}) '] is a string. (3)
~
then [' (t}) ']" A is a string. (4) If A is a string, then [' ']" A" [' are no other strings. Prove these theorems about this language. 26 In any string, the number of
(t)) '. It defines If A is a string,
(t}) '] is a string. (5) There
~ s is less then or equal to the number of
(t)) s plus one.
27. Every string has a unique phrase structure tree. Consider the "bat-and-ball" language consisting of vocabulary items' .-...... , and' defines strings inductively as follows: (1) ['
Q9
.-......
~ '] is a string.
Q9
contains no .-......
It
Q9 '] is a string. (3) If A is a .-...... '] is a string. (5) is a string,A" [' (2) ['
string, then A" [' ']" [' '] is a string. (4) If A There are no other strings. Prove these theorems about this language. 28. No string but
@ '.
Induction
29. Every string but
249
@
ends with a
~.
30. In any string, the number of ........ s is greater than or equal to the number of one.
@ s minus
31. Every string has a unique phrase structure tree. Consider the "cat-and-mouse" language consisting of vocabulary items '
~
" '
~
,
~ '. It defines strings inductively as follows: (1) [' ~ '] is a string. (2) If A is a string, ~ ']/\ A is a string. (3) If A is a string, then [' ~ 't A/\ [' ~ '] is a string. (4) There
and '
['
are no other strings. Prove these theorems about this language. 32. Any string contains at least one 33. In any string, the number of
~.
~
is greater than the number of
~ s.
34. Every string has a unique phrase structure tree. Let QL be the language of quantification theory without identity. Prove the following theorems about an arbitrary formula A of QL. 35. Every quantifier in A binds some variable in A. 36. If A is a formula, then A[c/d] is a formula. 37. If VxA is a formula, then A[c/x] is a formula. 38. No quantifiers on the same variable overlap in scope. 39. Give an example of wffs a and (3 in the language of SL and lists of symbols 8 and 'Y such that Conj(a,(3) = Conj(-y,8), but a -:j:. "f. 40. A formula of SL is in conjunctive normal form iff it is of the form (C 1 &C2 & ... &Cn ), where each Ci is of the form (L 1 V L 2 V ... L k ), for some k, and each L j is either a sentence letter or the negation of a sentence letter (these are called 'literals'). Prove by strong induction that every formula of SL is logically equivalent to some formulain conjunctive normal form. **41 . A formula ¢ is in Horn-clause form iff (a) ¢ is in conjunctive normal form, and (b) and each conjunction C i in ¢ contains at most one positive (unnegated) literal. Construct an efficient algorithm for deciding whether ¢ is satisfiable.
Appendix A
Plato's Users' Guide A.I
Introduction
Plato is a software application for both Windows and Macintosh computers developed at the University of Texas at Austin. It is intended to assist students who are learning formal logic. It can be used as a resource for any introductory course in formal logic that uses a similar approach. Plato is a flexible, extensible tool that enables the user to construct proofs in the formal languages of the Sentential Calculus and the Predicate Calculus. It can also be used to construct some formal mathematical proofs, including portions of set theory. Plato is not limited to a fixed stock of examples, and it can construct proofs of great complexity. It is so rapid and accurate that it can assist even advanced students in the construction of complex proofs, yet is so easy to use that it is accessible to the beginner. Plato helps the user to construct proofs in the system developed in chapters 3, 5 and 6. A proof in Plato is a sequence of formulae in our formal language, either the language of sentential logic (8L) or the language of quantificational logic (Q). A proof begins with one or more assumptions and a Show line indicating what is to be proved. The rest of the proof is constructed using deductive rules that apply to the existing proof and that add new lines at the end of the proof. Using Plato, one can construct proofs of several different types; each type of proof can also be used to construct subproofs within a larger proof. For SL, Plato can be used to construct three types of proof: direct proof, indirect proof, and conditional proof. For QL Plato can construct proofs of these three types, and it can also construct universal proofs. In constructing a proof using Plato, one uses deductive rules that are stored in rule files. Three rule files are provided with Plato: Basic Rules, Derived Rules, and Quantificational Rules. These rules are used to add new formulas to a proof within the framework provided by a proof type. The Basic Rules file includes the following deductive rules: Conjunction Exploitation, Conjunction Introduction, Disjunction Exploitation, Disjunction Introduction, Double Negation, Conditional Exploitation (sometimes called Modus Ponens), Biconditional Introduction and Biconditional Exploitation. The Quantification Rules file includes Existential Introduction, Existential Exploitation, and Universal Exploitation. In addition, the Derived Rules file includes a number of useful rules that can be derived using the rules in the other files. They include rules that correspond to DeMorgan's laws, Modus Tollens, and other rules that are useful in constructing a proof. Plato allows the user to use any rules from any open rule files in constructing a proof. Plato also allows the user to open any number of rule files. 251
Logic, Sets and Functions
252
A.2
Getting Started
In this section, we will describe how to get Plato running on the Apple Macintosh computer (Mac Plus and higher using System 6.0.5 or higher). Those running Platoin a Windows should use the installer application included on the accompanying floppy disks. Macintosh users simply have to copy the Plato application and the Rule files on to their hard disk, while dragging the Logic fonts folder into their computer's System folder. To use Plato, you should be familiar with the basic operation of your operating system, whether Windows or MacOS, including clicking and dragging with the mouse, and using menus. For use on the Mac, the Plato application requires the following files and folders: 1. The Plato application itself (the first icon illustrated below).
2. A suitcase containing three fonts: Terlingua, Salt Flat and Pittsburgh. 3. Some rule set files (the third icon below).
n[] I:"' I
The fonts in the suitcase should be installed on the Macintosh that you are using. If you are using System 6, you can use the FontjDA Mover to add the fonts to the System file, or you can use some other font utility to install these fonts. If you are using System 7 or later, you may simply drag the fonts onto the System Folder or the Fonts folder within the System Folder.
A.3
Starting Plato on the Macintosh
Find the icon for Plato: a picture of a hand holding a pencil to a deductive proof (illustrated above). Double-click on the icon, or click on the icon and select Open from the file menu. After a moment, the window for Plato will appear on your screen. When it starts, Plato will open an empty document. You can also double-click on a Plato document (a document having the middle icon in Figure 1.) to start Plato with that document, or you can drag a Plato document or rule file to the Plato application to open that document or rule file.
A.3.1
The Menus: An Overview
At the top of the screen, you will see the usual Apple, File, and Edit menus as well as six new menus: Proofs, Annotations, Font, Size, and Windows. The menu bar is illustrated below. You can pull down each menu to see what it contains. The various commands in each menu are described in the following sections. In this section, those menus and commands that are not directly related to the construction of a proof are described.
,. •
File
Edit
Proofs
Rnnotations
Font
Size
Windows
Plato Users' Guide
A.3.2
253
The File Menu
Most of the commands in the File menu are standard Macintosh File menu commands. • New - Creates a new document. (CMD-N). • Open... - Opens an existing document. (CMD-O). • Open Rule Set... - This command allows you to open a file containing a set of derivation rules. It presents a standard file dialog from which you can choose the file containing the rule set to be opened. The commands associated with the derivation rules in a rule set are added to the Proofs menu, as a submenu (a menu with a triangle to the right of the menu item). The submenu will be named after the rule set, and the menu items in that sub menu will be named after the rules that they are associated with. • Close - Closes the topmost document. If you have changed the contents of the document, you will be asked if you want to save the document before closing it. • Close Rule Set... - Presents a list of open rule sets and allows you to choose one to be closed. Once you close a rule set, the submenu associated with that rule set is removed from the Proofs menu. • Save - Saves the topmost document to a file. You may not overwrite any files that were not created by the Plato program. (CMD-S). • Save As... - Saves the topmost document to a file, prompting for a new file name. • Save As MacDraw... - Saves the topmost document as a PICT document, suitable for use with most drawing programs (e.g. MacDraw, SuperPaint, etc.). • Revert to Saved... - Reverts the topmost document to the most recently saved version, discarding all changes since the most recent Save or Save As operation. You will be asked if you really want to revert the document. • Quit - Quits the application. You will be asked if you wish to save any changes to open documents. (CMD-Q).
A.3.3
The Edit Menu
Most of the Edit menu commands are standard Macintosh Edit menu commands. However, the Undo and Redo commands implement infinite undo and redo. • Undo - Undoes the most recent change made to the topmost document. Once a change is undone, subsequent use of Undo undoes the changes made previous to the most recent change, in reverse chronological order. Repeated use of Undo steps the document back to progressively earlier states. (CMD-Z). • Redo - Redoes the most recently undone change to the topmost document. Once a change is redone, subsequent use of Redo redoes the previous most recent undone changes, in reverse chronological order. Redo can only be used after a command has been undone; if a new command is executed, the Redo command is not available. (CMD-R). • Copy - Copies the selected formula, or, if no formula is selected, copies the entire table as a picture that can be pasted into another application. The command-key equivalent (CMD-C) works in all dialogs where text is entered.
Logic, Sets and Functions
254
• Cut - Not enabled except for use with a desk accessory. However, the command-key equivalent (CMD-X) works in all dialogs where text is entered. • Paste - Not enabled except for use with a desk accessory. However, the command-key equivalent (CMD-V)works in all dialogs where text is entered. • Show Clipboard - Shows the contents of the clipboard (if they are either text or a picture) in a window.
A.3.4
The Annotations Menu
Plato documents may be annotated with a title and a comment. The title appears at the top of the document, above the proof. The comment appears at the bottom of the document, below the proof. Four menu items in the Annotations menu allow you to change the title and comment of a document.
• Edit Title...- This command allows you to enter a new title for the document or edit an existing title. Within the dialog box, the standard Macintosh editing command keys work, and one may type a carriage return by typing option-return. Typing a return closes the dialog box, saving the title (this is the same as clicking the "OK" button with the mouse). Typing command-period (CMD-.) closes the dialog box without saving the title (this is the same as clicking the "Cancel" button). One may also type option-return to insert a return into the text of a title.
Enter or edit the comment on the table below:
( Cancel )
K
OK
B
• Edit Comment ... - This command works just as the Edit Title command does, except that it allows you to edit the comment of a document rather than the title. The title of the document appears over the proof, and the comment appears below the proof. • Remove Title - This command removes the title from a document. • Remove Comment - This command removes the comment from a document.
Plato Users' Guide
A.3.5
255
The Font and Size Menus
The Font menu changes the appearance of your Plato documents; it controls the font in which a document is displayed. Only those fonts that have the proper logical characters are included in the Plato font menu. [Windows users do not have this option - all Plato files produced on the Windows platform employ the Windows version of the Terlingua font, Terlingu. ttf.] The Size menu changes the appearance of your Plato documents; it controls the size of the font in which a document is displayed. Choosing a number in the menu displays the topmost document in the font and the size chosen. The other menu items allow you to change the font size in other ways. The current size is indicated by a check mark next to a number in the menu, or by a check mark next to the Other menu item and a number in parentheses after that menu item if the font size does not appear in top part of the menu. • Larger - This menu command changes the font size of the font used to display a document to a point size one greater than the current font size. (CMD-}). • Smaller - This menu command changes the font size of the font used to display a document to a point size one smaller than the current font size. (CMD-[). • Other... - This menu command presents a dialog box (illustrated below) in which you can type a new font size from 1 to 1024. The document will be displayed in the font size that you type if you click OK or type a return. The font size will remain unchanged if you click Cancel, hit the escape key, or type CMD-.(CMD-period).
Please enter the font size: ( Cancel )
A.4
~~(iiiiiiOK~J
Homework Mode the Windows Version
Since the Windows version of Plato was developed most recently, it has several features that are unavailable on the Macintosh version. The most important of these added capabilities is the Homework Mode. When Homework Mode has been enabled, Plato will request the user's name and student ID number. These will be encrypted and stamped into every file produced while the user is in Homework Mode. When you are in Homework Mode, you may open existing Plato files, but only when those files include only assumptions and an initial Show line. If a file includes a proof already in progress, the file cannot be opened while you are in Homework Mode. This prevents you from simply copying someone else's homework and stamping your own identification into their files. To produce a Plato file for homework credit, you must start a proof from the very beginning while in Homework Mode. There is also a Grading Mode, which your instructor may use in checking and grading homework produced by Plato. When submitting homework to your instructor electronically, you should begin the name of every Plato file with your last initial and the last four digits of your
256
Logic, Sets and Functions
student ID. Otherwise, if the instructor receives several copies of a file named "hw1.pla", the latest file received will simply overwrite all of the earlier submissions. Unfortunately, Homework Mode is available only on Plato jar Windows. If you are using a Macintosh, you may use Plato for purposes of self-instruction, but you will not be able to produce submittable homework files. The following table provides Windows users with keystroke combinations that may be used in entering logical symbols in the Terlingua font:
Symbol V
f-->
\I
3
A.5
Terlingua Windows Keystroke Table Keystroke Combination Narne of Symbol Alt + 0253 Disjunction Alt + 0225 Conditional Alt + 0151 Negation Alt + 0226 Biconditional Alt + 0140 Universal quantifier Alt + 0180 Existential quantifier
The Windows Menu
• Cascade Windows - This command arranges the open windows within Plato so that you can easily see the title of the first few windows. It also reshapes the windows to fit the screen nicely. • The remaining items in the Windows menu allow you to bring a window to the top. This menu lists the titles of all open windows in the Plato application. The title of the topmost window is listed in gray text, and the titles of the other windows are listed in regular text in the order that they are stacked on the desktop. To bring a window to the top, you can select its title from the Windows menu. You can also click on a window to bring it to the top, but if the window is not visible, then the Windows menu is sometimes convenient.
A.6
Languages of Logic
Plato allows you to construct proofs using two formal languages, the language of sentential logic and the language of quantificational logic. The language of sentential logic is a sublanguage of the language of quantificational logic. We describe both briefly below. For more information and examples, see Chapters 2 and 5.
A.6.1
The Language of Sentential Logic
To add assumptions and Show lines to a proof in Plato, you must enter grammatical formulas using the Add Assumption or Add Show Line commands (see Section 4.3 below). A formula consists of a string of logical symbols that obeys the grammar of a logical language. Different logical languages can have different grammars. A formula that is grammatical, or that satisfies the conditions of the grammar for a particular logical language is sometimes called a well-formed-formula, or a WFF. For that part of logic known as sentential logic (SL), the permissible formulas within Plato are governed by the rules laid out in Chapter 2. Remember that parentheses are added whenever one of the four "binary" connectives, & , V, _, and lejtrightarrow, are used to combine two sentences into a compound sentence. This is the only time such parentheses can be added. Since the meaning of a formula can change if the
Plato Users' Guide
257
parentheses are wrong, and since Plato cannot guess what you really meant to say, you must be careful to follow the rules literally and precisely. When using Plato, you may not arbitrarily add or omit parentheses or other symbols. Plato will not recognize what we have called conventional abbreviations of formulas. Thus, the following strings of characters are not well-formed formulae: (-,A)
-,(A)
(A) A&B
((A) V (B)) (A&B) --> A
In contrast, the following are well-formed sentences of sentential logic:
((A&B) -,-,-,-,A
V
(B
-->
A))
-,-,(-,A +-+ -,(B V -,A)) (A V (A&(A +-+ A)))
There should always be exactly as many left parentheses in a sentence as there are right parentheses, and there should be exactly one pair of parentheses for each binary connective. Instead of numerical subscripts, as were used in Chapter 2, Plato allows us to add any number of prime symbols (') to a sentence letter. Consequently, Plato will not recognize PI as a sentence letter, but it will recognize P', P", P"', and so on.
A.6.2
The Language of Quantificational Logic
Predicate logic with identity adds three new logical symbols to our language: 'V , 3 , and =. The symbols 'V and 3 are quantifiers; they are the universal and existential quantifiers respectively. In addition, Q uses lower-case letters; the letters a, .. , s are called constants, and the letters t, u, v, w, x, y, and z are called variables. Unlike the language defined in Chapter 4, Plato does not allow us to add numerical subscripts to constants or variables. Instead, any lower-case letter may be followed by any number of prime signs ('), and the result will also be a constant or variable.
A.7
Systems of Proof
As we mentioned in the introduction, Plato allows the user to use several proof formats, more than a dozen basic rules, and many more derived rules in constructing proofs. In this section we briefly describe the proof system used in Plato. For more information you should consult Chapters 3 and 5.
A.7.1
Proof Formats
In the our system, one always states what one knows before the proof has begun in the form of assumptions. One also states what one wishes to prove before one begins, by means of a Show line. Then, one applies various rules of deduction to any available assumptions to derive intermediate results (if any) and eventually the desired result. Each new result is entered on a new line together with a line number and a notation explaining how the line was derived. When the desired result is obtained, the Show line that contains it is canceled, and the lines below that line are bracketed, indicating that they are no longer accessible. Assumptions are not required in every proof, though a Show line is always required. Assumptions are often not used in conditional, indirect and universal proofs. A sample proof is illustrated below, and the parts of a proof are discussed in more detail in the following sections.
258
Logic, Sets and Functions
1. (A-B) 2. (B-C) 3. (C-D)
~: [~ 6. 7. 8.
A A A (A-D) ACP
-E,I,5 -E,2,6 -E,3,7
B C D
Direct Proof A direct proof is the basic form of proof. It includes the elements discussed above. A finished direct proof has the following form.
1.
~
n. ~
A CtJ
~ [~ Notice that the Show line at the top of the proof has been canceled once the formula we set out to prove has been derived. Note also that the lines following the Show line have been bracketed. To cancel a Show line, the formula on that line (¢ on lines nand m in the example above) must not be enclosed in another bracket, and there must be no uncanceled Show lines between the Show line to be canceled and the line on which it has been derived.
Indirect Proof An indirect proof begins with a Show line of the form ¢ which is immediately followed by an assumption line containing -'¢, the opposite, or negation, of what one wishes to show. This assumption is labeled AlP, which stands for Assumption for Indirect Proof An indirect proof then includes the derivation of a contradiction from that assumption. A contradiction is a pair of sentences in which one is the denial of the other. Thus, it is impossible for both sentences in a contradiction to be true at the same time. Since the denial of what one wishes to show (the assumption) allows us to derive a contradiction, it cannot be true (i.e. it leads to contradiction, and contradictions cannot be true) and thus it must be false. If the assumption is false, then what one wishes toshow must be true. Proofs of this style are sometimes also known as Reductio ad Absurdum proofs since a contradiction is thought to be impossible or absurd. A finished indirect proof has the following form.
Plato Users' Guide
259
n.~ n;.O 1. 1.
m.
-,cp
cp
AlP
~ -,~
Conditional Proof A conditional proof begins with a Show line, on which a conditional of the form (¢> -> 'ljJ) occurs; it also has as an assumption ¢> on the next line, and the consequent 'ljJ is derived from that assumption. The assumption is labeled ACP for assumption for conditional proof. Below is a picture of what a completed conditional proof looks like.
m. The proof formats direct proof, indirect proof and conditional proof, comprise all the formats used for proofs within sentential logic. These proof formats may be used together; one may, for instance, make use of the conditional or indirect proof formats within a direct proof, and similarly a conditional proof may exploit the indirect or conditional proof formats as subproofs. Universal Proof The universal proof format resembles the conditional and indirect proof formats. It begins with a Show line that has the form of a universally quantified statement, Vv¢>. Immediately following the Show line and enclosed with the proof brackets is another Show line of the form ¢>[ajv], which represents a formula just like the formula represented by ¢>, except that every occurrence of the variable v in ¢ is replaced by an occurrence of the constant a. An important restriction on the constant a is that it cannot occur in any previous line of the proof in which the universal proof format is being employed. Here is what a universal proof format looks like.
A.7.2
Derivation Rules
Derivation Rules for sentential logic and quantificationallogic provide a means within the Deduction system for transforming assumptions into desired conclusions within the various proof formats just
Logic, Sets and Functions
260
described. For sentential logic, the basic rules allow us to introduce assumptions at the beginning of a proof (the Assumption rule) and provide ways of building up more complex truth functional formulas from their constituents that occur on earlier lines in the proof and of breaking down complex formulas into their simpler constituents. For each of the connectives, & V , -, , -., and f-t, the basic rules together with the conditional and indirect proof formats provide "introduction" rules and "elimination" or "exploitation" rules. These rules all use previous lines in the proof that are accessible. A line n is accessible at line m just in case it is (i) not an uncanceled Show line and (ii) does not occur within a subproof that does not contain line m.
A.7.3
The Proofs Menu
The Proofs Menu allows you to construct the basic structure of a logical derivation, or proof. This menu contains commands that add new information to a proof, that specify what the proof will show , and that help to construct proofs using various proof strategies. We will first look at the commands that are used in constructing proofs in sentential logic.
Basic Commands
• Add Assumption... - This command allows you to add an assumption to a proof. It implements the Assumption rule of the natural deduction system. The Add Assumption command presents a dialog box, illustrated below, which allows you to enter a well-formed formula. Within this dialog box, CMD-X, CMD-C, and CMD-V can be used to Cut, Copy, and Paste text from the clipboard. Clicking on the small buttons below the box in which the formula is edited inserts the logical character on that button into the formula at the insertion point. When you click "OK", the formula is added to the proof as an assumption with the annotation A. Typing an illegal character will cause the system to beep at you. New assumptions may not be added after anything has been derived within a proof (using the other commands).
261
Plato Users' Guide
Pl ease enter a new formula. Use the keyboard to type sentence 1et ters or varlables.
I
II
800888880 808GJ08000 ... 1 r I ~ore Symbols:.1
(. Concel .J ~
OK
JI
Note that there is also a pop-up menu in the dialog box; it is titled More Symbols. This menu allows you to type other symbols into a formula. These additional symbols are not needed for sentential or quantificational logic. They are required for set theory and other advanced topics. • Add Show Line... - This command allows you to add a well-formed formula to a proof as an intermediate goal. The dialog presented is the same as that presented by the Add Assumption command. This command can be used before beginning a proof using any of the methods of proof described earlier. • Cancel Show Line - This command allows you to cancel a Show line once the formula in that line has been derived. To use this command, select the Show line to be canceled by clicking on it or using the arrow keys, then shift-click on the formula which justifies canceling the Show line (that is the line where you have derived the formula on the Show line). Then select Cancel Show Line from the Rules menu. If the formula on the Show line has been derived without any additional assumptions, then the Show line will be canceled, and the word "Show" will be struck out. This canceled Show line is now accessible for further use in the proof. The lines of the proof below the Show line used to prove it, however, will be bracketed and will not be accessible for further use in the proof. Proof Strategies Commands The following commands are within the Proof Strategies submenu within the Proofs Menu. • Start Conditional Proof - This command allows you to begin a subproof for a conditional formula. To use this command, select a Show line containing a conditional, and then select the command from the menu. A new assumption will be added to the proof. The assumption will contain the antecedent of the formula on the selected line, and it will be annotated ACP for Assumption for Conditional Proof. When the consequent of the formula on the selected
262
Logic, Sets and Functions line has been derived, the Finish Conditional Proof command can be used to cancel the Show line containing the conditional.
• Finish Conditional Proof - This command allows you to finish a subproof for a conditional formula. To use this command, select the Show line to be canceled, and then shift click on the line which contains the consequent of the conditional. Then select the command from the menu. If the Show line is followed by an assumption for conditional proof (ACP) and the consequent of the Show line has been derived without any additional assumptions, then the Show line will be canceled, and the lines below it will be bracketed.
• Start Indirect Proof - This command allows you to begin an indirect proof. To use this command, select the Show line you want to use an indirect proof on, and choose Start Indirect Proof from the menu. A new assumption will be added to the proof; it will be annotated AlP for Assumption for Indirect Proof. If the formula on the selected Show line is a negation, then the new assumption will be the formula on the Show line with the negation removed. If the formula on the selected Show line is not a negation, the new assumption will be the formula on the Show line with a negation added. In either case, the subproof must be completed using the Finish Indirect Proof command.
• Finish Indirect Proof - This command allows you to finish a proof conducted in the indirect style. A proof in the indirect style is complete when a contradiction has been derived. This occurs when lines containing a formula and its negation have been derived. This is a contradiction since it amounts to the claim that something is both true and false, which is impossible. To use this command, select the Show line to be canceled, and shift-click on both of the contradictory lines before selecting the command from the menu. If the Show line is followed by an AlP line and the two secondary selections are indeed contradictories, then the Show line will be canceled and the lines below it will be bracketed.
• Start Universal Proof... - This command allows you to begin a proof of a universally quantified statement using the universal proof format described in section 3.2.4. To use this command, select the Show line that contains the universal statement, and choose the command from the menu. You will be presented with a dialog asking you to specify the variable and the constant to be used in the universal proof. You must specify one variable and one constant in this dialog. The variable must be the outermost quantified variable in the universal statement on the selected Show line, and the constant must not appear anywhere else in the proof. Once the proper items are specified and the OK button is clicked, a new line will be added to the proof. This line will be a new Show line with a formula constructed from the selected Show line by removing the universal quantifier and its associated variable, and then replacing all occurrences of the variable with the constant specified in the dialog. The dialog is illustrated below.
Plato Users' Guide
263
Please enter a variable and a constant to which the variable shoul d be instantiated. Replace the individual with the individual ( Caneel )
~~(iiiiiiiOK~»
• Finish Universal Proof - This command allows you to finish a proof of a universally quantified statement. To use this command, select the Show line to be canceled (i.e. the one containing the universal statement), and selected the command from the menu. If the selected Show line is followed by a canceled Show line originally constructed using the Start U niversal Proof command, then the selected Show line will be canceled and all the subsequent lines in the proof will be bracketed off.
Other Proof Commands
• Replacement - This command turns the replacement feature of Plato on and off. The replacement feature is on if this menu item has a check mark, and it is off if there is no check mark by this menu item. The use of those derivation rules that are equivalences between two formulas is affected by the state of this feature. The descriptions below describe how the derivations work when the replacement feature is off. When the replacement feature is on, their behavior is changed in the following way. The user is presented with a dialog box in which one must use the mouse to select the formula to which the rule is to be applied. If nothing is selected, it is assumed that the rule is to be applied to the whole formula in the way that would have occurred had the replacement feature been turned off. Once a formula is selected, the behavior of the rules is the same, except that if the user selects a subformula of the formula on the selected line, then the result of the rule's application is substituted for the selected subformula in the formula on the selected line when the new line is constructed.
Logic, Sets and Functions
264
Please select a subformula to which the rule should be applied.
I
( cancel)
€
OK
»
• Relations - Like the Replacement command, this command is a flag that you may turn on or off to change the way that Plato operates. When this menu command is checked, Plato allows you to use an extended syntax that is useful in set theory and formal mathematics. This syntax allows capital letters to represent individuals as well as predicates. • Hide Subproof - When the Show line of a completed subproof is selected, the Hide Subproof command will be enabled. Selecting this command will temporarily hide the subproof, replacing it with a small icon. If the Show line of a hidden subproof is selected, the command will change to Show Subproof, and selecting it will remove the icon and display the subproof.
A.7.4
Imported Rules
The only logical operations included in the Plato application are those described in Section 4.3 on the Proofs menu. Those operations are concerned primarily with entering new information into a proof and controlling the structure of your proof. Any actual deduction rules are provided by files that must be loaded into Plato, called rule files. These files may include any number of derivation rules. When they are loaded into Plato, the derivation rules that they include are added to a submenu of the Proofs menu. If the rule set contained in a rule file is closed using the Close Rule Set menu item, then the submenu containing the rules from that file is removed from the Proofs menu. Three different rule files are provided with the Plato application in the attached floppy disks. The Basic Rules file includes those rules that are concerned with the Sentential Calculus. These rules are generally justified by arguments outside the formal deductive system. The Derived Rules file includes those rules that are derived within the formal deductive system. Some of these rules are equivalence rules, to which the state of the Replacement menu item applies. The Quantification Rules file includes those rules that are associated with the manipulation of quantifiers and individuals in the Predicate Calculus. The Basic and Derived rules are discussed in detail in Chapter 3 above. The use of these rules in Plato is relatively straightforward. One uses the mouse to click on the first line to which the rule is to be applied. Then one uses shift-click to select the second or third lines (if any). Finally, one applies the appropriate Basic or Derived rule from its location in the corresponding sub-menu. If the rule does not apply to the formulas chosen, or if the wrong number of lines is selected, an error
Plato Users' Guide
265
message will inform you of this fact. If everything is in order, Plato will produce a new line on your proof, automatically applying the selected rule to the selected lines.
A.7.5
The Quantification Rule Set
• Existential Exploitation - This command allows you to eliminate the existential quantifier on a formula. Upon selecting an existentially quantified formula and choosing this command, you will be prompted for a variable and a constant. The variable must be the first quantified variable in the selected formula, and the constant must be new to the proof. If these conditions are met, a new line is created containing the selected formula without the initial quantification and with all instances of the variable replaced by the constant. Lines created using this command are labeled by 3E.
Please enter a variable and a constant to which the variable should be instant1ated. Replace the individual with the individual
_ _I _ _I
( Cancel )
€~(iiiiii;i;OK~»
• Existential Introduction - This command allows you to introduce an existential quantifier. To use it, select a formula that contains a constant and choose the command from the menu. You will be prompted for a constant and a variable. The constant must appear in the selected formula, and the variable must not appear in the selected formula. A new line will be created, containing the selected formula with all occurrences of the constant replaced by the variable, and with an existential quantifier and the variable inserted in the front. Lines created in this way are annotated 31. • Universal Exploitation - This command allows you to eliminate a universal quantifier. To use this command, select a universally quantified formula, and choose the command from the menu. You will be prompted for a variable and a constant. The variable must be the first quantified variable in the selected formula. There is no restriction on the constant. Once the variable and the constant are entered, a new line will be added to the proof, and it will contain the selected formula with the universal quantification over the specified variable removed, and with all instances of the variable replaced by the specified constant. Lines created with this command are annotated \IE. • Identity Exploitation - To substitute one constant for another given that the two are equal, use this command. Select the statement to be changed, and shift-click on the identity statement. Then, choose this command from the menu. You will be prompted for the constant
266
Logic, Sets and Functions to be replaced and the constant to replace it with. A new line will be created containing the selected formula with all occurrences of the first constant replaced by occurrences of the second constant. • Identity Introduction - To introduce a formula stating that a constant is equal to itself, choose this menu command. You will be presented with a dialog asking you for an individual. Once you have entered an individual (you may use no variables - only function symbols and constants), a new line will be created containing an identity expression with the individual you entered on both sides of the identity symbol.
Please enter on lndlvldual:
I
I ( cancel) [
OK
»
• Variable Rewrite - This command allows you to replace a variable in a logical formula with another variable. To use this command, select a quantified formula, and choose the command from the menu. You will be prompted for two variables. When both variables are accepted, a new line will be created containing the selected formula with all occurrences of the first variable replaced by occurrences of the second variable. The second variable that you specify may not appear anywhere in the selected formula before the replacement is performed. This derivation produces the annotation V R.
A.8
Macintosh Keyboard Shortcuts
Plato includes a number of keyboard shortcuts for various menu commands. Those that are specific to Plato and are not shared with other Macintosh applications are listed in the table below. These shortcuts consist of using the command key (which has either an apple or cloverleaf symbol, or both, on it) and one other key from the keyboard. Rules Menu Command Add Assumption . Add Show Line . Size Menu Command Larger Smaller
Shortcut Command-j Command-' Shortcut Command-] Command-[
Within the various dialog boxes used by Plato, a number of keyboard shortcuts are used. These shortcuts are common across all the dialog boxes used in Plato.
Plato Users' Guide Dialog Box Command Click "OK" ... Click "Cancel"
267
Shortcut Enter or Return Command-. or Escape
If a dialog box includes an area where text can be typed, the keyboard shortcuts for the Copy, Cut, and Paste commends can also be used (CMD-C, CMD-X, and CMD-V).
A.8.l
Mackintosh Keyboard Equivalents for Logical Symbols
In Plato, one can use the buttons provided in the formula entry dialogs box to enter any logical symbols that are needed. One can also use the following keyboard sequences to type the logical symbols directly when using the font Terlingua. Except for the existential quantifier the same key combinations work in Pittsburgh and Saltflat. Symbol Key Combination
..., v
V :J E
Option-shift-hyphen Option-v Option-e, a Option-i, a Option-Shift-q Option-Shift-e (Option-Shift-r) Option-e, e
Appendix B
Answers to Selected Problems Ch. 7.1 Extensionality # 2. No set is a proper subset of itself. 1. Skew -,3x(Sx&x C x) 2. 3x(Sx&x c x) AlP 3. Sa&a C a 3E,2 s;:; E, 3 4. as;:; a&a i- a =1 5. a=a &E,4 6. a i- a # 6. 1. Skew -,3x3y(Sx&Sy&x C y&y ex) 2. 3x3y(Sx&Sy&x c y&y c x) AlP 3. Sa&Sb&a c b&b c a 3E,2 4. a s;:; b&a i- b c Def, 3 5. b s;:; a&b i- a c Def, 3 6. a= b Th 7.2, 4, 5 7. a i- b &E,4 Ch 7.3 Pair sets, unit sets and enumeration
#
2. 1. Vx(Ux ....... Vy(Sy ~ x 2. Shew -,3xUx 3. 3xUx 4. Ua 5. Vy(Sy-->atf.y) 6. S{a} 7. atf.{a} 8. a E {a}
tf. y))
A
AlP 3E VE, 1,~, 4 Th 7.6 VE, 5,~ E, 6 Th 7.7
# 10. 1. Shew VxVyx E {x, y} 2. ShElw a E {a, b} 3. a= a
a =aVa=b 5. a E {a,b} Ch. 7.4 The null set 4.
#
=1 VI,3 Pair, 4
2.
269
270
Logic, Sets and Functions
1. £hew -,3x(0 ~ x&x 2. 3x(0 ~ x&x f. 0) 3. 0 ~ a&a i= 0
f. 0)
4.
aC 0
5.
Sa&Vy(y E a -> y E 0) Sa&a f. 0 3yy E a bE a bE a -> b E 0 bE 0 b\f.0
6.
7. 8. 9.
10. 11.
AlP 3E,2 ~ Def, 3
c,4 &E, 5, &E, 3, &1 Cor 7.1, 6 3E,7 VE,5 -> E, 8, 9
o
# 4. 1. £hew -,3x(Sx&0 = {x}) 2. 3x(Sx&0= {x}) 3. Sa&0 = {a} 4.
a E {a}
AlP 3E,2 Th 7.7 = E, 3, 4
aE 0 a \f. 0 Ch 7.5 Binary unions, intersections and complements 5. 6.
o
# 12. 1. £hew a - b C a 2. £hew Vx(x E a - b -> x E a) 3. £hew c E a - b -> c E a 4. cE a- b 5. c E a&c \f. b 6. cE a 7. a - bc a
AlP
&E,3 Def -, 4 &E,5 c,2,Th7.17,A
# 14. Show an b = b n a. Both are sets by Th 7.15, so we need Vx(x E an b ...... x E b n a). (-> ). Assume c E anb. Then c E a&c E b, by n . By commutativity of & , c E b&c E a, so c E bna, by n. (~ ) Similar. So, an b = b n a, by Extensionality. #16. Assume a C C and b C c. To show that a - b = an (c - b), since both are sets, we can use Extensionality. (-> ) d E a-b. So, d E a&d \f. b, by Def. -. Since a C c, dEc. Since d E c&d \f. b, dEC - b, by Def. -. Since d E a&d E c - b, dE an (c - b), by n . (~ ) dE an (c - b). By nand Def. -, we get: dE a&d E c&d \f. b. So, dE a&d \f. b, hence d E a-b. #18. Suppose a C c&b C c. We will show ace - b <-> a and b are disjoint by biconditional proof. (-» a C c - b. Since an b is a set, we need only to show that it is empty to establish that it is 0, by Th 11. Suppose for contradiction that d E an b. Then d E a&d E b, by n . Since aCe - b, dEc - b. So, d E c&d \f. b, by Def. -. But then, dEb and d \f. b, a contradiction. So, a nb is empty, and a and b are disjoint, by Djt E. (~ ) Suppose a and b are disjoint, Le.,a n b = 0 . To show that a C c - b, we need to show that an arbitrary member of a is a member of c - b. Assume d E a. Since a C c, dEc. Suppose for a contradiction that dEb. Then d E a n b, by n . Since an b = 0, d E 0, by = E. But d ¢ 0, a contradiction. So d ¢ b. Hence d E c&d ¢ b, so dEC - b, by Def. -. #20. Since both a - 0 and a are sets, we can use Extensionality. (-> ). Suppose b E a - 0. Then bE a&b ¢ 0, by Def. -. So, bE a. (--- ) Suppose b Ea. By 0, b ¢ 0.
Answers
271
So, b E a&b f/. 0. By Def. -, b E a - 0. Thus, by Extensionality, a - 0 = a. #22. Suppose that a C b. To show that an c C b n c, since both are sets by Th. 1.5.3, it is enough to show that if d E a n c, then dEb n c. Suppose d E an c. Then d E a&d E c. Since a c b, dEb. So d E b&d E c. Hence, dEb n c, by n . Thus, by c , an c C b n c. #24. We need to show en (a - b) = (a n c) - b. Since both are sets, by Th. 1.5.3 and Th. 1.5.5, we can use Ext. (-> ) Suppose dEC n (a - b). By nand Def. -, d E c&d E a&d f/. b. By Comm. and Assoc. of & , dE a&d E c&d (j. b. By nand Def. -, d E (a n c) - b. (f- ) Similar. Hence, by extensionality, en (a - b) = (a n c) - b. #26. We show (a - c) n (b - c) = (a n b) - c, by Extensionality. (-> ). Assume d E (a - e) n (b - e). By nand Def. -, we have (d E a&d f/. c)&(d E b&d f/. e). By SL, it follows that (d E a&d E b)&d (j. e. By nand Def. -, we have that dE (a n b) - e. (f- ). Similar. Since both are sets, by Th 7.153 and 7.17, (a - c) n (b - c) = (a n b) - c, by Extensionality. #28. Suppose a = {c} and b = {d}. First, we will show that if a and b are not disjoint, then a = b. Assume an b =1= 0. So, by Cor. 1.4.1 and Th 1.5.3, there is some e such that e E an b. So, e E a&e E b, by n . Since a = {e}, e = e. Likewise, e = d. By =E, e = d. Therefore, {e} = {d}, also by = E. Finally, it follows that a = b. By SL, it follows that (a and b are disjoint) Va = b.
#30. Suppose a C c&b C c. Show bee - a ~ aCe - b. (-> ). Assume that bee - a. Assume that d E a. Since aCe, dEc. We need to show that d f/. b. Assume for contradiction that dEb. Then, since bee - a, dEC - a. By Def. -, d E c&d f/. a. But, d E a, a contradiction. Thus, d E c&d f/. b. So, dEC - b, by Def. -. Since a and c - b are both sets, we have proved that aCe - b, by subset proof. (f- ) Similar. #32. Show that a C b ~ a n b = a. (-> ). Assume that a C b. We can show that an b = aby Extensionality. (->). Assume that c E a n b. Then c E a&c E b, by n . So, c E a. (f- ). Assume that e E a. Since a C b, c E b. So, c E a&c E b. By n , c E an b. (f- ) Assume that an b = a. We show that a C b by a subset proof, since both are sets. Assume that e E a. By =E, c E an b. So, c E a&c E b, by n. Thus, c E b. We have shown that a C b. #34. Show aU b = aU (b - a). (-» Assume c E aU b. Either c E a or c f/. a. Case 1. c Ea. So, c E a VeE (b - a). Therefore, c E aU (b - a). Case 2. c f/. a. By Union, c E a or c E b. So, c E b&e f/. a. Therefore, c E (b - a), and thus c E aU (b - a). (f- ) Assume c E a U (b - a). Either c E a or c E (b - a). Case 1. e E a. So, c E aU b. Case 2. eE(b-a). So,cEb,andcEaUb.
7.6 Unions and intersections of single sets #2. Since both n{ a} and a are sets, we may use Extensionality. (-> ) Assume b E n{a}. Then, by nl, Vx(x E {a} - t b E x). a E {a}, by Th. 7.7. By VE, bE a. (f- ) Assume bE a. We need to show that Vx(x E {a} -> bE x). Assume c E {a}. By Unit, c = a. Since b E a, bEe, by = E. Thus, Vx(x E {a} - t b E x). By n 1 , bE n{a}.
Logic, Sets and Functions
272
4. Suppose a E band V'x(x E b --l Sx). Suppose that an e = 0. To show that nb n c = 0, it is sufficient, by Th 1.4.3 and Th 1.5.3, to show that nb n c has no members. Assume for contradiction that d E nbn c. Then d E nb&d E c, by n. By n1, V'x(x E b ---4 dE x). Since a E b, by V'E, dE a. Thus d E a&d E e, so d E an c, by n . Since an c = 0, dE 0. But d 0, a contradiction. Thus, nb n c has no members and is therefore identical to 0.
rt
#6. Suppose that a E b. Assume that c E nb. By n 1 , V'x(x E b ---4
C
E x). So, by V'E, c E a. Since
nb and a are both sets, nb C a, by C . #8. Suppose a C b. Assume that c E nb. By n 1 , V'x(x E b ---4 C E x). We need to show that V'x(x E a -. e E x). Suppose dE a. Since a C b, dEb. Since V'x(x E b ---4 e E x), by V'Ee Ed. Thus, V'x(x E a -. e E x). So, by n 1 , e E na. Since both nb and na are sets, nb c na, by c .
#10. Suppose V'x(x E a -. x n b = 0). Show that Ua n b = 0. Suppose for contradiction that e E Ua n b. Then, by n , e E Ua&c E b. Since c E Ua, there is a d such that c E d&d E a (by U1 ). By V'E, since d E a, d n b = 0. Since e E d&e E b, c E d n b, by n . By = E, e E 0. But e rt 0, a contradiction. Therefore, Ua n b has no members. Since it is a set, by Th 7.15, it is identical to the empty set. So, Ua and b are disjoint. #12. Since u0is a set, it is sufficient, by Th 1.4.3, to show that it has no members. Assume for a contradiction that a E u0. Then, by U1 , there is a b such that a E b&b E 0. But b 0, a contradiction. Thus, U0is empty, and so u0 = 0.
rt
#14. Assume that a =I- 0 and V'x(x E a --l x is transitive). To show that na is transitive, we must show that V'xV'y(x E y&y E na --l x E na). Assume that b E c&c E na. We need to show V'x(x E a --l b E x). Let d E a. Since e E na, V'x(x E a ---4 c E x), by n 1 . Since dE a, c E d. By assumption and V'E, d is transitive, i.e., V'xV'y(x E y&y E d -. x E d). Since b E c and c Ed, bEd, by V'E. Since d was arbitrary, we have shown that V'x(x E a ---4 b EX). By n 1 , b E na. Since band c were arbitrary, we have that V'xV'y(x E y&y E na --l x E na). Therefore, na is transitive. #20. Assume that V'x(x E a --l x is transitive). Show that Ua is transitive. We need to show that V'xV'y(x E y&y E Ua ---4 x E Ua). Let band c be arbitrary. Assume that bE c&c E Ua. By U1 , there is a d such that c E d&d E a. Since d E a, d is transitive. Thus, V'xV'y(x E y&y E d --l X E d). Since b E c&c E d, bEd, by V'E. Since b E d&d E a, b E Ua, by u 1 . Since band c were arbitrary, we have shown that V'xV'y(x E y&y E Ua --l x E Ua). So Ua is transitive. #22. Assume that Sa&V'x(x E a --l Sx). Assume that Ua C na, and that a =I- 0. Since a is a nonempty set, there is a b such that bE a, by Cor. 7.1. We will show that a = {b}, by extensionality. (-. ) Suppose c Ea. We will show that c = b, by extensionality. Since V'x(x E a --l Sx), both band c are sets. (-.) Assume dEc. Then, d E c&c E a, so d E Ua, by U1 . Since Ua C na, dEna. So, V'x(x E a -. d E x). Since b E a, dE b. (<- ) By a similar argument. Thus, c = b. bE {b}, by Th. 7.7, so c E {b}, by = E. (- ) Suppose c E {b}. Then c = b, by Unit. Since b E a, c E a, by = E. Therefore, a = {b}. So 3ya = {y}.
Ch. 7.7 Powersets #2. Show: 0 is a member of every set. Assume a is a set. 0 C a, so 0 E P(a). #4. Assume P(a) = P(b). Show a = b. By extensionality. (--l ) Assume c Ea. So, {c} C a, and {c} E P(a). By hypothesis, {c} E P(b), so {c} (- ) Symmetrical.
c
b, and c E b.
Answers
273
#6. Assume Vx(x E a -> Sx). Show a C P(Ua). Assume bE a. By hypothesis, b is a set. Show: be Ua. Assume c E b. Since b E a, c E Ua. So, be Ua, and bE P(Ua). #8. Show pea) U PCb) c pea U b). Assume c E pea) U PCb). So, either c E P(a)or c E PCb). Case 1. c E pea). So, c Ca. Therefore,c C aU b, and c E pea U b). Case 2 is symmetrical. #10. Show a C b iff pea) C PCb). (-» Assume a C b. Show pea) c PCb). Assume c E pea). So, c C a. Since a C b, c c b. So, c E PCb). (<- ) Assume pea) c PCb). Show a C b. Assume c E a. Then, {c} C a, and {c} E pea). Since pea) C PCb), {c} E PCb), and {c} C b. Therefore, c E b. #12. Show VxP( {x}) = {0, {x}). Show P( {a}) = {0, {an. By Extensionality. (-> ) Assume b E P({a}). be {a}. Assume for contradiction that b f:. 0 and b f:. {a}. Since b is a nonempty set, there is acE b. If the only member of b was a, then b = {a}, by extensionality. So, there must be acE bsuch that c f:. a. Since b C {a}, c E {a}. So, c = a. Contradiction. So, b = 0 or b = {a}. By pair set, bE {0, {a}}. (<- ) Assume b E {0, {an. Either b = 0 or b = {a}. Case 1. b = 0.0 C {a}, so 0 E P({a}). So, bE P({a}). Case 2. b = {a}. {a} C {a}, so {a} E P({a}). Therefore, bE P({a}). Consequently, bE P({a}). #14. Assume a is transitive, and that every member of a is a set. Show P(Ua) C pea). Assume bE P(Ua). So, be Ua. Show b Ca. Assume c E b. So c E Ua. Therefore,c Ed and d E a, for some d. Since a is transitive, c Ea. Since a is a set of sets, b is a set, and be a. Therefore, bE pea). 16. Assume pea) = {a}, and that a is a set. Show that a = 0. Since 0 is a subset of every set, E pea). So, 0 E {a}. By unit set, 0 = a. 18. Show np(a) = 0. Assume for contradiction that b E nP(a). Since 0 is a member of every powerset of every set, 0 E pea). By n 1 , bE 0. But b fj. 0. Contradiction. Ch. 8.2 Cartesian products # 2. {(0,0),(0,1)} #4. {(O, 0), (0, 1), (1,0), (1, I)} #6. {(O,O), (1,0), (2,0)} #8. {(a, 1), (1,1), (2, 1),(0,2),(1,2),(2,2)} #10.0 #12. Assume Sa&Sb. Assume a x b = 0. It suffices to show that if a f:. 0, then b = 0. Assume a =1= 0. Then, by Cor. 7.1, there is a c such that c E a. Suppose for contradiction that b =1= 0. Then, by Cor. 7.1, there is a d such that dEb. So, by x , (c, d) E a x b. By = E, since a x b = 0, (c, d) E 0, a contradiction. #14. Suppose Sa&Sb. Assume a x b = b x a, and that a =1= 0 and bl 0. By Cor. 7.1, for some c, c E a and for some d, dEb. Since a and b are sets, we will show that a = b by Extensionality. (-> ) Assume e E a. Then (e,d) E a x b. By = E, (e,d) E b x a, so e E b. (<- ) Similar. #16. Assume Sa&Sb&Sc. Assumea =1= 0&a x be a x c. By Cor. 7.1, there is a d such that d E a. We will show that b C c by a subset proof. Assume that e E b, for arbitrary e. By xE, (d, e) E a x b. Since a x be a x c, (d, e) E a x c. By x , e E c. Thus, be c. #18. By Extensionality. (-> ) Assume d E a x (b n c). Then d = (e, J) and e E a and fEb n c. By n , fEb and J E c. Thus, (e, J) E a x band (e, J) E a x c. By n , (e, J) E (a x b) n (a x c). By = E, dE (a x b) n (a x c). (<- ) Suppose d E (a x b) n (a x c). Then d E a x band d E a x c. Then d = (e, J) and (e, J) E a x b, and d = (g, h) and (g, h) E a x c. Since d = d, by Seq =, e = g and J = h. Thus, (e, J) E a x c. By x , e E a, fEb, and J E c. By n , fEb n c. By x , (e, J) E a x (b n c). By = E, dE a x (b n c).
o
Logic, Sets and Functions
274
#20. By Extensionality. Assume d E a x (b - c). Then, d = (e, f) , e E a, and f E (b - c). Thus, fEb and f 'i- c. Clearly, (e, f) E a x b, and (e, f) 'i- a x c. So, dE (a x b) - (a x c).
Ch. 8.3 Relations # 2. 1J: N, R: N,F: N. #4. 1J: Z, R: Z,F: Z. #6. 1J: the half-open interval (0,1], R : [0,1), F: [0,1]. #8. 1J : 0, R : 0, F : 0. #10. 1J: {O, I}, R : {O, I}, F : {a, I}. #18. We show that R(R) - R(S) c R(R - S), by a subset proof. Suppose a E R(R) - R(S). Then, by Def -, a E R(R) and a 'i- R(S). By RE, there is a b such that Rba. It cannot be that Sba, since if it were, we would have a E R(S). Thus, (b, a) E R&(b, a) tf- S. So, by Def -, (b, a) E R - S. By RE, a E R(R - S).
#20. Show D(RnS) c D(R) nD(S). Assume a E D(RnS). So, (a,b) E RnS. Moreover, (a, b) E R and (a, b) E S. Thus, a E 1J(R)and a E 1J(S), and therefore a E D(R) n D(S). #22. Show R D(R) x R(R).
c
1J(R) x R(R). Assume (a, b) E R. Then, a E D(R) and b E R(R). So, (a, b) E
#24. Assume R = A x B, and that A and Bare nonempty sets. Show D(R) = A. By Extensionality. (--> ) Assume c E D( R). Then (c, d) E R. Then, c E A.
(<-) Assume c EA. Let dEB. Then, (c,d) E A x B, and so (c,d) E R. Therefore, c E D(R). Ch 8.4 Properties of Relations [Whenever a relation is labelled as "asymmetric", it is also anti- symmetric] #2. Reflexive, Anti-symmetric, Transitive. #4. Symmetric. #6. Irreflexive, Asymmetric, Transitive, Connected. #8. Reflexive, Anti-symmetric, Transitive, Connected & Strongly Connected. #10. Irreflexive, Symmetric, Connected. #12. Reflexive, Anti-symmetric, Transitive. #14. Reflexive, Anti-symmetric, Transitive. #16. On sets, including 0: Anti-symmetric. On the class of non-empty sets: Irreflexive, asymmetric, intransitive. #18. Interpret the relation to be: {(x, y) : y = a - x}, where a is a non-empty set. Then, the relation is: irreflexive, intransitive. #20. Anti-symmetric. #22. Anti-symmetric. #24. Anti-symmetric. #26. Anti-symmetric. #28. Reflexive, Symmetric, Transitive. #30. Symmetric. #32. Irreflexive, Asymmetric. #34. Irreflexive, Asymmetric. #36. Irreflexive, Symmetric. #38. Irreflexive, Asymmetric, Transitive. #40. Any relation is both reflexive and irreflexive on 0. Otherwise, this is impossible. #42. Identity. #44. < on reals. #46. is no more than 1 year older than. #48. is either 1 year older than or 1 year younger than. #50. is exactly 1 year older than.
Answers
#68. #70. #72. #74. #76. #78. #80. #82.
275
Irreflexive, symmetric, asymmetric, anti-symmetric, transitive, intransitive. Irreflexive, asymmetric, anti-symmetric, transitive, intransitive, connected. Same as 70. Reflexive, symmetric, antisymmetric, transitive. Anti-symmetric, transitive, connected. Same as 76. Symmetric, connected. Same as 80.
Ch. 8.5 Ordering Relations #2. Linear ordering. #4. Strict linear ordering. #6. Strict linear ordering. #8. Linear ordering. #10. Strict linear ordering. #12. Linear ordering. #14. Linear ordering. #16. Partial ordering. #18. Partial ordering. #20. Partial ordering. #22. Partial ordering. #24. Partial ordering. #26. None. #30. Strict partial ordering. #32. Assume R strictly partially orders A. Show that S is irreflexive and transitive on A. (a) Irreflexivity. Assume that b E A. Since R is irreflexive, ,Rbb. So, ,Sbb. (b) Transitivity. Assume Sbe and Sed, with b, e, d E A. We have Rde and Reb. Since R is transitive, we also have Rdb. Thus, Sbd. #34. Show that Q is irreflexive, asymmetric, transitive and connected and R x S. (a) Irreflexivity. Assume for contradiction that Q(a, b)(a, b) . By definition of Q, either Raa or (a = a&Sbb). Either disjunct leads to a contradiction, since both Rand S are irreflexive. (b) Asymmetry. Assume Q(a, b)(e, d) . Either Rae or (a = e and Sbd). Case 1. Rae. Since R is asymmetric, we have -.Rea. Since R is irreflexive, we have a =I- e. So, we have neither Rca nor (e = a and Sdb). By definition of Q, 'Q(e, d)(a, b) . Case 2. a = e and Sbd. Since R is irreflexive, we have ,Rca, and since S is asymmetric, we have ,Sdb. By definition of Q, -'Q(e, d)(a, b) . (c) Transitivity. Assume Q(a, b)(e, d) and Q(e,d)(e,f) . By definition of Q, we have either Rae or (a = e&Sbd). Case 1. Rae. By definition of Q, we have either Ree or (e = e and Sdf). Case 1a. Ree. Since R is transitive, we have Rae. So, Q(a, b)(e, f) . Case lb. (e = e&Sdf). Since e = e, we have Rae. So, Q(a, b)(e, f) . Case 2. (a = e&Sbd). By definition of Q, we have either Ree or (e = e&SdJ). Case 2a. Ree. Since a = e, we have Rae. So, Q(a,b)(e,f) . Case 2b. (e = e&Sdf). So, a = e. Since S is transitive, Sbf. So, Q(a,b)(e,f) . (d) Connectedness. Assume (a, b) =I- (e, d) . Show that either Q(a, b(c, d) or Q(c, d) (a, b) . Either a =I- e or a = c. Case 1. a =I- e. Since R is connected, either Rac or Rca. Consequently, either Q(a, b) (e, d) or Q(e, d) (a, b) . Case 2. a = c. Therefore, b =I- d. Since S is connected, either Sbd or Sdb. Consequently, either Q(a,b)(e,d) or Q(e,d)(a,b) . Ch 8.6 Relations between relations
I1 I
Logic, Sets and Functions
276
#26. Assume R is reflexive on A. Let b E A. So, Rbb. By def. of converse, R-1bb. So, R-I is reflexive. #28. Assume R is symmetric on A. Show R- 1 is. Assume R-1ab. By def. of converse, Rba. Since R is symmetric on A, Rab. So, R-1ba. #30. Assume R is antisymmetric on A. Show R-I is. Assume R- l ab and R- 1ba. So, Rba and Rab.
1
I !
j
Since R is antisymmetric, a = b. #32. Assume R is connected on A. Show R-1 is. Assume a :/=- b. Since R is connected, Rab or Rba. By definition of converse, R-I ba or R- l ab. #34. Assume R is intransitive on A. Show R- 1 is. Assume R- 1ab and R- 1bc. So, Rcb and Rba. Since R is intransitive, we have ,Rca. So, oR-lac. #36. Assume Rand S are strongly connected. Show R 0 S is. Let a, b E A. Since R is strongly connected, we have either Rab or Rba. Since S is strongly connected, we have Saa and Sbb. Case 1. Rab. We have Saa&Rab, so (a, b) E R 0 S. Case 2. Rba. We have Sbb&Rba, so (a, b) E R 0 S. #38. Show (R U S)-I = R- 1 US-I. By Extensionality.
(-+) Assume (a, b) E (R U S)-I. (R U S)ba. Either Rba or Sba. Case 1. Rba. R- l ab. So, (a, b) E R- 1 U S-I. Case 2. Symmetrical. (t- ) Assume (a, b) E R- 1 US-I. Either R- 1ab or S-Iab. Case 1. R- 1ab. Rba. So, (b,a) E RUS. So, (a, b) E (R U S)-I. Case 2. Symmetrical. #40. Show (R - S)-1 = R- 1 - S-l. By Extensionality. (-+ ) Assume (a, b) E (R - S)-l. (b, a) E (R - S). So, Rba and --Sba. R- 1ab and __ S-l ab. Thus, (a,b) E R- 1 - S-l. (t-) Assume (a,b) E R- 1 _ S- I . So, R- l ab and oS-lab. Rba and ,Sba. Thus, (b,a) E R-S. So, (a, b) E (R- S)-I. #42. Show D(S 0 R) c D(R). Assume a E D(S 0 R). So, (a, b) E (S 0 R), and Rac and Scb, for some band c. Since Rac, a E D(R). #44. Show To (S 0 R) = (T 0 S) 0 R. By Extensionality. (-+ ) Assume (a, b) E To (S 0 R). Then, (a, c) E (S 0 R)and Tcb. Moreover, Rad and Sdc. Since Sdc and Tcb, (d, b) E (T 0 S). Since Rad, (a, b) E (T 0 S) 0 R. (t- ) Assume (a, b) E (T 0 S) 0 R. Then Rad and (d, b) E (T 0 S). Moreover, Sdc and Tcb. Since Rad and Sdc, we have (a, c) E (S 0 R). Since Tcb, (a, b) E To (S 0 R). #46. Show (S U T)
R) U (T 0 R). By Extensionality. R. So, Rac and (c, b) E S U T. Either Scb or Tcb. Case 1. Scb. So, (c, b) E S U T. Since Rac, (a, b) E (S U T) 0 R. Case 2. Similar. (t- ) Assume (a, b) E (S 0 R) U (T 0 R). Either (a, b) E (S 0 R) or (a, b) E (T 0 R). Case 1. (a, b) E (S 0 R). Rac&Scb. (, c, b) E S U T. So, (a, b) E (S U T) 0 R. Case 2. Similar. #50. Show (A x B) 0 (A x B) c A x B. Assume (c, d) E (A x B) 0 (A x B). So, (c, e) E A x B, and (e,d) E A x B. Thus, c E A and dEB, so (c,d) E A x B. 0
R = (S
(t- ) Assume (a, b) E (S U T)
0
0
#52. Assume R is reflexive on A. Show D(R) = D(R- l ). By extensionality. (-+ ) Assume b E D(R). So, Rbc, and c E A (since R is a relation on A). Since R is reflexive on A, Rcc. So, R- 1cc, and c E D(R- l ). (t- ) Similar.
I I i
j l
I I
1
I
I ! 1
j i
1 J
j I I
l
#54. R is asymmetric on A iff Rand R- are disjoint. (-+) Assume R is asymmetric on A. Assume for contradiction that Rand R- 1 are not disjoint. So Rab and R- 1ab, for some a and b. Thus, Rab and Rba, contradicting the asymmetry of R. (t- ) Assume Rand R- 1 are disjoint. Show R is asymmetric on A. Assume for contradiction Rab and Rba. Since Rba we have R- l ab, so (a, b) ERn R- l , and Rand R- 1 are not disjoint. #56. Show R is transitive on A iff R 0 R c R
I j i
"1
I i
Answers
277
. (--> ) Assume R is transitive. Assume (a, b) E R 0 R. So, Rae and Reb. Since Ris transitive, Rab. (f-- ) Assume R 0 R c R. Show R is transitive. Assume Rab and Rbe. So, (a, e) E R 0 R. Consequently, (a, e) E R, i.e., Rae. #58. Show: R is reflexive, symmetric and transitive on A iff R-l 0 R = R. (--> ) Assume R is reflexive, symmetric, and transitive on A. Show R- 1 0 R = Rby extensionality. (--> ) Assume (a, b) E R- 1 0 R. So, Rae and R-1eb. Thus, Rbe. Since R is symmetric, Reb. Since R is transitive, Rab. (<- ) Assume Rab. Since R is reflexive, Raa. Since R is symmetric, Rba. Hence, R-1ab. Therefore, (a, b) E R- 1 0 R. (<- ) Assume R- 1 0 R = R. (a) Show R is reflexive. [We have to assume that R is nonempty.] Rab. R-1ba. So, (a, a) E R-l oR. Therefore, Raa. (b) Show R is symmetric. Assume Rab. Then, R-1ba. Since R is reflexive, Rbb. So, (b, a) E R- 1 oR. By hypothesis, Rba. (c) Show R is transitive. Assume Rab and Rbe. Show Rae. Since R is symmetric, Reb. So, R-1be. Since Rab, we have (a, e) E R- 1 0 R. By hypothesis, Rae. 60. x
+1<
62. x < y 66. x > y
Y
64. Z x Z 68. x 2:: y 72. Z x Z 76. xC z
70. Z x Z 74. x E z 78. x is a parent of y, and y has a brother.
80. x is a sibling or half-sibling of y, or x = y 84. x is a mother-in-law of y
82. x is a son of y, and x is not an only child 86. x is a sister-in-law of y (wife of a brother).
= y&x is married(assuming monogamy). Otherwise, x and y share a spouse. Ch. 8.8 Equivalence relations and partitions
#88. x
#2. Only one (0 itself). #4. Five: {{1,2,3}},{{1,2},{3}},{{1},{2,3}},{{1,3},{2}},{{1},{2},{3}}. #8. Assume R is an equivalence relation on R. Show R-l is. See problems # 26, # 28, and # 31 in Ch. 8.6. #12. Assume B partitions A, andC E B. By definition of partition (clause 1), uB = A. Since C E B, C
c
uB (see problem
#
9, Ch. 7.6), so C CA.
Ch 9.1 Functions and relations #2, #4, #6, #8: all yes. #10. No. For instance, V2 tf- N. #12. No, for the same reason as 10. #14. & #16. yes. #18. No. For instance, -1/2 to the -1/2 power, i.e.,
~is not a real number.
V- 1!2
#20. No. If x is a negative even integer, then the xth root of x is not a real number. #22. Yes. #24. Yes. #26. No. x! is not defined when x is not in N. #28. No. For instance, ~ is not in N. #30. No. For instance, -1 is not in N.
#32. Yes. #34. No. is undefined. #36. No. In(x) is undefined if x
!
#42. Show 9
0
f :A
--->
C.
~
O.
278
Logic, Sets and Functions
(a) Show 1)(g 0 f) = A. By extensionality. (-+ ) Assume d E 1)(g 0 f). So, (d, e) Ego f. So, fdh and ghe. Thus, dE 1)(1), so d EA. (<- ) Assume d E A. So, dE 1)(1). Thus, fde, for some e E B. e E 1)(g), so geh, for some h E C. Therefore, (d, h) Ego f. So, d E 1)(g 0 f). (b) Assume (g 0 f)dh and (g 0 f)de. Show that h = e. By definition of composition, we have fdj and gjh, for some j, and fdk and gke, for some k. Since f is a function, j = k. Consequently, we have both gjh and gje. Since 9 is a function, we have h = e. (c) Show go f c A x C. Assume (g 0 f)de. By def. of composition, fdh and ghe, for some h. So, d E A and e E C, since f C A x Band 9 C B xC. #46. Show f c 9 iff [1)(1) c 1)(g) and ' ) Assume f C g. Clearly, V(I) C V(g). Assume b E V(I). Then, (d, f(d)) E f· Since f C g, (d,f(d)) E g. Since dE V(g), (d,g(d)) E g. So, f(d) = g(d), since 9 is a function. (<- ) Assume V(I) C 1)(g) and '
Ch 9.2 Properties of functions #2. Injective. #4. None. #6. Injective. #8. None. #10. Injective. (R+ = the set of positive real numbers) #12. Injective. #14. Injective (assuming 0 (j. Z+) #16. None. {YO = 1 = 0. #18. None. #20. Injective #22. Bijective. #24. Injective. 0 (j. R(I). #26. Surjective. (Assume thatn + ~ is rounded up to n + 1.) #29. Assume f : A - t B. Show: f- 1 : B - t A iff f is bijective. (-> ) Assume f- 1 : B ----> A. Show f is bijective from A to B. 1)(1-1)
= B, so R(I) = B. Therefore, f is surjective. We must show that it is injective. Assume that d, e E A, and that d -=j:. e. Show f(d) -=j:. f(e). Assume for contradiction that f(d) = f(e). Since f is a function, (d, f(d)) E f and (e, f(e» E f. Consequently, (f(d), d) E f- 1 and (f(e), e) E f-l. Since f(d) = f(e), and f- 1 is a function, d = e, contrary to our assumption. (<- ) Assume f is bijective. Show f- 1 : B - t A. (a) Show 1)(1-1) = B. Since f is surjective, R(I) = B, so 1)(1-1) = B. (b) Show: uniqueness of values. Assume dEB, and that f- 1de and f- 1dg. Show e = g. Assume for contradiction that e -=j:. g. We have fed and fgd. Since f is injective, d -=j:. d, which is impossible. (c) Show: f- 1 c B x A. Immediate, since f C A x B. #33. Assume that f : R ----> R is strictly increasing. Show that f is injective. Assume thata, b E R, and a -=j:. b. Show f(a) -=j:. f(b). Since a -=j:. b, either a < b or b < a. Case 1. a < b. Then, f(a) < f(b), since f is strictly increasing. But then, f(a) -=j:. f(b). Case 2. Symmetrical. 1ba. Thus, fab. #42. Show R(I-l) = 1)(1). By extensionality. (-+ ) Assume a E R(I-l). So,
r
So a E V(I). (<- ) Similar. #44. Assume f is a bijection. Assume a ED. Show fof-1(a) = a. We know that fof-l : D ----> D. Consequently, (a,for1(a)) E fof-l. Consequently, (a, b) E f- 1 and (b,fof-1(a)) E f, for some b. By definition of converse, we have fba. Since f is a function, a = f 0 f-l(a).
Chapter 10.1: Natural numbers and definition by recursion
Answers
279
#2. Show VxVy(x + y = Y + x). By induction on x. Base case. ShowVy(O + y = Y + 0). This follows immediately from Th 10.5 and Cor. 10.4. Inductive hypothesis. Assume Vy(k + Y = Y + k). Show Vy((k + 1) + y = Y + (k + 1)) [by universal proof] Let m be an arbitrary number. (k+1)+m =(k+m)+1,byCor.1O.5 = (m + k) + 1, by inductive hypothesis, 'IE = m + (k + 1), by Th 10.5 #6. ShowVxVyVz(x, y, zEN -+ (x· y).z = x . (y. z)). By induction on x. Base case. Show VyVz(O· y) . z = o· (y . z). Let a and b be arbitrary numbers. O· a = 0, and O· b = 0, so (0· a) . b = O. O· (a· b) = O. Inductive hypothesis, Assume VyVz(k . y) . z = k . (y . z). Show VyVz((k + 1) . y) . z = (k + 1) . (y. z). Let a and b be arbitary numbers. (k + 1) . a = k· a + a, by Cor. 10.7. (k . a + a) . b = (k . a) . b + a· b, by the distribution law (problem # 3). By inductive hypothesis, (k . a) . b = k . (a· b). So,
((k+1)·a)·b
k·(a·b)+a·b (k+1)·(a·b)
by Cor. 10.7.
#10. Prove: succ(m) . n = (m· n) + n. By induction on n. Base case. Show succ(m) ·0= (m· 0) + O. succ(m) ·0 = 0, by Th. 10.6(a). (m . 0) = 0, andO + 0 = 0, by Th. 10.5(a). Inductive hypothesis. Assume Vmsucc(m) . k = (m· k) + k. Show Vmsucc(m) . succ(k) = (m· succ(k)) + succ(k) succ(m) . succ(k) = succ(m)· k + succ(m) by Th 10.6(b) = Um· k) + k) + succ(m) by inductive hypothesis. = succ(((m· k) + k) + m) by Th. 1O.5(b). = succ(((m· k) + m) + k) by assoc. & comm. of +. = ((m· k) + m) + succ(k) by Th. 10.5(b). = (m· succ(k)) + succ(k) by Th. 1O.6(b). Ch 10.2 Weak induction on the natural numbers #4. Base case. ~o i 2 = 12 = 1 1·(1+1)(2'1+1) _ 1·2·3 - Q - 1 6
-
6
-6-
Inductive hypothesis. Assume ~k i2 = k(k+1~(2k+1). Show
",k+ ·2 _ LJ
t
-
(k+1)(k+l+1)(2(k+l)+1) 6 .
~k+1 i2 = ~k i2 + (k + 1)2 = k(k+l)6(2k+1) + (k
= = = = =
2
+ (k + 1)2 (k + 1)[(2k:+k) + (k + 1)] (k + 1)[(2k:+k) + 6(kt l j (k + 1)[(2k:+kl + (6kt 6 )] (k + 1)[ (2k +kt 6k + 6 ) j (k+1)(;k +k)
2
+ 1)2,
by Ind. Hyp.
Logic, Sets and Functions
280
= (k + 1)[(k+2)~2k+3)] + 1)[(k+l+l)(~(k+l)+l)] Show 2:~0(2i)3 = 2(n2 + n)2.
= (k
#8. Base case. Show 2:?=0(2i)3 = 2(0 2 + 0)2 Inductive Hypothesis: assume 2:7=0(2i)3 = 2(k 2 + k)2 Show 2::7:~(2i)3 = 2((k + 1)2 + (k + 1))2 2::::~ i 3 = 2:::=0 i 3 + [2(k + 1)]3
= 2(e + k)2 + 8(k + 1)3
= 2k 2 (k + 1)2 + 8(k + 1)3 = 2(k + 1)2[2k 2 + 4k + 4]
= 2(k + 1)2(k + 2)2 = 2[(k
+ 1)(k + 2)]2
= 2(k 2 + 3k + 2)2
= 2((k 2 + 2k + 1) + (k + 1))2 = 2((k + 1)2 + (k + 1))2
#12. Show 22n - 1 is divisible by 3, for each positive integer n. Base case. Show 22 . 1 - 1 is divisible by 3. 22 . 1 - 1 = 3, and 3 = 3·1. Inductive hypothesis. Assume 22k - 1 is divisible by 3. So, 22k - 1 = 3· m, for some integer m. Show 22(k+1) - 1 is divisible by 3. 22(k+l) _ 1 = (2 2k . 22 ) - 1 = 4· 22k - 1 = 4· 2 2k - 4 + 3 = 4(2 2k - 1) + 3 = 4(3m) + 3, by inductive hypothesis = 3[4m + 1]. So 2 2(k+l) - 1 is divisible by 3. #16. Assume a is an integer, a < 0, and n ~ O. Show (a + l)n - 1 is divisible by a. By induction on n. Base case. Let n = O. (a + 1)0 = 1, and 1 -1 = O. 0 is divisible by a, i.e., O·a = O. Inductive hypothesis. For all a, (a + l)k - 1 is divisible bya. Show (a + l)k+1 - 1 is divisible by a. (a + 1)k+1 = (a + 1)k· (a + 1). By IH, (a + l)k - 1 is divisible by a, i.e., (a + l)k - 1 = b· a, for some number b. So, (a + 1)k+l = (b· a + 1) . (a + 1) = b· a2 + b· a + a + 1. So, (a + 1)k+1 - 1 = b· a2 + b· a + a = a· (b· a + b + 1). So it is divisible by a.
#20. By induction on n. Base case. Let n = 1. Then D = {I}. We have that R(I, 1), and that R is a strict partial order, and so irreflexive. Contradiction. Inductive hypothesis. Letn > 1. Assume that there is no strict partial ordering R on {I, 2, ... , n} such that Rnl and for all i such that 1 :s; i < n, Ri( i + 1). Show that there is no strict partial ordering R on {I, 2, ... , n, n + I} such that R( n + 1) 1 and for all i such that 1 :s i < n + 1, Ri(i + 1). Assume for contradiction that there is such an R. Thus, we have Rn(n + 1). Since R is transitive, we also have Rn1. Let R' be the restriction of R to {1,2, ... ,n}, that is: R' = {(a,b) : Rab&a,b E {1,2, ... ,n}}. R' is clearly irreflexive. Moreoever, R'nl and, for all i such that 1 :s; i < n, R'i(i + 1). It only remains to show that R' is transitive. Suppose we have R'ab and R'be. Then we have Rab and Rbe, and also Rac, since R is transitive. Since a and c are in the field of R' , we also have R'ac. So, R' is a strict partial ordering on {I, 2, ... , n}. Contradiction.
j
Answers
281
Thus, we have Rn(n + 1). Since R is transitive, we also have Rnl. Let R' be the restriction of R to {1,2, ... ,n}, that is: R' = {(a,b) : Rab&a,b E {1,2, ... ,n}}. R' is clearly irreflexive. Moreoever, R'nl and, for all i such that 1 :::; i < n, R'i(i + 1). It only remains to show that R' is transitive. Suppose we have R'ab and R'bc. Then we have Rab and Rbc, and also Rac, since R is transitive. Since a and c are in the field of R', we also have R'ac. So, R' is a strict partial ordering on {I, 2, ... , n}. Contradiction. #30. Prove: if 5 is a set with k members, and A is a proper subset of 5, then I A 1< k. Let f be a 1 - 1 function from 5 into the number k + 1 (thought of as a set). Since A is a proper subset of 5, A C V(f). Moreoever, there is some b such that b E Sand b 1. A. There are three possibilities: (i) k is not in the range of f, (ii) f(c) = k, and c 1. A or (iii) f(c) = k, and c E A. Case (i). If k is not in the range of f, then the restriction of f to A is a 1-1 function from A into k. So, I A 1< k. (ii) If f(c) =k, and c 1. A, then once again the restriction of f to A is a 1 - 1 function from A into k. So, I A 1< k. (iii) If j(c) = k, and c E A, then consider f(b). j(b) is not in the range of j restricted to A. Define f* : A ~ k as follows:
f* (d) It is easy to check that
f*
is 1 - 1, and sol A
j(d), ifd =I c j(b), ifd = c
1< k.
Ch. 10.3 Induction on Sets #2. (a) (00,), (1,0), (1, 1), (1,2), (2,0), (2, 1), (2, 2). (b) Prove that 2m ~ n, for (m, n) E T. Base case. (m, n) = (0,0) . 2· 0 ~ o. Inductive cases. (i) (m, n) = (j + 1, k) , and 2j ~ k. Clearly, 2j + 2 ~ k. (ii) (m, n) = (j + 1, k + 1)4, and42j ~ k. Clearly, 2j + 2 ~ k + l. (iii) (m, n) = (j + 1, k + 2) , and 2j ~ k. Clearly, 2j + 2 ~ k + 2. #6. Assume that X is whitish. Prove that X is indistinguishable from the standard white patch. Base case. Xis indistinguishable from the standard white patch. Inductive case. X is indistinguishable from Y, and Y is (by inductive hypothesis) indistinguishable from the standard white patch. Since indistinguishability is assumed to be transitive, X is also indistinguishable from the standard white patch. (b) Perceptual indistinguishability is not really transitive. #8. Assume that the range of b is {T}. Show that j(¢) = T. Base case. ¢ is a sentence letter. So, b(¢) = j(¢). Since the range of b is{T}, j(¢) = T. Inductive cases. There are two cases: conjunction and disjunction. (i) ¢ is a conjunction, ("l/J&o.). By inductive hypothesis, j("l/J) = T = j(o.). So, j(¢) = T. (ii) ¢ is a disjunction, ("l/J Va.). By inductive hypothesis, j("l/J) = T = j(o.). So, j(¢) = T.
Ch. 10.4 Graphs, Trees, and Lists
#4. (a) [aJ, [bJ, [aba], [abb], [baa], [bab]. (b) Prove that the length of w is odd. Base cases. w = [a] or w = [b]. In each case, the length of w is one, so it is odd. Inductive cases. Assume thatv E B, and w = [ab]l\v or w = [ba]l\v. By inductive hypothesis, the length of v is odd. The length of w is l(v) + 2, so it is also odd. #10. Prove: 'VxV(+- x) = V(x).
282
Logic, Sets and Functions
Inductive case. The connected graph G has k + 1 vertices. Let v be a vertex in G. The graph G' that results from erasing v and every edge to v from G has k vertices. By hypothesis,G' has at least k - 1 edges. Since G is connected, there must be at least one edge from v to a vertex in G'. Thus, G has at least kedges. Ch. 10.6 Formal languages #4. Show Vxs(x) = 1 + r(x). Let A be an arbitrary formula. Base case: A = [c], and c is a sentence letter. Then s(A) = 1, andr(A) = 0, so s(A) = r(A) + 1. Inductive hypothesis: Assume Vx(rnk(x) < rnk(A) ~ x(x) = r(x) + 1) Assume A = ['-, 'V' B or A = [,(,V' B/\ */\ C/\[')'], where * is a binary connective, and where rnk(B) < rnk(A), andrnk(C) < rnk(A). Case 1. A = ['-, ']/\ B, for some formula B such that rnk(B) < rnk(A). By IH, s(B) = r(B) + 1. s(A) = s(B), and r(A) = r(B), so s(A) = r(A) + 1. Case 2. A = ['(']/\ B/\ */\ C/\[')'], where * is a binary connective, and where rnk(B) < rnk(A), and rnk(C) < rnk(A). By IH, s(B) = r(B) + 1, and s(C) = r(C) + 1. s(A) = s(B) + s(C). r(A) = r(B) + r(C) + 1. So, s(A) = r(B) + 1 + r(C) + 1 = r(A) + 1. #12. Show: S contains exactly one 'a'. Base case: S = ['a']. ['a'] contains exactly one 'a'. Inductive hypothesis: Vx(rnk(x) < rnk(S) ~ x contains exactly one 'a'). S = ['(']/\C/\['b']I\[')'], and rnk(C) < rnk(S). So, by IH, C contains exactly one 'a'. S contains neither more nor less 'a's than does C. So S contains exactly one 'a'.
#16. A string is syntactically ambiguous iff it has two distinct grammatical trees. Consider the string ['0','1']. This can be constructed in two ways: (1) start with the sequence ['0'], by clause 1, and apply clause 3 to concatenate ['1'] to the end of the sequence; (2) start with the sequence ['1'], by clause 2, and concatenate ['0'] to the beginning of the sequence. fa'i6t
#20. A string contains at least one Base cases:
m 00
Jl&.
s if it contains an even number of """2E s.
1i'ffiit
Case 1. A = ['
m 00
1i'ffiit '].
In this case, A contains at least one
m 00
•
Case 2. A = [' . . . ']. In this case, A does not contain an even number of Inductive case: Assume rnk(A) > o. IH: Vx(rnk(x) < rnk(A) A = [' . . ']/\8/\['
~
~ (x contains an even number of
.. s
Assume A contains an even number of
~ s, then 8 contains at least one Ir ~ s. Then, 8 contains an even number of
A has exactly two more. Therefore, 8 contains at least one #26. Show: in any string, the number of
~
plus one. Letr(x) = the number of an arbitrary string. Show r(A) ::; s(A) Base cases:
= ['
~
'].
~ x contains at least one
Ir
'], and rnk(8) < rnk(A).
By IH, if 8 contains an even number of
Case 1. A
~ s.
r(A)
~
m 1m 00
,
. . s, since
and so does A.
s is less than or equal to the number of
sin x. Let s(x)
+ 1.
= 1, s(A) = 0.1::; 0+ 1.
(fi))
= the number of (fi)) sin x. Let
s
A be
Answers
283
. Case 2. A = [' ~ ']. rCA) = 0, seA) Inductive case: assume rnk(A) > O.
IH:Vx(rnk(x) < rnk(A)
---+
rex) ::; sex)
=
1. 0 ::; 1 + 1.
+ 1)
Cases:
A = [' ~ 'V'B, and rnk(B) < rnk(A). By IH, reB) ::; s(B) + 1. rCA) = reB). s(B) = seA) -1. So, by = E, rCA) ::; (s(A) -1) + 1. Therefore,r(A) ::; seA). Thus, rCA) ::; seA) + l.
Case 3.
~ ']"B"[' ~ '], and rnk(B) < rnk(A). By IH, reB) ::; s(B)+1. reB) = r(A)-l. s(B) = seA) - 1. By = E, rCA) - 1 ::; [seA) - 1] + 1. So, rCA) ::; seA) + 1.
Case 4. A = ['