Presupposition and Implicature in Compositional Semantics Edited by
Uli Sauerland and Penka Stateva
Palgrave Studies in Pragmatics, Language and Cognition Series Editors: Noël Burton-Roberts and Richard Breheny Series Advisors: Kent Bach, Anne Bezuidenhout, Robyn Carston, Sam Glucksberg, Francesca Happé, François Recanati, Deirdre Wilson Palgrave Studies in Pragmatics, Language and Cognition is a new series of high quality research monographs and edited collections of essays focusing on the human pragmatic capacity and its interaction with natural language semantics and other faculties of mind. A central interest is the interface of pragmatics with the linguistic system(s), with the ‘theory of mind’ capacity and with other mental reasoning and general problem-solving capacities. Work of a social or cultural anthropological kind will be included if firmly embedded in a cognitive framework. Given the interdisciplinarity of the focal issues, relevant research will come from linguistics, philosophy of language, theoretical and experimental pragmatics, psychology and child development. The series will aim to reflect all kinds of research in the relevant fields – conceptual, analytical and experimental. Titles include: Anton Benz, Gerhard Jäger and Robert van Rooij (editors) GAME THEORY AND PRAGMATICS Reinhard Blutner and Henk Zeevat (editors) OPTIMALITY THEORY AND PRAGMATICS María J. Frápolli (editor) SAYING, MEANING AND REFERRING Essays on François Recanati’s Philosophy of Language Corinne Iten LINGUISTIC MEANING, TRUTH CONDITIONS AND RELEVANCE The Case of Concessives Ira Noveck and Dan Sperber (editors) EXPERIMENTAL PRAGMATICS Uli Sauerland and Penka Stateva (editors) PRESUPPOSITION AND IMPLICATURE IN COMPOSITIONAL SEMANTICS Christoph Unger ON THE COGNITIVE ROLE OF GENRE Forthcoming titles: Hans-Christian Schmitz OPTIMAL ACCENTUATION AND ACTIVE INTERPRETATION
Palgrave Studies in Pragmatics, Language and Cognition Series Series Standing Order ISBN 0–333–99010–2 Hardback 0–333–98584–2 Paperback (outside North America only) You can receive future titles in this series as they are published by placing a standing order. Please contact your bookseller or, in case of difficulty, write to us at the address below with your name and address, the title of the series and the ISBN quoted above. Customer Services Department, Macmillan Distribution Ltd, Houndmills, Basingstoke, Hampshire RG21 6XS, England
Presupposition and Implicature in Compositional Semantics Edited by
Uli Sauerland and
Penka Stateva
© Selection and editorial matter © Uli Sauerland and Penka Stateva 2007 Chapters © their authors 2007 All rights reserved. No reproduction, copy or transmission of this publication may be made without written permission. No paragraph of this publication may be reproduced, copied or transmitted save with written permission or in accordance with the provisions of the Copyright, Designs and Patents Act 1988, or under the terms of any licence permitting limited copying issued by the Copyright Licensing Agency, 90 Tottenham Court Road, London W1T 4LP. Any person who does any unauthorised act in relation to this publication may be liable to criminal prosecution and civil claims for damages. The authors have asserted their rights to be identified as the authors of this work in accordance with the Copyright, Designs and Patents Act 1988. First published in 2007 by PALGRAVE MACMILLAN Houndmills, Basingstoke, Hampshire RG21 6XS and 175 Fifth Avenue, New York, N.Y. 10010 Companies and representatives throughout the world. PALGRAVE MACMILLAN is the global academic imprint of the Palgrave Macmillan division of St. Martin’s Press, LLC and of Palgrave Macmillan Ltd. Macmillan® is a registered trademark in the United States, United Kingdom and other countries. Palgrave is a registered trademark in the European Union and other countries. ISBN-13: 978–0–230–00533–4 hardback ISBN-10: 0–230–00533–0 hardback This book is printed on paper suitable for recycling and made from fully managed and sustained forest sources. Logging, pulping and manufacturing processes are expected to conform to the environmental regulations of the country of origin. A catalogue record for this book is available from the British Library. Library of Congress Cataloging-in-Publication Data Presupposition and implicature in compositional semantics / edited by Uli Sauerland and Penka Stateva. p. cm.—(Palgrave studies in pragmatics, language, and cognition) Includes bibliographical references and index. ISBN-13: 978–0–230–00533–4 (cloth) ISBN-10: 0–230–00533–0 (cloth) 1. Semantics. 2. Presupposition (Logic) 3. Implication (Logic) 4. Pragmatics. I. Sauerland, Uli. II. Stateva, Penka, 1965– P325.P655 2007 401 .43—dc22 10 9 8 7 6 5 4 3 2 1 16 15 14 13 12 11 10 09 08 07 Printed and bound in Great Britain by Antony Rowe Ltd, Chippenham and Eastbourne
2006047823
Contents
Notes on Contributors
viii
1 Introduction
1
Uli Sauerland and Penka Stateva
2 Quantifier Dependent Readings of Anaphoric Presuppositions
12
Sigrid Beck 1 2 3 4
Introduction The presupposition of again Quantifier dependent again Conclusions and consequences
3 Licensing or
12 14 22 31 34
Regine Eckardt 1 2 3 4 5 6
Introduction Boolean and mis-behaved ‘or’ Explicit existential quantification and ‘or’ Modal existentials Interactions between different licensers Summary
4 Free Choice and the Theory of Scalar Implicatures
34 38 46 53 62 67 71
Danny Fox 1 2 3 4 5 6 7 8 9
Some background on scalar implicatures The problem of free choice permission Other free choice inferences Chierchia’s puzzle Sauerland’s proposal An alternative perspective Recursive exhaustification and FC Other existential quantifiers Singular indefinites v
72 80 85 89 90 95 102 105 106
vi
Contents
10 Other FC effects 11 Remaining issues 12 Conclusion
5 Partial Variables and Specificity
108 109 111 121
Gerhard Jäger 1 2 3 4 5 6
Introduction Specificity and scope Solution strategies Partial variables Partial variables and presuppositions Conclusion
6 Negated Antonyms: Creating and Filling the Gap
121 122 125 136 150 154 163
Manfred Krifka 1 Double negatives 2 Attempts to explain double negatives 3 Pragmatic strengthening within an epistemic theory of vagueness 4 Conclusion
7 A Pragmatic Constraint on Adverbial Quantification
163 165 168 175
178
Orin Percus 1 A possible line of argumentation, and reasons to reject it 2 A constraint on the use of sentences with adverbial quantifiers 3 Some consequences of this constraint 4 Consequences for ‘semantic partition’ 5 Concluding remarks
8 Transparency: An Incremental Theory of Presupposition Projection
178 183 191 201 206
214
Philippe Schlenker 1 The dynamic turn and the Transparency theory 2 The projection problem: basic results of the Transparency theory 3 The triggering problem: against a lexical treatment 4 Problems and prospects
215 223 232 236
Contents
9 Aspects of the Pragmatics of Plural Morphology: On Higher-Order Implicatures
vii
243
Benjamin Spector 1 2 3 4 5
A puzzle about plural indefinites Sketch of the analysis The proposal More complex cases An enigma that confirms the hypothesis: the modal presupposition induced by plural indefinites 6 Sauerland’s alternative 7 Conclusion
Index
243 245 251 257 264 267 271
282
Notes on Contributors Sigrid Beck is Professor of Linguistics at the University of Tübingen. She studied linguistics, computer science and psychology at Tübingen and Edinburgh. Her 1996 University of Tübingen dissertation is on Intervention Effects in questions. She has taught at the Universities of Massachusetts, Connecticut, and Potsdam. Her current work is on semantics and its interfaces to syntax and pragmatics, and addresses topics such as questions, plural predication, reciprocals, comparison constructions, and decomposition adverbs. Regine Eckardt is Professor of English at Göttingen University. She is working on topics in semantics and pragmatics, specifically focus semantics, events, aspect, and discourse particles. Her interests extend to historical linguistics and semantic reanalysis. The monographs Adverbs, Events, and Other Things and Meaning Change in Grammaticalization as well as the collected volume Words in Time reflect her work in these fields. In addition, Eckardt published several research articles and is Coeditor of the Journal of Semantics. Danny Fox is currently Associate Professor at the Massachusetts Institute of Technology, where he received his PhD in 1998. From 1998 to 2001, he was a junior fellow at the Harvard Society of Fellows. Among his publications are Economy and Semantic Interpretation, ‘Antecedent Contained Deletion and the Copy Theory of Movement’ (Linguistic Inquiry), and ‘Cyclic Linearization of Syntactic Structure’ (with David Pesetsky, Theoretical Linguistics). Gerhard Jäger is Professor of Linguistics at the University of Bielefeld in Germany. He received his PhD from Humboldt University in 1996, with a thesis on dynamic semantics. He is the author of Anaphora and Type Logical Grammar. Together with Anton Benz and Robert van Rooij, he co-edited Game Theory and Pragmatics (Palgrave Macmillan, 2005). Manfred Krifka is Professor of General Linguistics at Humboldt University, Berlin, and director of ZAS (Center for General Linguistics, viii
Notes on Contributors
ix
Berlin). He has taught at the University of Tübingen and the University of Texas, Austin. He is the author of two books and many other publications on the semantics and pragmatics of aspect, mass nouns and measure constructions, polarity items, and information structure. Orin Percus recently joined the faculty of the University of Nantes. He received his PhD from MIT in 1997 with a thesis on the semantics of indefinites. Since then he worked in Amherst, Milan, and Tübingen. His work investigated general principles that constrain the semantic contribution of embedded clauses, the way in which lexical material contributes to the contextual requirements of sentences, and the semantic composition of copular sentences. He is a founding editor of the journal Snippets. Philippe Schlenker is currently Professor of Linguistics at UCLA, and is affiliated with Institut Jean-Nicod (CNRS) in Paris. He studied philosophy and linguistics in Paris and Boston (PhD in linguistics, MIT, 1999; PhD in philosophy, EHESS, Paris, 2002; Habilitation, EHESS, 2005). His initial work was concerned with the semantics of indexicals and attitude reports (‘A Plea for Monsters’, Linguistics and Philosophy, 2002). His current interests include issues in natural language semantics (intensional phenomena, anaphora, presuppositions) and in philosophical logic (paradoxes). Benjamin Spector is currently a Junior Fellow at the Harvard Society of Fellows. He has presented his PhD dissertation in 2006 (University of Paris 7). From 2003 to 2006, he has been teaching linguistics and logic at Ecole Normale Supérieure in Paris. His work’s main focus is formal pragmatics, especially scalar implicatures and the semantics and pragmatics of questions and answers.
This page intentionally left blank
1 Introduction Uli Sauerland and Penka Stateva Zentrum f ür Allgemeine Sprachwissenschaft (ZAS), Berlin
A speaker of a language can quickly judge the grammaticality, truth and felicity of a sentence on the basis of his beliefs about what is the case and who is talking. A basic goal of semanticists is to model what we do when we judge a sentence to be true. A basic goal of pragmaticists, on the other hand, is to model what we do when we judge a sentence to be felicitous. Semanticists and pragmaticists have each proposed explicit formal models of these mental mechanisms. These models map a pair of a sentence and a model-theoretic representation of what we may believe to be the case onto a truth value and a measure of felicity. Semantics and Pragmatics have both established themselves as independent fields in the 1970s and 1980s. During that period, the distribution of labor between semantics and pragmatics was intensively debated. (Davis (1991) and Kasher (1998) have collected much of the relevant literature in useful volumes.) The separation of semantics and pragmatics has made enormous progress possible in both fields. At the same time though, it has had the unfortunate consequence that interconnections between semantics and pragmatics may have not received enough attention. To begin, we briefly survey the main lines of inquiry in semantics and pragmatics that this volume builds on. For the computation of truth conditions, the dominant view is that this process involves at least three things: One is the structured mental representation of a sentence that syntax provides. The second are the interpretations of the primitive parts of the sentence, which are provided by their lexical entries. The third is the compositional interpretation procedure: A recursive procedure that assigns an interpretation to any sub-structure of the syntactic representation on the basis of the interpretations of the immediate parts of this sub-structure. Recursive application of the interpretation function then successively builds up the 1
2 Presupposition and Implicature in Compositional Semantics
interpretation of larger and larger sub-structures, until the interpretation of the entire sentence results. This view can be attributed to Montague (1974) and Lewis (1970) and has been developed much further since. In the following, we assume the formal model of the interpretation function laid out in the textbook by Heim and Kratzer (1998). For the computation of felicity conditions, the dominant view distinguishes at least two kinds of felicity conditions: Presuppositions (Frege 1879, Strawson 1952) and Implicatures (Grice 1989). Here and in the following, we follow Karttunen and Peters (1979), Horn (1990) and others in assuming that Grice’s Conventional Implicatures are to be identified with presuppositions, and use the term Implicatures to exclusively refer to Grice’s Conversational Implicatures. The mechanisms giving rise to the presuppositions and implicatures of a sentence are also thought to be fundamentally different. We consider first presuppositions, then implicatures. Presuppositions are represented as felicity conditions in the lexical entries of certain morphemes. The computation of the presuppositions of a sentence proceeds in a compositional fashion analogous to the computation of the sentences truth conditions. Technically, presuppositions can be represented as domain restrictions of the functions that represent the meanings of lexical entries, constituents, and sentences (Blau 1978, Heim 1992). A presupposition failure then arises when the value of a function–argument combination is undefined. For example, the domain of the definite determiner the is restricted to properties P such that at least one individual has property P. Therefore, the sentence (1) leads to a presupposition failure if there are no books.
(1)
I didn’t see the books.
Implicatures, on the other hand, are generally held to arise from the interaction of sentence meaning and conditions on felicitous discourse. Grice (1989) proposed one view of the computation of implicatures that continues to be very influential. Consider, for example, the analysis of (2). (2) is true in a situation where I know that the Philharmonic played all Beethoven symphonies this season, however, it is infelicitous because the more informative sentence The Philharmonic played all symphonies by Beethoven is also true.
(2)
The Philharmonic played some symphonies by Beethoven this season.
Introduction 3
Therefore, (2) implicates that the Philharmonic did not play all nine symphonies by Beethoven this season. Horn (1972) observes that not all sentences that are more informative than what is said are considered for implicature computation. For example, They played only some Beethoven symphonies this season is more informative than (2), but (2) does not implicate that they didn’t play only some Beethoven symphonies Horn amends Grice’s proposal with a restriction to scalar alternatives; for example, many and all are scalar alternatives, but not many and only many. Horn (1989), Gazdar (1979), Hirschberg (1985), Matsumoto (1995), Sauerland (2004) have developed this view further, in particular for the analysis of scalar implicatures. In the literature cited above, the three components of sentence interpretation have been considered separately. This separation is justified on theoretical and methodological grounds and has made substantial progress in all three areas possible. However, the separate treatment of the truth conditions, implicatures, and presuppositions, while it was important to make progress, bears certain risks because of the narrow view adopted: It may overlook important interactions between the three components. Furthermore, it may induce researchers to focus on phenomena that fit neatly into one of the three components. This reasoning provides the motivation for the present volume: The contributions concern, on the one hand, empirical phenomena that do not fit neatly into one of the three components and, on the other hand, the interaction of the three components. Specifically, three of the contributions (those by Eckardt, Fox, and Percus) focus on the question of which component should best provide the explanation for a specific empirical phenomenon. Questions of this type are pertinent to the other chapters too, but the other chapters concern directly the interaction between the different components. Two of them concern the interaction between implicatures and truth conditions (those of Krifka and Spector). Three other chapters focus on the interaction between presuppositions and truth conditions (those by Beck, Jäger, and Schlenker). Percus reconsiders the analysis of restrictions on the use of individual level predicates. As is well known, many examples that contain both a temporal adverbial and an individual level predicate are judged ungrammatical, as is illustrated by (3). (3)
*Ingrid was always Swedish.
Percus establishes a parallel between (3) and sentences like (4), which have the same relevant properties and same status but become acceptable if used under suitable pragmatic conditions.
4 Presupposition and Implicature in Compositional Semantics
(4)
(*)The student who finished first was always Swedish.
If, for example, (4) is uttered in a situation in which a number of exams are held during the past week, and different people finish first each time, the speaker can successfully use that sentence to inform the audience of the nationality of the winners. (4) crucially shows that these restrictions on the use of individual level predicates have pragmatic, rather than grammatical origin (pace Kratzer 1995). To identify these pragmatic factors, Percus considers pairs of situations tied to examples like (4). One member of each pair renders the sentence fully acceptable while in the other the sentence gets a much worse status. Examining the relations within these pairs of situations Percus concludes that the status of sentences containing individual level predicates and temporal quantifiers depend on two additional factors: (i) the time intervals that context makes salient and (ii) what the parties to conversation think is being taken for granted. He argues that given standard assumptions about (ii) if the truth of a quantified statement is implied by a variant of it which involves a subset of the time intervals the original involves the original sentence is predicted to sound bizarre. This idea is made precise by establishing a general pragmatic principle which precludes the usage of otherwise legitimate logical forms and is tested in examining related phenomena. Percus’s contribution targets mainly the division of labor between grammar and pragmatics but it also gives an intriguing twist to the study of the interaction between sentence meaning and conditions on felicitous discourse. As he notes, the principle he argues for, has a very specific status among other familiar pragmatic principles since it involves comparison among propositions that are equally informative. It would fall therefore into the Gricean category of Manner, however, unlike other Manner maxims Percus principle does not apply to a surface form. Both Eckardt’s and Fox’s chapters concern the free choice effect with disjunction. The basic problem discovered by Kamp (1973) is that (5a) and (5b) are perceived to be equivalent, but this is not predicted on the basis of the standard logical meanings of permissive modality, conjunction, and disjunction. (5)
a. You may have coffee or tea. b. You may have coffee and you may have tea.
Recent work on the topic has therefore claimed that the standard logical meanings of permissive modality and disjunction are not correct for
Introduction 5
natural language (Zimmermann 2000 and others). But this conclusion is challenged by Eckardt’s and Fox’s chapters. Both chapters independently make the novel observation that plural indefinites also give rise to something akin to a freeehoice effect in examples like (6). (6)
a. Some linguists were having coffee or tea. b. Some linguists were having coffee and some linguists were having tea.
Such examples show that the free choice effect should not be analyzed as tied to an analysis of the interaction of disjunction and existential modality. Fox strengthens this point further by showing that the logical duals of existential quantification and disjunction give rise to the same effect if negated. All these facts argue for an analysis based on the pragmatics of existentials and disjunction. Specifically what Eckardt and Fox both pursue is an analysis of the free choice effect as a conversational implicature. However, the two chapters pursue two different lines of analysis from this point on. Eckardt pursues the intuition that the implicature leading from (6a) to (6b) has the same basis as that from (7) to (6b). The core of her proposal is the claim that (6a) involves distributive quantification over a specific set of linguists. This predicts that the indefinites should behave like the plural definite or universal in (7). (7)
The/All linguists were having coffee or tea.
Eckardt then goes on to extend this account to other cases of free choice effects. Fox’s contribution presents in detail the conflict between the free choice effect and the neo-Gricean approach to scalar implicatures. The neo-Gricean account in the version developed by Sauerland (2004), as Fox shows, predicts that (8) should be an implicature of (5a), but this would contradict (5b). (8)
The speaker is not sure whether you may have coffee and the speaker is not sure whether you may have tea.
Fox develops a new account for scalar implicatures that does not derive them in the same way as other conversational implicatures. Rather, Fox argues that scalar implicature computation must precede the computation of other conversational implicatures and therefore places them in
6 Presupposition and Implicature in Compositional Semantics
the compositional semantic system (cf. Chierchia 2004). Fox shows then that free choice effects can be derived in this architecture on the basis of a novel proposal for exhaustification that can apply recursively. In example (6a), exhaustification at the second level involves a comparison among the three propositions in (9). (9)
a. You may have tea or coffee but not both. b. You may have tea but no coffee. c. You may have coffee but no tea.
Because (9a) is entailed by (9b) and (9c), Fox predicts as the result of the second level of exhaustification the free choice effect: You may have tea or coffee, but not both, and neither tea nor coffee is forbidden. The chapter by Spector is one of the two that concerns the interaction between implicatures and truth conditions. His chapter also presents a second application of Fox’s idea to apply exhaustification recursively; namely, to plural noun phrases. Plural noun phrases in many cases require plural reference, but there are exceptions as recently summarized by Sauerland et al. (2005). Spector asks whether the plurality requirement of plural noun phrases is part of the truth conditions, presuppositions or an implicature, arguing in favor of the latter. Strong initial support for the implicature analysis is provided by the observation that in the scope of negation the plurality requirement does not apply. This is illustrated by the fact that (10b) does not express the negation of (10a). (10)
a. The homework contains difficult problems. b. The homework doesn’t contain difficult problems.
The difference between (10a) and (10b) follows if we assume that the plural does not exclude singular reference. Outside the scope of negation, however, use of the plural implicates that the uniqueness implication of the singular would have been too strong. The result is a plurality implicature. This effect has been observed in previous work in semantics, but all previous attempts of explaining it have been unsatisfactory as Spector argues. The difficulty arises from the assumed logical relation of singular and plural sentences: The implicature analysis assumes that the singular example (11a) entails the plural example (10a) as long as implicatures are put aside. But, then the negated plural example (10b) is predicted to entail the negated singular example (11b). The standard implicature analysis predicts therefore that there should be a clear difference in interpretation between (10b) and (11b). However,
Introduction 7
the two sentences are more or less equivalent – Spector derives what difference there exists in the final section of his chapter. (11)
a. The homework contains a difficult problem. b. The homework doesn’t contain any difficult problem.
Spector elegantly solves this problem by applying the idea of recursive exhaustification from Fox’s chapter. On Spector’s proposal, the first level of exhaustification only compares the plural (10a) to the proposition expressed by (12a) which is equivalent to (10a). At the second level, however, (10a) is compared to (12b), which is the exhaustified version of (12a), and since (12b) entails (10a), the plurality implicature is predicted. For the negated plural in (10b), however, this line of argumentation does not go through since the negation of (12b) is not the exhaustified version of (12a). In this way, Spector provides an nice account of the plurality requirement of the plural. (12)
a. The homework contains at least one difficult problem. b. The homework contains exactly one difficult problem.
The chapter by Krifka discusses another case where the application of negation does not lead to the logical inverse. Consider (13): (13)
a. John is happy. b. John is not unhappy.
Contrary to expectation, the double negative in (13b) is not synonymous with the positive predicate in (13a). Krifka offers a solution to this puzzle by appealing to a process of pragmatic strengthening within an epistemic theory of vagueness (cf. Williamson 1994). This process involves an intricate interplay of a number of pragmatic principles which are applied in the use of negated antonyms. The final three chapters concern presuppositions and their relation to truth conditions. Beck focuses on an observation due to Soames (1989) and Heim (1992) that presuppositions triggered by again are tied to an anaphoric element. Speakers relate Heim’s (14a), for example, but not (14b) to an inference that John’s birthday precedes in time Mary’s birthday. (14)
a. We will have pizza on John’s birthday, so we shouldn’t have pizza again on Mary’s birthday.
8 Presupposition and Implicature in Compositional Semantics
b. We will have pizza on John’s birthday, so we shouldn’t have pizza on Mary’s birthday. It is clear that again triggers a presupposition that there is some time interval preceding Mary’s birthday in which we have already had pizza. But among the many possible previous pizza-eating occasions only one is relevant: that which involves John’s birthday, hence the inference pattern. This intuition points to the fact that the presuppositions of again in (14) involve a specific time period. Beck takes this fact to suggest that again must be related to an anaphoric temporal element for which the assignment gives John’s birthday as a value that satisfies the presuppositions of again. In other words, this temporal element has the properties a free variable in (14). The parallel between this anaphoric element and natural language variables like pronouns, restrictions on quantifier domains and traces is not accidental: Beck proposes that again comes with a lexical entry which specifies that its most internal argument is that anaphoric element, represented as a temporal variable. She further argues that like other variables, it is discernable in syntax and comes in the same varieties: it can be free, bound, or involve a complex structure. Beck tracks original examples of each kind addressing in most detail the interaction of again with quantifiers and the resulting anaphoric presuppositions. It is shown that the proposed logical forms lead to definedness and truth conditions that are in concord with speakers’ intuitions. The chapter concludes with an extention of the proposal to the anaphoric presuppositions of too and also. Jäger’s chapter is concerned with the analysis of specific indefinites. The topic is familiar from the numerous discussions in the literature of their notorious scopal properties. Unlike other quantified expressions, specific indefinites can escape scope islands: (15)
a. If most experiments succeed, John will be happy. b. If an experiment succeeds, John will be happy.
Jäger proposes a theory of specific indefinites from which these properties naturally follow while he avoids the pitfalls of other influential proposals. His theory is a synergy of ideas from DRT, the choice function approach, and theories that associate specificity with presuppositions. Indefinites are proposed to be analyzed as partial variables: local assignments determine the value of the variable which, however, comes with definedness conditions and only if these are fulfilled the indefinite can refer. Since
Introduction 9
indefinites are translated as variables, they can be bound by existential closure which can apply at every sentential level thus providing for the observed scopal freedom of specific indefinites. The core proposal in the chapter follows a trend in the literature to view specificity as a kind of presupposition phenomenon. Furthermore, it is extended to suggest that in general presuppositions are restrictions on partial variables. This view elegantly accounts for presupposition projection and accommodation. Specific indefinites are treated on a par with standard pesupposition triggers but, in addition, they are subject to the Novelty Condition (Heim 1982). Schlenker’s contribution is concerned with basic issues of presupposition triggering and presupposition projection. He does not accept the notion of presupposition as basic, but seeks to derive it from general pragmatic principles. Specifically, Schlenker proposes a new pragmatic principle, Be Articulate, that constrains the use of lexical items that express more than one separate truth-conditional contribution simultaneously. Namely, Be Articulate requires that only one of these components of the meaning must introduce new information. He shows that this straightforwardly derives the principles of presupposition projection: For example, assume that to stop Q-ing expresses two separate truth-conditional contributions: not doing Q and having done Q in the past. Then the use of stop in (16) satisfies Be Articulate when either one of the contributions of to stop smoking is part of the common ground. (16)
Sue stopped smoking.
If the sentential context in which stop occurs, however, establishes that one of the contributions of stop is redundant, nothing is required of the common ground. This provides an explanation for examples of presupposition filtering like (17a). Schlenker’s account, furthermore, straightforwardly extends to examples like (17b), where it is the other contribution of stop that is redundant. (17)
a. If Sue ever smoked, she has stopped smoking. b. If Sue isn’t smoking, she has stopped smoking.
The work presented in this volume is one of the results of a research network on Semantics and Pragmatics called the Semantiknetzwerk. The network has been funded by the German research council DFG (grants SA 925/2-1 and SA 925/2-2, principal investigator Uli Sauerland), and we are very grateful for this financial support. The support has allowed
10
Presupposition and Implicature in Compositional Semantics
us to hold five meetings up to now with a sixth meeting planned for September 2006. Four of the meetings were held in Berlin at the Centre for General Linguistics, and one was held in Nijmegen, Holland. Many of the chapters contained in this volume have been presented and discussed at one of the five meetings. In addition to the authors and editors of the present volume, the following researchers took part in these meetings regularly: Richard Breheny Irene Heim, Graham Katz, Cecile Meier, Doris Penka, and Ede Zimmermann. Furthermore the following researchers took place in some of the meetings: Bridget Copley, Cornelia Endriss, Bart Geurts, Beáta Gyuris, Andreas Haida, Elke Kasimir, Jason Mattausch, Robert van Rooij, Katrin Schulz, Magdalena Schwager, Yael Sharvit, Arnim von Stechow, Kazuko Yatsushiro, and several others. We would like to thank all participants for making the meetings of the Semantiknetzwerk a lively and stimulating forum for the discussion and exchange of ideas leading to this book and several other publications.
References Blau, U.: 1978, Die dreiwertige Logik der Sprache. Ihre Syntax, Semantik und Anwendung in der Sprachanalyse. de Gruyter, Berlin, Germany. Chierchia, G.: 2004, ‘Scalar implicatures, polarity phenomena, and the syntax/pragmatics interface’, in A. Belletti (ed.), Structures and Beyond, pp. 39–103. Oxford University Press, Oxford, UK. Davis, S. (ed.): 1991, Pragmatics. A Reader. Oxford University Press, Oxford, UK. Frege, G.: 1879, Begriffsschrift. Eine der arithmetischen nachgebildete Formelsprache des reinen Denkens. Neubert, Halle, Germany. Gazdar, G.: 1979, Pragmatics: Implicature, Presupposition, and Logical Form. Academic Press, New York, N.Y. Grice, P.: 1989, Studies in the Way of Words. Harvard University Press, Cambridge, Mass. Heim, I.: 1982, ‘The semantics of definite and indefinite noun phrases’, PhD dissertation, University of Massachusetts, Amherst. Heim, I.: 1992, ‘Presupposition projection and the semantics of attitude verbs’, Journal of Semantics 9, 183–221. Heim, I. and A. Kratzer: 1998, Semantics in Generative Grammar. Blackwell, Oxford, UK. Hirschberg, J.: 1985, ‘A theory of scalar implicature’, PhD dissertation, University of Pennsylvania, Philadelphia, Penn. Horn, L. R.: 1972, ‘On the semantic properties of logical operators in English’, PhD dissertation, University of California, Los Angeles. Horn, L. R.: 1989, A Natural History of Negation. University of Chicago Press, Chicago, Ill. Horn, L. R.: 1990, ‘Hamburgers and truth: why Gricean explanation is Gricean’, in K. Hall et al. (eds), Proceedings of the Sixteenth Annual Meeting of the Berkeley Linguistics Society, pp. 454–71. Berkeley Linguistics Society, Berkeley, Calif.
Introduction 11 Kamp, H.: 1973, ‘Free choice permission’, Proceedings of the Aristotelian Society, New Series 74, 57–74. Karttunen, L. and S. Peters: 1979, ‘Conventional Implicature’, in C. Oh and D. Dinneen (eds), Presupposition (Syntax and Semantics 11), pp. 1–56. Academic Press, New York, NY. Kasher, A. (ed.): 1998, Pragmatics: Critical Concepts. Routledge, London, UK. Kratzer, A.: 1995, ‘Stage-level and individual-level predicates’, in G. N. Carlson and F. J. Pelletier (eds), The Generic Book, pp. 125–75. University of Chicago Press, Chicago, Ill. Lewis, D.: 1970, ‘General semantics’, Synthese 22, 18–67. Matsumoto, Y.: 1995, ‘The conversational condition on Horn scales’, Linguistics and Philosophy 18, 21–60. Montague, R.: 1974, ‘The proper treatment of quantification in ordinary English’, in R. Thomason (ed.), Richard Montague: Selected Papers, pp. 247–70. Yale University Press, New Haven, Conn. Sauerland, U.: 2004, ‘Scalar implicatures in complex sentences’, Linguistics and Philosophy 27, 367–91. Sauerland, U., J. Andersen, and K. Yatsushiro: 2005, ‘The plural is semantically unmarked’, in S. Kepser and M. Reis (eds), Linguistic Evidence – Empirical, Theoretical, and Computational Perspectives, vol. V, pp. 413–34. Mouton d’Gruyter, Berlin, Germany. Soames, S.: 1989, ‘Presupposition’, in D. Gabbay and F. Guenthner (eds), Handbook of Philosophical Logic, vol. IV, pp. 553–16. Reidel, Dordrecht, Netherlands. Strawson, P.F.: 1952, Introduction to Logical Theory. Methuen, London, UK. Williamson, T.: 1994, Vagueness. Routledge, London. Zimmermann, T. E.: 2000, ‘Free choice disjunction and epistemic possibility’, Natural Language Semantics 9, 255–90.
2 Quantifier Dependent Readings of Anaphoric Presuppositions Sigrid Beck Universität Tübingen
It has been observed (e.g. Cooper (1979), Chierchia (1993), von Fintel (1994), Marti (2003)) that the interpretation of natural language variables (overt or covert) can depend on a quantifier. The standard analysis of this phenomenon is to assume a hidden structure inside the variable, part of which is semantically bound by the quantifier. In this chapter I argue that the presupposition of the adverb ‘again’ and other similar presuppositions depend on a variable that gives rise to the same phenomenon.
1
Introduction
Von Fintel (1994), following Heim (1991), notes that (1a) has the reading paraphrased in (1b) and proposes the analysis sketched in (1 ). (1)
(1 )
a. Only one class was so bad that no student passed. b. Only one class was so bad that no student in that class passed. a. [[noC ]](A)(B) = 1 iff (g(C) ∩ A) ∩ B = { } b. [[only one class] [λx[ tx was so bad that [IP [DP nof(x) student ] [ passed]]]] c. only one class x: x was so bad that nof(x) student passed d. f: x → {y: y is in x}
The crucial property of the analysis is that the quantifier is assumed to have a covert restriction variable (cf. e.g. Westerstahl (1984), Rooth (1992), von Fintel (1994)), its resource domain variable (C in (1 a)). The value of this restriction varies depending on the quantifier ‘only one class’. Hence it is proposed that the resource domain variable itself 12
Quantifier Dependent Presuppositions 13
has a hidden internal structure, consisting of a variable x bound by ‘only one class’ and a free function variable that assigns to x the set of students in x. The suggestion that an expression of natural language that normally functions as a variable (this is meant in the following when I say ‘natural language variable’) is internally complex and can for this reason covary with a quantifier (without being bound by it) is not new. It has been made by Cooper (1979) to analyze certain pronouns, and extended to traces by Chierchia (1993). Interesting consequences especially for the theory of Logical Form have been investigated by Marti (2003). Thus a natural language variable can be free, bound or internally complex. In this chapter I argue that the adverb again gives rise to interpretations that similarly show covariation with a quantifier, and that the analysis sketched for the resource domain variable case can be extended to again. To give an example, consider (2), (3). (2) Bill was sick on his birthday. He was sick again on Thanksgiving. (2 ) Bill’s birthday is before Thanksgiving. (3) In 1995, 1996 and 1998, Bill was sick on the day of the department party. In each of these years, he was sick again on Thanksgiving. (3 ) In each of these years the department party was before Thanksgiving. Intuitively, the adverb again makes reference to a time preceding the time that the sentence with again talks about. Thus (2) triggers the inference (2 ), because the use of again in the second sentence requires there to be a previous time at which Bill was sick, and the only relevant time in the context is the one mentioned in the first sentence, Bill’s birthday. The example in (3) shows that which earlier time is picked can depend on a quantifier – here: ‘in each of these years’. Example (3) triggers the inference (3 ). The use of again requires that for each of the years 1995, 1996 and 1998, the department party that year precedes Thanksgiving that year. I will extend the analysis of (1) to (3) and related examples. The next section of this chapter introduces again and the theory of again’s presupposition I will use. In the subsequent Section 3 I examine the quantifier dependent cases and develop an analysis for them. Conclusions are presented in Section 4.
14
Presupposition and Implicature in Compositional Semantics
2
The presupposition of again
2.1
Basics
The adverb again indicates repetition, as illustrated by the simple example below. (4a) is only appropriate if (4b) holds. Thus it is standardly assumed that (4a) presupposes (4b). (4b) is given slightly more formally in (4c). t1 stands for the time interval that the sentence (4a) is about. (4)
a. John snored again. b. John had snored before. c. ∃t [t < t1 & John snored at t ]
I adopt a view of presuppositions according to which they are restrictions on appropriate contexts (compare Stalnaker (1973), Heim (1990), Kadmon (2001) among others). The restriction that (4a) imposes on contexts in which it can be uttered is given in (5a) (informally) and (5b) (more precisely); c is Stalnaker’s context set; c entails p iff c is a subset of p. (5)
a. (4a) is only felicitous in a context that entails that John had snored before. b. c ⇒ λw.∃t [t < t1 & John snored in w at t ]
Again triggers this presupposition. It is associated with a lexical entry like the one in (6) (see for example Stechow (1996), Fabricius-Hansen (2001), Jäger and Blutner (2003), van der Sandt and Huitink (2003) for recent discussion; the entry in (6) will serve as a starting point for present purposes). The definition (6) entails that the truth value of a sentence containing again may be undefined in a world. I take it that a sentence S is felicitous in a context c only if the presuppositions of S are true in all worlds in c; or in other words, S can only be added to c if c entails the presuppositions of S (cf. (7 )). This is how presupposition as undefinedness in simple sentences is mapped to appropriateness in a context (von Fintel (2003) calls (7 ) ‘Stalnaker’s Bridge’). The application to the example is illustrated in (7); t1 is a free time variable that is assigned a time interval by the variable assignment function g; in the example, suppose that is the interval t1. Thus (4a) can only be uttered felicitously in context c if (5b) is the case, as desired. (6)
[[again]] (p<τ , <s,t)(t)(w)
=1 if p(t)(w) & ∃t [t
Quantifier Dependent Presuppositions 15
(7)
(7 )
[[ [[[ John snored] again ] t1 ]]]g (w) is defined only if ∃t [t < g(t1) & John snored in w at t ] If defined, it is true iff John snored in w at g(t1). S is only felicitous in c if for all w ∈ c: [[S]]g (w) is defined.
As a general notational convention, I try to use boldface t1, t2 etc for free metavariables over time intervals and regular t1, t2 etc as time pronouns in the object language (the linguistic structures of English). Later on, the notation in (7 ) for the information in (7) will be handy, where @ marks presupposed material (cf. Beaver (2001)). (7 )
John snored in w at g(t1) & @(∃t’[t’
While (6) is (roughly) what is standardly used, it seems upon closer inspection that we need a more sophisticated version of again’s presupposition. It has been argued by Soames (1989) (who refers to unpublished work by Kripke), and by Heim (1990) that the presupposition introduced by again is about a specific time, not existential. An example I take from Heim (1990) is given below. (8)
a. We will have pizza on John’s birthday, so we shouldn’t have pizza again on Mary’s birthday. b. We will have pizza on John’s birthday, so we shouldn’t have pizza on Mary’s birthday.
In (8a) but not in (8b) one infers that John’s birthday precedes Mary’s birthday. This must come from the presence of again in (8a). But as long as we take the appropriateness constraint imposed by again to be (9a), there is no explanation for the inference: presumably, we had pizza plenty of times in the past, so this presupposition would be met, independently of John’s birthday. Intuitively, however, one takes John’s birthday to be the relevent prior time at which we had pizza. This follows if the appropriateness condition imposed on the context is as in (9b): of a particular earlier time, the context must entail that we had pizza at that time. The presupposition of again is thus specific rather than existential: (10a). In the example, the prior time is assumed to be the time of John’s birthday (since this is the only time relevant in the context), hence the inference that John’s birthday is earlier than Mary’s. (9)
a. c ⇒ λw.∃t [t
16
(10)
Presupposition and Implicature in Compositional Semantics
a. that we have pizza at t & t
This argument amounts to the claim that the content of the presupposition triggered by again depends on an anaphoric element (as pointed out by Kripke and emphasized in the semantic discussion of Kamp and Rossdeutscher (1994)). I will follow the argument and represent that element, like other implicit anaphoric elements, as a free variable, to be assigned a value by the contextually relevant variable assignment. A revised lexical entry for again is given in (11). The free variable is the first argument of again in (11) and shows up as a subscript in the syntactic structure in (12) (see also Beck (2006)). (12) gives the interpretation of the example. We try to choose a variable assignment function g such that the presupposition triggered by again is met. The only salient candidate is the time talked about in the first sentence in (8a), John’s birthday. Thus the anaphor is resolved to that − g(t )=John’s birthday.1 Note that the fact that the content of the presupposition again gives rise to depends on an anaphoric element makes it necessary to give again that variable as another argument. This was not necessary in the old version (6). The analysis in (11)–(12) assumes that again’s first argument is represented in the syntactic structure. (11)
[[again]] (t ) (p) (t) (w)
= 1 if p(t)(w) & p(t )(w) & t
(12)
[[ [[[we have pizza] againt ] t1 ] ]]g (w) is defined only if we have pizza in w at g(t ) and g(t ) < g(t1). If defined, it is true iff we have pizza in w at g(t1).
On this analysis, notice, there is no relevant difference between the variable called t in (11) and natural language variables like pronouns, traces and quantifier domain restrictions. They all show up in the syntactic structure that is the input to compositional interpretation, and should have similar properties. We will exploit this to provide an analysis of examples like (3) from the introduction. For completeness, I will mention an aspect of the interpretation of again that has attracted much attention in the literature. Sentences with again often exhibit the so-called repetitive/restitutive ambiguity illustrated in (13).
Quantifier Dependent Presuppositions 17
(13)
Otto opened the door again.
(13 )
a. (13) presupposes that Otto had opened the door before. Then, it is true iff Otto opened the door. = Otto opened the door, and he had done that before. (repetitive) b. (13) presupposes that the door had been open before. Then, it is true iff Otto opened the door. = Otto opened the door, and the door had been open before. (restitutive)
(14)
a. Jack opened his lips to speak, but shut them again, [. . .] (Jack and Jill, by Luisa May Alcott; from the Project Gutenberg offical homesite http://promo.net/pg/) b. ‘shut them again’ is appropriate because there is a previous time at which Jack’s lips were shut (namely the beginning of the opening).
I will assume a structural analysis of this ambiguity in the style of von Stechow (1996) (see Stechow (1996), Fabricius-Hansen (2001), Jäger and Blutner (2003), Beck (2005), among others, for discussion). Application of the structural theory to our example in (13) is illustrated informally in (15)–(17). Crucial is the decomposition of the transitive verb open into the adjective open and an empty verb expressing a change of state. (15)
a. Otto opened the door. b. underlying structure: [VP Otto [ ØV [SC openAdj [the door]]]] c. surface structure: [VP Otto [ [Ø + openAdj ]V [SC t [the door]]]]
(16)
a. [[ØV ]] = λpλxλtλw.x does something in w at t that causes that p comes to be in w at t b. [[SC]] = λtλw.the door is open in w at t
(17)
λtλw.Otto does something in w at t that causes that the door comes to be open in w at t
Decomposition creates two possible adjunction sites for the adverb again, above and below the empty verb: (18a,b). The two structures lead to two different interpretations, paraphrased in (19a,b), which correspond to the repetitive and the restitutive reading respectively. The
18
Presupposition and Implicature in Compositional Semantics
ambiguity is thus analysed as a scope ambiguity between again and the empty verb. (18)
a. b.
[VP [VP Otto [ ØV [SC openAdj [the door]]]] again] [VP Otto [ ØV [SC [SC openAdj [the door]] again]]]
(19)
a. Once more, Otto did something that caused the door to become open. b. Otto did something that caused the door to become once more open.
Under this analysis, the semantics of again remains the same on the restitutive reading as on the repetitive reading, and whatever effects we find regarding the presupposition should show up on both readings. 2.2
Anaphoricity
Again on both the repetitive and restitutive reading should be recognizably anaphoric, given the analysis above. That is, we should always be able to identify the relevant prior time of an occurance of again in discourse. It has been observed (Fabricius-Hansen (2001), Kamp and Rossdeutscher (1994), Klein (2001), and following them Beck (2006)) that this is indeed a property of again. Below are some of Fabricius-Hansen’s examples as well as some that I myself collected that illustrate this. I italicise the predicate that again combines with as well as the preceding predicate that causes again’s presupposition to be satisfied. The relevant prior time that the presupposition triggered by again refers to would be the time interval associated with the first predicate. The context provides an appropriate salient antecedent for again’s first argument. (20)
Fabricius-Hansen (2001): a. Er konnte sich an nichts erinnern, nicht einmal daran, dass ich ihn gebeten hatte, sich diese drei Dinge zu merken. Ich wiederholte den Test und liess ihn diesmal die Bezeichnungen der drei Gegenstaende aufschreiben; wieder hatte er sie vergessen [. . .] He remembered none of them or indeed that I had even asked him to remember. I repeated the test, this time getting him to write down the names of the three objects; again he forgot [. . .] b. Elias hoerte nicht bloss, er sah das Toenen. Sah, wie sich die Luft unaufhoerlich verdichtete und wieder dehnte. Elias not only heard the sounds, he also saw them. He saw the air incessantly contracting and expanding [again; SB].
Quantifier Dependent Presuppositions 19
(21)
Project Gutenberg – repetitive uses: a. She rather suspected it to be so [. . .] from his listening to her so attentively while she sang to them; and when the visit was returned by the Middletons’ dining at the cottage, the fact was ascertained by his listening to her again. (Sense and Sensibility, Jane Austen) b. [. . .] the old lady shook the bottle with the air of a general who had routed the foe before and meant to do it again. ( Jack & Jill, Louisa May Alcott)
(22)
Project Gutenberg – restitutive uses: a. As she said this, she sunk into a reverie for a few moments; – but rousing herself again, [. . .] (Sense and Sensibility, by Jane Austen) b. And away she went; but returning again in a moment, [. . .] (Sense and Sensibility, by Jane Austen)
I should point out that while Fabricius-Hansen (2001) also calls the presupposition of again anaphoric, there is a difference in what we mean by this. According to her, the presupposition of again is anaphoric in the way that all presuppositions are anaphoric – in the sense of van der Sandt (1992). This is not what is meant here. I do not regard the presupposition triggered by again itself as an anaphor. For me, it is an admittance condition. But its content depends on an anaphoric element – i.e. what presupposition precisely is triggered depends on the context. This is not a property of all presuppositions. The content of the presupposition triggered by the definite article, for example, does not (necessarily) depend on an anaphor. 2.3
Discourse inferences
The data we have seen show that anaphoric presuppositions play an important role for discourse coherence. They guide us in organizing the information provided by individual sentences into a coherent whole. This becomes most obvious when they lead to inferences that would not otherwise have been justified. The pizza example repeated below illustrates this. (23)
We will have pizza on John’s birthday, so we shouldn’t have pizza again on Mary’s birthday.
20
Presupposition and Implicature in Compositional Semantics
The observation is that (23) supports the inference (24a). Why does it do that? The context to which the second sentence of (23) is added entails of a time t that we had pizza at t ((24b)). We take t to be the antecedent of again’s first argument. The use of again then imposes on the context the requirement (24c). The context does not entail that t
a. b. c. d.
John’s birthday precedes Mary’s c ⇒ that we have pizza at t c ⇒ that we have pizza at t & t
The pizza example is not unique in supporting such inferences, and is therefore not unique as an illustration of the anaphoric nature of the presupposition that again gives rise to. Both of the following examples support inferences that are partial presupposition accommodations, and both times the key to what is accommodated is the anaphoric first argument of again. In neither of the examples would this particular inference arise if again were not anaphoric, i.e. did not make reference to a particular time. Kamp and Rossdeutscher (1994) call such effects ‘presupposition justification’. (25)
A group from Jehova’s witnesses was on my doorstep at 1pm. An hour later, they were there AGAIN.
(25 )
a. (25) ⇒ they left in between 1pm and 2pm b. c ⇒ that a group from Jehova’s witnesses was on my doorstep at t c. c ⇒ that a group from Jehova’s witnesses was on my doorstep at t and t
(26)
In January 2000, we were skiing in Grafton. In January 2001, we were skiing in Vermont AGAIN.
(26 )
a. (26) ⇒ Grafton is in Vermont b. c ⇒ that t
Quantifier Dependent Presuppositions 21
d. accommodate: we were in Vermont at t hence: Grafton is in Vermont When there are no additional inferences, the role of the anaphoric presupposition seems to be to confirm an interpretation arrived at on independent grounds. The example below illustrates that. The relevant aspects of the interpretation of (27) are given in (27 a). The temporal order of the events can be derived from the use of the expression ‘return the visit’. The use of again confirms that this is the temporal organization of the events described (cf. (27 b)). (27)
She rather suspected it to be so [. . .] from his listening to her so attentively while she sang to them; and when the visit was returned by the Middletons’ dining at the cottage, the fact was ascertained by his listening to her again. (Sense and Sensibility, Jane Austen)
(27 )
a. that he listened to her at t1 that he listened to her at t2 & t1
Compare Fabricius-Hansen (2001), Klein (2001), van der Sandt and Huitink (2003), Saeboe (2004) and Beck (2006) for further comments on various aspects of the behaviour of again in discourse. We will assume the analysis in (11), in which again combines with a free variable first that determines the content of its presupposition. That analysis has received further support in this and the previous subsection. 2.4
Bound occurences of again’s first argument?
All the examples we have considered so far were cases in which the variable that is the first argument of again occurs free. Can it also occur bound? Perhaps. The following example may be a case in point: the temporal quantifier ‘every time’ would be the binder (I omit the world parameter in these examples for simplicity). Given the general facts of presupposition projection in the scope of a quantifier illustrated by example (29), the analysis makes the prediction that the presupposition of the sentence with again is as in (28b). (28)
a. Last semester, Bill was sick on Tuesday three times. Every time he was sick again two days later.
22
Presupposition and Implicature in Compositional Semantics
b. ∀t[ t is one of the three Tuesdays → [[again]] (t) (λt . Bill was sick at t )(t + 2 days)] Assertion: ∀t[ t is one of the three Tuesdays → Bill was sick at t + 2 days] PSP: ∀t[ t is one of the three Tuesdays → Bill was sick at t & t
a. Every player brought her TV Derendingen warm-up t-shirt. b. PSP: Every player had a TV Derendingen warm-up t-shirt.
Some support for such an analysis of (28) comes from the variant in (28 a), which licenses the inference in (28 b). (28 )
a. Last semester, Bill was sick on Tuesday three times. Every time he was sick again on the day of the colloquium. b. Every time, the (relevant) colloquium was after Tuesday. c. PSP: ∀t[ t is one of the three Tuesdays → Bill was sick at t and t < the (relevant) colloquium]
In general, I find it hard to identify conclusively a bound reading as opposed to a more indirectly quantifier dependent reading, to which we turn in the next section. Note that if again’s first argument can be bound, it behaves like pronouns and traces; if it cannot be bound, it is, as far as I know, like a quantifier’s resource domain variable in that respect. I have no particular expectation either way.
3
Quantifier dependent again
3.1 E-type pronouns, complex traces, quantifier domain restrictions To my knowledge, the first proposal that natural language variables can have a hidden internal complexity concerns pronouns; a relevant early reference is Cooper (1979). A Cooper-style analysis for the example in (30) is given in (30 ) (example and details taken from Heim and Kratzer (1998)). (30)
a. Every host bought just one bottle of wine and served it with dessert. b. For every host x: x bought just one bottle of wine and x served the bottle that x had bought with dessert.
Quantifier Dependent Presuppositions 23
(30 )
a. For every host x: x bought just one bottle of wine and x served [the [R x]] with dessert b. "it" = [ the [R x]] R : = λu.λv.u had bought v and v is a bottle of wine
Chierchia (1993) applies a similar strategy to traces in so-called functional wh-questions. An example and its analysis is given in (31)–(31 ). (31)
a. Who does every Italian love? b. example answer (functional reading): his greengrocer
(31 )
a. [whoi [ every Italiank loves ti,k ]] b. which f: for every Italian x, x loves f(x)? c. ti,k = [ t
> pro > ]] [[ ti,k ]]g = g(ti) (g( prok)) = f(x)
We are already familiar with example (1)–(1 ) from the introduction, in which the variable – the domain restriction of the quantifier – is itself covert. It seems to be a general properties of the expressions that appear function as variables in natural language that they permit an analysis in which they consist of a free variable part combining with a part that can be bound. (1)
a. Only one class was so bad that no student passed. b. Only one class was so bad that no student in that class passed.
(1 )
a. [[noC ]] (A)(B) = 1 iff (g(C) ∩A) ∩B = { } b. [ [only one class] [lx [ tx was so bad that [IP [DP nof(x) student ] [ passed]]]] c. only one class x: x was so bad that nof(x) student passed d. f: x → {y: y is in x}
Marti (2003) provides arguments to the effect that such covert variables, while invisible, are nonetheless syntactically real. This supports an analysis that assumes additional internal structure including a bound variable as in (1 ). The view that emerges from this discussion is that quite generally, natural language variables can be more complex than appearance would indicate, and that that complexity is reflected in their syntactic structure.
24
Presupposition and Implicature in Compositional Semantics
3.2 Again and quantifiers We are now ready to consider the interaction of again and quantifiers. We established above that the presupposition of again contains an anaphoric variable, and also showed that variables in natural language can be quantifier dependent. We expect that again’s anaphoric variable should be able to depend on a quantifier as well, and this is going to be our view of example (3) from the introduction and related data. We begin with a slightly simpler example, (32) below. (32)
In 1995, 1996 and 1998, Bill was sick on Labour Day. In each of these years, he was sick again on Thanksgiving.
Intuitively, the earlier time that again makes reference to in (32) is the time of illness on the relevant Labour Day. Thus there is not one earlier time, but three. In other words, the anaphoric variable in the semantics of again depends on the quantifier ‘in each of these years’. In contrast to example (28) on the analysis discussed above, the earlier time is not itself in the domain of the quantifier, but depends on it indirectly. I propose the analysis in (33) for the slightly simplified version (33a). (33)
a. Every year, Bill was sick again on Thanksgiving that year. b. ∀t[ year(t) & C(t) → [[again]] (f(t)) (λt .Bill was sick at t )(Thanksgiving in t)] c. f: t → the Labour Day in t C = {1995, 1996, 1998} d. ∀t[ year(t) & C(t) → Bill was sick on Thanksgiving in t & @(Bill was sick on Labour Day in t & (Labour Day in t < Thanksgiving in t))] e. ∀t[ year(t) & C(t) → Bill was sick on Labour Day in t & (Labour Day in t < Thanksgiving in t)]
Given the general facts of presupposition projection in the scope of a quantifier already mentioned, the analysis makes the prediction that the presupposition of the sentence with again is as in (33e). This seems right, and the presupposition can be assumed to be met in (32) by virtue of the first sentence plus world knowledge. (34) would be the presupposition of (33a) using the semantic analysis of again in (11) but without a functional analysis of the anaphoric variable. This amounts to the claim that Bill was sick at a time preceding 1995 - which strikes me as counterintuitive. (35) would be the result of using the original semantics of again in (6), with the predicted presupposition (35b) from (35a). This looks perhaps
Quantifier Dependent Presuppositions 25
a little better, even if it is still weaker than (33e). But of course it uses an analysis of again that we had retired for independent reasons already. (34)
∀t[ year(t) & C(t) → Bill was sick at t & t
(35)
a. ∀t[ year(t) & C(t) → Bill was sick at t & @(∃t [t
I think that the analysis in (33) is the most intuitive view of what happens in (32). The case for functional anaphoric variables in again’s presupposition can be strengthened when we consider examples that require partial presupposition accommodation (like the pizza example), hence give rise to additional inferences. I have constructed a quantificational version of the pizza example below. The sentence intuitively supports the inference in (36b). It is analysed in (37). Following the same steps as in the previous example, we arrive at (37d) as the presupposition of (37a). The context establishes that we had pizza at Johns birthday every year, but it is not asserted that John’s birthday preceded the department picnic. In the same way as in the non-quantificational version, we assume that this part of the presupposition is in fact true. The assumption surfaces as an inference. (36)
a. Every year, we had pizza on John’s birthday, and decided that we shouldn’t have pizza again at the department picnic. b. Every year, John’s B-day preceded the department picnic
(37)
a. ∀t[we have pizza at John’s B-day in t & we decide that shouldn’t [again(f(t)) (λt . we have pizza at t )(the picnic in t)]] b. f: t → John’s B-day in t c. ∀t[we have pizza at John’s B-day in t & we decide that shouldn’t [we have pizza at the picnic in t] & @(we have pizza at John’s B-day in t & John’s B-day in t < the picnic in t)] d. PSP: ∀t[we have pizza at John’s B-day in t & John’s B-day in t
Suppose that we had used a standard, simple first argument of again. The predicted presupposition is then (38): that a particular time t precedes all annual picnics, and we had pizza at t . That would be John’s birthday
26
Presupposition and Implicature in Compositional Semantics
in the first of the years talked about. The presupposition is not suitable to support the intuitive inference. (38)
∀t[we have pizza at t & t < the picnic in t]
Similarly, the existential analysis of the presupposition of again (given below) is not suitable to support that inference – it is too weak. Thus we can conclude that we have found further support for an analysis of the presupposition of again as involving an anaphoric element. The support comes from the additional, quantificational data themselves as well as from the general fact that our hidden variable in the semantics of again behaves like other hidden variables. (39)
a. ∀t[we have pizza at John’s B-day in t & we decide that shouldn’t [we have pizza at the picnic in t] & @(∃t [we have pizza at t & t < the picnic in t)]] b. PSP: ∀t[∃t [we have pizza at t & t < the picnic in t)]]
Example (3) from the introduction is of course just the pizza example in disguise. I add example (40) as a plausible case of quantifier dependent restitutitve again. The verb forget is decomposed into BECOME + ‘not know’. Once more the presupposition in (41e) seems more appropriate to me than the one in (42). (40)
a. She has told me her name three times. b. Every time / mostly /. . . I forgot it again five minutes later.
(41)
a. ∀t[t is one of the three times → I forgot it again at t+5min] b. ∀t[t is one of the three times → BECOME (t+5) ([[again]] (f(t)) (λt .I do not know it at t ))] c. f: t → t−1 (the time immediately preceding t) d. ∀t[t is one of the three times → BECOME (t+5) (λt .I do not know it at t ) & @(I do not know it at t−1)] e. PSP: ∀t[t is one of the three times → I do not know it at t−1]
(42)
∀t[t is one of the three times → I do not know it at t & t
Quantifier Dependent Presuppositions 27
The phenomenon discussed here is natural enough; an informal search via Google turned up the following two instances: (43)
From the diary of a cat owner:
Heute haben unsere Schätzchen Besuch gehabt. Wir haben Nachbarskater Thommy herein gelassen. Erstes Problem war, dass Merlin in der Zeit, bis Thommy sich bequemt hat, reinzuhüpfen, schon 3x nach draußen geflitzt ist. Otto hat ihn dann jedes Mal wieder geschnappt und hineingetragen. Today our darlings had a visitor. We let in the neighbours’ cat Thommy. The first problem was that until Thommy agreed to hop in, Merlin had run out three times. Otto cought him again every time and carried him inside. http://www.janko.at/WM/Tagebuch2004_09.htm (44)
Description of a grassroots group party event:
Höhepunkt war für viele das Modell der Castor-Strecke von Lüneburg nach Dannenberg. Die mehr als sechs Meter lange Modellbahnstrecke, mit unendlicher Liebe zum Detail gebaut, hatte alles, worauf man sich schon mal wieder vorbereiten sollte: Jede Menge bunte protestierende Leute, viel Polizei, ein rollender Castorzug, ein als Hochsicherheitstrakt ausgebauter Verladekran, [. . .] und als Höhepunkt eine Brücke die rechtzeitig vor dem Zug zusammenbricht. Das machte sie während des Festes 471-mal. Der Zug musste jedes Mal wieder zurückfahren. For many the highlight was the model of Castor’s train line from Lüneburg to Dannenberg. The model railway was longer than six meters, constructed with endless love for detail, and had everything one ought to be prepared for: loads of colorful protesting people, a lot of police, a rolling Castor train, a high security crane, [. . .] and as the highlight a bridge that collapses in time in front of the train. It did that 471 times during the party. The train had to go back again every time. http://www.graswurzel.net/281/castor.shtml An anonymous reviewer points out that these examples might involve bound first arguments of again rather than complex ones. I think the uncertainty stems from the fact that the restriction of the relevant quantifier (‘every time’) is largely pragmatic and there may be more than one plausible candidate for what is quantified over; this makes it difficult
28
Presupposition and Implicature in Compositional Semantics
to decide whether the ‘earlier’ time depends on the quantifier directly (through binding) or indirectly (via internal complexitiy). Data like (3), (32) and (36) are clearer in that respect. A final remark: I think that there are data that receive a natural analysis in terms of again’s hidden variable being dependent on a nominal quantifier. A candidate is given below (let the notation tx stand informally for the time at which x is told x’s password). I leave the details of the analysis for another occasion. (45)
a. The principal called us into the study and told us our individual password. b. Everyone forgot it again immediately/within the hour.
(45 )
a. ∀x[BECOME (tx +1hour) (again f(x) (λt.x does not know x’s password at t))] b. f: x → the time just before x was told x’s password c. PSP: ∀x[x does not know x’s password at f(x)]
3.3
Some further instances: also, stop
The adverb again is not the only expression in natural language that has been suggested to trigger a presupposition dependent on an anaphoric element. The literature (once more Soames (1989) following Kripke, and Heim (1990); see also Beaver (2001)) points out that a similar argument can be made for too/also and stop. Let’s look at too/also. Standardly, it is assumed that (46a) (in which Laura associates with too) has the presupposition in (46b). According to our understanding of presupposition, this means (46c). A semantics for too/also in this spirit is given in (46d), where I combine directly with the associate for simplicity (instead of using a focus semantics). (46)
a. b. c. d.
LAURA was at the conference, too. ∃x[x =Laura and x was at the conference] c ⇒ λw.∃x[x =Laura and x was at the conference in w] [[also]] (y)(P)(t)(w) is defined only if ∃x[x =y and P(x)(t)(w). If defined, it is true iff P(y)(t)(w).
This analysis runs into a problem with the contrast in (47a) vs. (47b). It is not clear why (47b) should be bad, since the first sentence states that the conference was well-attended.
Quantifier Dependent Presuppositions 29
(47)
a. Peter was at the well-attended recent conference in Mount Kisco. LAURA was at the conference, too. b. # Peter wasn’t at the well-attended recent conference in Mount Kisco. LAURA was at the conference, too.
Intuitively, the problem in (47b) is Peter. We just stated that Peter was NOT at the conference, and Peter seems to be what too wants to relate to. This can be captured once we recognize that the presupposition triggered by too/also is specific rather than existential, as described in (48). The context must provide a particular person of which it implies that that person was at the conference. The only person specifically mentioned is Peter, but Peter was not at the conference. The presupposition is thus met in (47a) but not in (47b), and (47b) is correctly predicted to be infelicitous. Thus we revise our (simplified) lexical entry for too/also to (48c), in which it first combines with a hidden anaphor.3 The analysis of the example is given in (49). The same point is usually made on the basis of the infelicity of examples like (50) out of the blue. (48)
a. ∃x[x =Laura and c ⇒ λw.x was at the conference in w] b. that x was at the conference & x =Laura c. [[also]] (x)(y)(P)(t)(w) is defined only if x=y and P(x)(t)(w). If defined, it is true iff P(y)(t)(w).
(49)
[[ LAURA t1 be at the conference, too(x) ]]g (w) is defined only if g(x) =Laura and g(x) was at the conference in w at g(t1). Then, it is true iff Laura was at the conference in w at g(t1).
(50)
Sam is having dinner in New York tonight, too.
The similarity of again to too/also is particularly clear when we consider examples like (51). Assume that too associates with with ‘in the fall of 1997’. (51)
a. In the fall of 1997, the conference was in Mount Kisco again. b. In the fall of 1997, the conference was in Mount Kisco, too.
The only difference between the two sentences is that (51a) requires a context that establishes for some other time than fall 1997 that the
30
Presupposition and Implicature in Compositional Semantics
conference had been in Mount Kisco, whereas in (51b) the other time has to be earlier than fall 1997. So in the semantics of again, we have ti
a. Every boy invited his girlfriend. b. Every boy also invited LUISE / HIS CLIMBING PARTNER. c. inference: Luise/his climbing partner is not the girlfriend of any of the boys
(52 )
a. ∀x: also (f(x)) (Luise) (λz. x invited z) b. f: x→ x’s girl friend c. ∀x: x invited Luise & @(x’s girlfriend =Luise & x invited x’s girlfriend) d. PSP: ∀x: x’s girlfriend =Luise & x invited x’s girlfriend
A final example for an anaphoric presupposition I will mention is the one triggered by stop. Consider (53), once more from Heim (1990). (53)
John is cooking now. He will stop cooking when tomorrow’s basketball game starts.
The discourse suggests that John will be engaged in a continuous cooking event from now until tomorrow. If ‘John stopped cooking’ simply presupposed that John had cooked at some immediately preceding time interval, we should not have to interpret the discourse in such an implausible way. Once more, the presupposition is anaphoric in that it refers back to a salient time interval in the discourse, and that is the cooking John is engaged in now. We must suppose that this time interval extends until the basketball game. (53 ) is a quantificational version. (53 )
During the World Championship last year, John would cook the evening before an important match.
Quantifier Dependent Presuppositions 31
Every time / mostly / . . . , he stopped cooking when the match began the next day. We see that there is a class of presuppositions that share with again the property of depending on an anaphoric element, as well as the empirical consequences this property has.
4
Conclusions and consequences
We have investigated a set of data in which the content of the presupposition triggered by again depends on a quantifier. I have proposed an analysis for these data in which the presupposition that again introduces is determined by an anaphoric element. That element is represented in the syntax and makes the first argument of the adverb again. It can be free, (perhaps) bound, or internally complex, like other natural language variables. Thus the quantifier dependent examples receive an analysis parallel to other examples in which the content of a variable depends on a quantifier, for instance quantifier domain restrictions. The quantfier dependent data lend further support to the analysis of again’s presupposition containing a free variable. A similar argument can be made for other anaphoric presuppositions.
Acknowledgements I am grateful to the audiences to whom various parts of this work were presented, and to Irene Heim and Luisa Marti for their help with anaphoric presuppositions and complex variables respectively. Many thanks to an anonymous reviewer for two important comments on the prefinal version.
Notes 1. This can be made more explicit in a dynamic framework in which contexts are sets of world-assignment pairs, but I will not do so here. 2. We tend to take the time immediately preceding the topic time t1 to be a time at which the proposition expressed by the sentence with again is not true. See e.g. Stechow (1996) for discussion. This effect is ignored in our formal lexical entry for again, but would have to be used in a complete account of (25). 3. I have put the non-identity requirement for the associate and the antecedent in the semantics of also. See Beaver (2001) for an argument that this may not be the right place.
32
Presupposition and Implicature in Compositional Semantics
References Beaver, David (2001). Presupposition and Assertion in Dynamic Semantics. Stanford, CSLI. Beck, Sigrid (2005). ‘There and Back Again: A Semantic Analysis’. Journal of Semantics 22, 3–51. Beck, Sigrid (2006). ‘Focus on Again’. Linguistics and Philosophy 29, 277–314. Chierchia, Gennaro (1993). ‘Questions with Quantifiers’. Natural Language Semantics 1, 181–234. Cooper, Robin (1979). ‘The Interpretation of Pronouns’. In: F. Heny & H.S. Schnelle (eds): Selections from the third Groningen Round Table (= Syntax and Semantics 10). New York, Academic Press, 61–92. Fabricius-Hansen, Cathrine (2001). ‘Wi(e)der and Again(st)’. In: Caroline Fery and Wolfgang Sternefeld (eds): Audiatur Vox Sapientiae: A Festschrift for Arnim von Stechow. Akademie Verlag, Berlin, 101–30. von Fintel, Kai (1994). ‘Restrictions on Quantifier Domains’. PhD dissertation, University of Massachusetts Amherst. von Fintel, Kai (2003). ‘Pragmatics; Notes on Presupposition’. Ms, MIT. Heim, Irene (1990). ‘Presupposition Projection’. In: R. van der Sandt (ed.): Presupposition, Lexical Meaning and Discourse Processes. Workshop Reader, University of Nijmegen. Heim, Irene (1991). ‘Artikel und Definitheit’. In: Stechow, Arnim von & Dieter Wunderlich (eds): Handbook of Semantics. De Gruyter. Heim, Irene & Angelika Kratzer (1998). Semantics in Generative Grammar. Blackwell. Jäger, Gerhard & Reinhard Blutner (2003). ‘Competition and Interpretation: The German Adverb wieder (‘again’)’. In: E. Lang, C. Maienborn & C. FabriciusHansen (eds), Modifying Adjuncts. Mouton de Gruyter, 393–416. Kadmon, Nirit (2001). Formal Pragmatics: Semantics, Pragmatics, Presupposition and Focus. Blackwell. Kamp, Hans and Antje Rossdeutscher (1994). ‘DRS-construction and Lexicallydriven Inference’. Theoretical Linguistics 20: 165–235. Klein, Wolfgang (2001). ‘Time and again’. In: Caroline Fery and Wolfgang Sternefeld (eds): Audiatur Vox Sapientiae: A Festschrift for Arnim von Stechow. Akademie Verlag, Berlin, 267–86. Marti, Luisa (2003). ‘Contextual Variables’. PhD dissertation, University of Connecticut. Rooth, Mats (1992). ‘A Theory of Focus Interpretation’. Natural Language Semantics 1, 75–116. Saeboe, Kjell Johan (2004). ‘Conversational Contrast and Conventional Parallel: Topic Implicatures and Additive Presuppositions’. Journal of Semantics 21: 199–217. van der Sandt, Rob (1992). ‘Presupposition Projection as Anaphora Resolution’. Journal of Semantics 9, 333–77. van der Sandt, Rob & Janneke Huitink (2003). Again. Proceedings of the 2003 Amsterdam Colloquium. Soames, Scott (1989). ‘Presupposition’. In: D. Gabbay & F. Günthner (eds): Handbook of Philosophical Logic. Kluwer, Dordrecht.
Quantifier Dependent Presuppositions 33 Stalnaker, Robert (1973). ‘Presuppositions’. Journal of Philosophical Logic 2, 447–57. Stechow, Arnim von (1996). ‘The Different Readings of Wieder “Again”: A Structural Account’. Journal of Semantics 13, 87–138. Westerstahl, D. (1984). ‘Some Results on Quantifiers’. Notre Dame Journal of Formal Logic 25, 152–70. Project Gutenberg official homesite: http://promo.net/pg
3 Licensing or Regine Eckardt University of Göttingen
This chapter addresses the so-called free choice effects of disjunction under existential quantification. The effect has received most attention in modal contexts, specifically as free choice permission. Consequently, a broad range of analyses trace the reason for free choice effects at its interaction with modality. I will argue that free choice effects arise with any kind of existential quantification, and that an analysis should hence not essentially rest on modality. I propose that a Boolean analysis of or, in concert with specific interpretation of existentials, is sufficient to derive the observed effects. It is moreover proposed that the reinterpretation of existential quantifiers is made necessary by a principle of or-licensing which is the second cornerstone of the analysis.
1
Introduction
Our logic classes taught us that the word or denotes Boolean disjunction ∨. However, looking at uses of or in everyday life, it seems that the word rarely ever exhibits its pure meaning. The problem seems to be that disjunctive statements A or B contain the simpler parts A and B. Why did the speaker fail to make one of the simpler statements and chose the longer, and less informative disjunction instead? A survey of the data leads to the following cases: In downward monotone contexts, A or B statements are indeed more informative than simpler statements. In such sentences, or can unproblematically denote Boolean disjunction. In certain other uses, the disjunctive statement may be longer than A alone, or B alone, and less informative than A and B, but gives rise to extra implicatures and hence leads to motivated utterances. We could label such uses as ‘Boolean ∨ plus pragmatic inferences’. 34
Licensing or 35
In a last kind of cases, however, or occurs in a sense that is hard to explain as the result of Boolean or plus pragmatic inferences. These are the so-called free choice use of or in the scope of existential modals. The following example illustrates the case.
(1.1) (1.2)
Judy may take coffee or juice. Judy may take coffee, and Judy may take juice.
The puzzle is this. The sentences in (1.1) and (1.2) seem intuitively equivalent. However, a simple Boolean analysis of or, together with a simple possible-world analysis of may as existential quantification over deontically accessible worlds (‘worlds that show what Judy is allowed to do’), fails to predict this intuition. The failure rests on the well-known fact that, in predicate logic, formula (1.3) does not imply (1.4). Kamp (1973) is the classical reference which brought the free choice puzzle on the agenda of formal semantics.
(1.3) (1.4)
∃x( A(x) ∨ B(x) ) ∃x( A(x) ) ∧ ∃x( B(x) )
In recent years, several strategies to solve this puzzle have been explored. As the effect arises predominantly in modal contexts, some authors attribute it to a revised interpretation of modal existentials (plus, possibly, further pragmatic effects). This line was followed, for example, by Schulz (2002), Aloni (2003a,b), or Simons (2005a,b). The more radical kind of rescue, however, consists in the claim that the word or did not denote Boolean disjunction in the first place. Zimmermann (2000) develops an analysis of or as conjoined epistemic alternatives and derives free choice effects in modal contexts from this analysis. The more recent Geurts (2005), while deviating from Zimmermann’s proposal in detail, can be seen as a conservative extension of this kind of approach. The authors leave it somewhat open whether they would allow for a reading of or as Boolean disjunction in addition to their analyses. However, Zimmermann seems to defend a maximal position in that the chapter lists several examples where his denotation for or is problematic; a second use in a Boolean sense is never his proposed solution. I will therefore at some points refer to his approach as the most consequently non-Boolean analysis.
36
Presupposition and Implicature in Compositional Semantics
Apart from free choice examples, any analysis of or should be able to cope with the following range of cases. (1.5)
a. Nobody was sick or on holiday. b. Nobody was sick, and nobody was on holiday.
Sentence (1.5) illustrates the use of or in a downward entailing context (here: the scope of negation). Whatever the exact delimitations of contexts may be, elementary predicate logic tells us that the disjunction (a) logically entails the conjunction in (b.). Semantic intuition suggests that Boolean ∨ is a fully adequate denotation of or in these examples. (1.6) (1.7)
Joe is drunk or sick. (As far as I can see,) Joe might be drunk, and Joe might be sick (and I don’t know which)
In positive contexts, epistemic uncertainty appears to be the standard reason to use or. If no other reason can be thought of, the hearer will assume that the speaker makes an unspecific assertion because he does not know any better. (1.8)
Everybody ordered a beer or a pizza.
This sentence shows or in an upwards monotone context. If or is taken to denote Boolean disjunction, then (1.8) is true if everyone ordered a beer (and no pizza), if everybody ordered a pizza (and no beer), or in mixed cases. In fact, our semantic competence tells us that (1.8) should only be used in mixed cases (unless the speaker has imperfect knowledge). For instance, it can felicitously be uttered by the waitress who has just taken the corresponding order. We can assume that she has perfect knowledge about the order; e.g. having noted it as a list. In such a situation, she does not seem to express a list of epistemic possibilities, neither globally (‘possibly everybody ordered a beer, and possibly everybody ordered a pizza, and I don’t know which’) nor locally (‘for each customer, it is possible that he ordered beer, and its possible that he ordered pizza’). In these cases, however, a simple Gricean argument can account for this fact. If the speaker knew that everybody in fact ordered a beer, he could have said so and been briefer and more informative. Similarly for a homogeneous order for pizzas. Hence, the mixed cases are the only ones that warrant possible utterances. (Matters change, of course, as soon as or is justified by the
Licensing or 37
speaker’s imperfect knowledge.) Analogous examples arise in necessity statements (i.e. for modal universal quantification). The cases listed so far (including free choice examples) are the common stock in the literature. The last two types of examples, to my knowledge, have not yet received extensive attention. (1.9)
The whole school was posing in the yard to welcome the queen. Everybody was dressed in nice clothes for the great day. a. Some children waved little flags or threw flowers. b. Some children waved little flags. And some children threw flowers.
The sentence in (1.9.a) seems to be uttered by the speaker with a certain group of children in mind. He might vaguely intend to talk about those children who took part in cheerful activities. No epistemic uncertainty is expressed, and we understand that the disjunction is used because among the active children, there were some who waved and some who threw flowers. The conjunction in (b.) follows. Note that (1.9) is the analogue of the classical free choice constellation in the domain of nominal quantification. Klinedinst (2005) is another source where such examples were discussed.1 I sympathize with his findings, which, however, need to be sharpened to do justice to the data. Finally, consider (1.10). (1.10)
What new privileges does Judy gain when she’s 18 years old? # Judy may drive a car, or marry without her parents’ consent or vote in elections.
The striking fact about (1.10) is that it does not offer an instance of the free choice pattern. Specifically, the disjunct does not imply the conjoined list of privileges of those over-age in (1.11). (1.11)
Judy may drive a car, and Judy may marry without her parents’ consent, and Judy may vote in elections.
Free choice or essentially occurs in just those cases where the disjunction lists deontic alternatives with respect to one specific decision. (1.10) in a free choice interpretation sounds as if the three things depended on each other, somehow as if Judy’s driving a car would exclude her voting in elections. I will come back to the no list of privileges example (1.10) at several places. Specifically, I will use it to explore the nature of deontic alternatives that are addressed by true free choice sentences (Section 4).
38
Presupposition and Implicature in Compositional Semantics
Note that Zimmermann (2000) predicts that it should be a natural free choice example and imply (1.11) (Section 2). This chapter is organized as follows:2 In Section 2, I will recapitulate uses of Boolean and ‘ill-behaved’ or and offer a brief review of the most recent literature. It will be concluded that or requires licensing in most contexts, and that licensing or can require a non-standard use of other parts of the sentence (specifically existential quantifiers). In Section 3, I will elaborate my analysis of examples like (1.9). I propose that these examples rest on quasi-specific existential quantification. The hearer will understand that the speaker has a certain group of individuals in mind, and in order to be truthful has to use disjunction. Section 4 shows how this analysis can be carried over to the modal case, i.e. free choice examples of the type in (1.1). Section 5 discusses some examples where various kinds of licensing or interact. It highlights the program underlying the present chapter (as many others before): The word or can show unexpected effects in various ways. It is preferable to assume one uniform semantic contribution, namely Boolean disjunction, which interacts with other parts of the sentence in different ways which, in turn, lead to different side messages.
2
Boolean and mis-behaved ‘or’
Boolean disjunction is potentially problematic in terms of pragmatics. Disjunction offers us a means to denote a weaker property by using a more complex expression. This constellation, in many contexts, posits a contradiction to the combined M- and I-principle (Levinson 2000) or classically, Grice’s maxims of manner and quantity. Such conflicts arise already in the simplest possible disjunctions, the disjunction of two positive atomic sentences. (2.1) offers an example. (2.1)
Joe is drunk or sick
Instead of one of two possible, more specific statements (‘Joe is drunk’, ‘Joe is sick’), the speaker chooses to attribute the broader property of ‘being drunk or sick’ to Joe. Of course, it is not principally problematic to utter sentences with general properties. A sentence like ‘Joe is German’ is pragmatically unproblematic, even though the speaker might know that Joe is in fact from Bavaria. The speaker did not mention Bavaria, and if the context does not render the German counties relevant, the sentence Joe is German does not implicate Joe is not from Bavaria or The speaker does not know whether Joe is from Bavaria. Disjunctions, however,
Licensing or 39
are different. In using a disjunction, the speaker necessarily has to mention two properties which are usually more specific. These properties are presented as salient and relevant. The simpler sentences are salient alternative utterances in context. The hearer hence will look for a reason why the speaker chose a more complex expression in order to give less information. Going back to example (2.1), it is a textbook case in pragmatics to infer that (2.1) implicates that (2.2). (2.2)
Speaker does not know that Joe is drunk, and speaker does not know that Joe is sick.
Let me call a surrounding sentence in which a disjunction may be embedded a sentence frame. Then we can posit a general requirement on uses of or: A sentence frame does not pragmatically license disjunction if the simpler sentence(s) entail the frame plus disjunction . (2.3)
unlicensed disjunction: (A) → (A ∨ B)
The use of disjunction is pragmatically licensed if the simple sentence does not entail disjunction (2.4.a), and specifically if sentence frame with disjunction entails the simple sentence (2.4.b). Moreover, or is licensed if there are implicatures such that disjunction plus its implicatures are not entailed (2.4.c). (2.4)
a. licensed disjunction: (A) −/→ (A ∨ B) b. licensed disjunction: (A ∨ B) → (A) c. licensed disjunction: (A) −/→ (A ∨ B) ∧ with implicature .
It can be seen easily that disjunctions in downward entailing contexts instantiate the (b.) case. Negation as well as the antecedent of a conditional, or the restrictor of universal quantifiers, are good pragmatic licensing contexts for Boolean disjunction. (2.5) (2.6)
If you get an A or a B in the exam, I will take you out for dinner. (Either condition will be sufficient) Nobody was bored or annoyed. Gordon is the youngest man who ever climbed the Chimborazzo or the K2. Gordon eats neither meat nor fish.
40
Presupposition and Implicature in Compositional Semantics
Note that the last kind of use under negation must have been perceived so genuinely a useful or pattern in English that the original construction not either . . . or together with negative concord n- was lexicalized as neither . . . nor. I therefore take the neither nor construction as a particularly clear use of or under negation. A disjunction in the antecedent of a conditional is likewise useful because it allows the speaker to express two conditionals in one. Let us now turn to some cases where a sentence plus a pragmatic inference can license the use of a disjunction. We already saw (1.8) above, but more cases show that the scope of universal quantifiers uniformly give rise to implicatures. Consider the sentence in (2.7). (2.7)
(In order to get a credit,) You have to write a paper or kill a rabbit.
Simple Boolean disjunction under universal modal quantification will lead to a sentence meaning which holds true • if in fact, I intend to hand out a credit for a paper, and under no other circumstances • if in fact, I intend to hand out a credit for a dead rabbit, and under no other circumstances • if in fact, I intend to hand out a credit exactly if one or the other requirement is fulfilled. Our everyday understanding of (2.7) suggests that it only should be true in the third case. Neither case should be mentioned spuriously (even though practical reasoning would strongly suggest the first scenario!). Like in (1.8), we can argue that the author of the announcement, having full control over matters, should have used a simpler and more informative sentence in the first and second scenario. Similar observations hold for or in the scope of proportional quantifiers with a strict ratio (e.g. ‘exactly half’). All these contexts allow the use of or plus implicatures. None of these cases needs to involve epistemic uncertainty. Once again, epistemic uncertainty could be the reason for the speaker to utter (2.7) – for instance for a student who can not remember the announcement very well. But the point is that epistemic uncertainty need not be understood. A third class of sentences require one further step in order to allow for a motivated use of or. They are such that in their literal interpretation, the use of or would not be licensed (unless expressing epistemic uncertainty). However, such sentences can be improved by resorting
Licensing or 41
to re-interpretations or secondary readings. (2.8) offers an example. Imagine the following report by the class teacher after class: (2.8)
(Today’s math class was just a disaster. I was not able to inspire anyone for the beauty of the subject. The kids just did not react.) Some pupils were sleeping or daydreaming.
If all parts of (2.8) are understood in their normal, common way, we get an existential statement that is weaker than the simpler statements without disjunction. (2.9.a) is entailed by b. and c. (I use capital variables for plural objects. Predicates are assumed to apply to pluralities in a distributive manner. The details are given in the next section.) (2.9)
a. ∃X( PUPIL(X) ∧ [ SLEEP(X) ∨ DAYDREAM(X) ] ) b. ∃X( PUPIL(X) ∧ [SLEEP(X)] ) c. ∃X( PUPIL(X) ∧ [DAYDREAM(X)] )
Hence we would expect that (2.8) is pragmatically marked. One way to remedy this markedness could consist in an implicature about epistemic uncertainty. In this reading, the speaker is not sure whether the pupils were asleep or awake, daydreaming. However, under normal circumstances the sentence intuitively seems a shorter way to express (2.10). (2.10)
(My class today was just a disaster.) Some pupils were sleeping, and some pupils were daydreaming.
This stronger statement seems to come about in two subsequent steps. First, the indefinite ‘some’ is not interpreted as a mere existential quantification. The speaker appears to have a specific group of pupils X in mind that she wants to talk about. Using ‘some pupils’ in a specific sense, she can refer to this group. Secondly, given that the referent is fixed, the speaker appears to have reason not to use one of the simpler sentences ‘some pupils were sleeping’ or ‘some pupils were daydreaming’. The hearer will hence infer that among the pupils that the speaker intends to talk about, some were sleeping and not daydreaming, and some were daydreaming and not sleeping. Sentence (2.10) is entailed by the overall information conveyed by (2.8), but crucially, the groups of ‘some pupils’ in (2.10) are not the larger group that the speaker was talking about in (2.8). Note that the pattern corresponds to the free choice inference.
42
Presupposition and Implicature in Compositional Semantics
The free choice puzzle has received renewed attention in the last years, starting with the analysis by Zimmermann (2000), followed by Geurts (2005), Aloni (2003b), Schulz (2002), Simons (2005a,b) and Klinedinst (2005). A very detailed and insightful discussion of earlier treatments, including Kamp (1973) and (1979), is given in Schulz (2002). The majority of papers rest on a modified analysis for or. Zimmermann 2000 (and Geurts, elaborating on Zimmermann 2000) claims that or does not contribute logical disjunction, but presents a series of epistemic alternatives which the speaker conjoinedly entertains. The paradigm case of this pattern is shown by sentences like (2.11) which comes down to the statement in (2.12): (2.11) (2.12)
Joe is drunk or sick. (As far as I can see,) Joe might be drunk, and Joe might be sick (and I don’t know which)
Zimmermann assumes that (a) all alternatives mentioned as a disjunct have non-empty intersection with the epistemic background of the speaker; that (b) there is no subset of epistemic alternatives of the speaker which supports none of the alternatives mentioned, and (c) there is an implicature about ‘independence’ of the alternatives mentioned. Assumption (a) is designed to derive genuineness: no disjunct is uttered spuriously. Assumption (b) predicts exhaustivity under normal circumstances. No life option should remain unmentioned. Condition (c) is motivated by the observation that the hearer usually seems to understand that the alternatives are mutually exclusive in some sense. The examples discussed in this chapter rest on logical independence, but the author states that more general (topological) notions of non-overlap are operant in the general case. Free choice effects for epistemic might follow straightforwardly from the analysis. In order to generalize the analysis to deontic modality, Zimmermann proposes the authority principle. If the speaker can be assumed to be an authority in the issue in question, epistemic possibility (‘maybe’) can be strengthened to epistemic certainty (‘in fact’). In spite of its initial plausibility, Zimmermann (2000) has problems in those cases where or does not show modal flavour. Among the problematic cases are those without modal component (e.g. (2.8) and (2.6)), but problems arise also for deontic must as well as the no-list-of-privileges case. Consider once more the announcement in (2.7). If we interpret (2.7) as
Licensing or 43
a list of epistemic possibilities, we get something like (2.13)
Possibly, you have to write a paper, and possibly, you have to kill a rabbit. ♦((KILL-A-RABBIT) ) ∧ ♦((WRITE-A-PAPER))
Given that the author of the requirement has full knowledge about it, we can apply the authority principle in order to go beyond the ‘ignorant student’ reading of (2.7). The authority principle leads to (2.14)
(KILL-A-RABBIT) ∧ (WRITE-A-PAPER)
This states that the requirement for a credit consists of two parts: killing a rabbit and writing the paper. This does not match the natural understanding of (2.7). Zimmermann’s prediction for example (1.10), repeated here as (2.15), is likewise problematic. Remember that the sentence was not an acceptable way to state the conjunction in (2.16). (2.15)
(2.16)
(What new privileges does Judy gain when she’s 18 years old?) # Judy may drive a car, or marry without her parents’ consent or vote in elections. Judy may drive a car, and Judy may marry without her parents’ consent, and Judy may vote in elections.
The sentence in (2.15) will be assigned the meaning in (2.17). Following Zimmermann, I use for deontic possibility here. (2.17)
♦((DRIVE-JANE-CAR)) ∧ ♦((MARRY-JANE)) ∧ ♦((VOTE-JANE))
This list of epistemic possibilities is subject to Zimmermann’s requirements to lists, (a)–(c). In particular, logical incompatibility (c) could, but need not necessarily, be required for (DRIVE–JANE–CAR), (MARRY–JANE) and (VOTE–JANE). In actual practice, however, no form of incompatibility (or non-overlap) is intended in deontic free choice examples. This chapter proposes to apply the authority principle to (2.17). If the speaker has full knowledge about Jane’s privileges, he will not only hold all three alternatives as possible, but as actual facts. Specifically, they can hence not be mutually exclusive. (2.18)
(DRIVE–JANE–CAR) ∧ (MARRY–JANE) ∧ (VOTE–JANE)
44
Presupposition and Implicature in Compositional Semantics
(2.18) however expresses exactly the list of privileges that Jane (as well as any 18 year old person) will enjoy. The analysis, far from explaining that (2.15) is unacceptable, will predict that it is a perfect statement of lists of privileges. Schulz (2003) argues in favour of a pragmatic basis for the modal approach in Zimmermann (2000). Going back to Gazdar’s seminal work on implicature, Schulz assumes a systematic set of possible implicatures of disjunctions under modal operators which are evaluated in context. The remaining implicatures are then fed into a highly developed modal logic framework and give rise to the desired entailments. Schulz’ analysis moreover extends to obligation sentences and puts a wide range of data into a coherent common framework. While the present proposal shares Schulz’ perspective of a modular semantic-pragmatic setup, Schulz once again offers an analysis which focusses on modal quantification and is inherently designed to cover free choice effects in those, and only those examples. Schulz’ division of labour into pragmatics and semantics will avoid modal readings for those sentences where Boolean disjunction plus Gricean inferencing is sufficient (see (1.8) Everybody ordered beer or pizza), but the setup is unsuited to capture free choice effects in nonmodal contexts. The following example replicates the effects shown in (2.8) and (1.9). (2.19)
(Gordon is not as strict a vegetarian as he wants to make you believe. I have observed him for quite a while and. . .) Gordon sometimes DID eat meat or fish. → Gordon sometimes ate meat. And Gordon sometimes ate fish.
Two recent proposals by Simons (2005a,b) rest on a modified analysis for or but do not posit a modal component. The core idea of both proposals lies in the assumption that the disjuncts in the sentence should be collected in a set which is then available for further semantic computation. In Simons (2005b), the set of disjuncts is moreover exploited for a scoping mechanism in that, at an appropriate place, set union will be performed, reflecting the semantic contribution of or. Simons’ proposal so far rests on a classical boolean view of or which can not explain the free choice puzzle. At this point, both accounts stipulate semantic requirements on the set of disjuncts which yield the desired entailments. Simons (2005a) discusses the requirement of being a supercover, while Simons (2005b) resorts to the weaker symmetry condition. Let me concentrate on this more recent proposal. The symmetry condition states that there should be some
Licensing or 45
salient property P which is shared by all disjuncts. Simon illustrates the idea on basis of examples like (2.20). (2.20)
Judy may take coffee or juice.
The condition refers to the set of disjuncts { Judy take coffee, Judy take juice} and requires one salient common property for all these disjuncts. In absence of any common property which is entailed logically at this point, Simons stipulates that the requirement could be ‘that both propositions have nonempty intersection with the deontically accessible worlds’. With this extra requirement, the free choice entailment follows. While the symmetry condition offers the correct results, it remains a convenient mystery that hearers always resort to the right kind of common property. Simons explicitely refrains from discussing the motivation underlying symmetry. The following kind of abuse can therefore not be excluded so far: (2.20 )
Judy may take coffee or juice. Salient common property: Judy take coffee and Judy take juice are both desirable options for Judy. (They have nonempty intersection with Judy’s buletic alternatives.) Fact: In fact, Judy is only allowed to take juice.
(2.20’) is predicted to be acceptable by Simons’ symmetry condition. In fact, it is not a good situation for the sentence to be true. I suspect that a pragmatic motivation of the symmetry condition would reveal it to summarize the pragmatic inferences that will be at the heart of the approach in the present chapter. One important aspect in Simons (2005b) consists in that she acknowledges the existence of pure Boolean disjunction in downward entailing contexts. She quotes the observation that or with modals, if negated, is absolutely well-behaved. I will evaluate her solution in section five, where interacting licensing factors are discussed. A recent proposal which is very close in spirit to the present chapter was made by Klinedinst (2005). He is the only author, as far as I can see, who has ever pointed out free choice effects in nonmodal contexts. Consequently, any analysis of these effects should not rest on modality in a way which can not be transferred to ordinary existential quantification. It is this criticism that stands, in Klinedinst’s as well as my opinion, against all approaches that were reviewed so far (as well as related ones), in spite of their sophistication and adequate predictions in the modal domain. While Klinedinst’s examples, as well as (1.8), (1.9), are essential in the
46
Presupposition and Implicature in Compositional Semantics
understanding of free choice effects, he does not explore explicit existential quantification in sufficient detail and therefore fails to understand free choice effects as the consequence of semantic re-interpretations plus pragmatic inferencing. He does not discuss specificity effects, and related effects in the modal domain which lead to the no-list-of-privileges facts. The present approach to the free choice puzzle is very conservative in its analysis of or and modality. I propose that or is Boolean disjunction, and that existential statements are interpreted as specific under suitable circumstances. If a disjoint property is attributed to a specific set of objects, we can derive an implicature that all disjuncts must be exhibited by some elements of this set. The implicature arises by simple allusion to the Gricean maxims of manner (‘be brief – comparing the salient alternative utterances’) and quality (‘do not say what you believe to be false’). The analysis will receive initial support from data with explicit existential quantification. They give us clear access to the discourse pragmatics of the effects under investigation. We can then proceed to the generalization to the modal case.
3
Explicit existential quantification and ‘or’
In the present section, I will investigate free choice effects with explicit nominal or adverbial existential quantification in more detail. I will use the term ‘explicit existentials’ for these quantifiers, because unlike modal existentials, they allow us to specify restrictions on the domain of quantification, and to investigate its status in discourse. Explicit existentials offer a test case for free choice effects where the interacting interpretive and pragmatic mechanisms are easier to access than in the modal case. After revisiting some phenomena, I will offer a semantic analysis for free choice effects for explicit existentials which will be applied to modal existentials in Section four. Let me start by illustrating the free choice effects for explicit existentials with some more examples. All examples are given with an appropriate context. The ties between free choice existentials and the discourse context will be discussed below. (3.1)
(3.2)
(Our class visited the Zoo yesterday. After watching animals in the morning, we took a break near the kiosk at noon.) Some pupils had chips or ice cream. (Others went to the playground.) (Discussing whether our friend Gordon is a vegetarian or not: No, I don’t think that he is. We visited a summer school together, and I remember clearly that . . . ) Gordon sometimes DID eat meat or fish.
Licensing or 47
The conjoined statements in (3.3 ) and (3.4) are intuitively entailed by the respective disjunctions. (3.3) (3.4)
Some pupils had chips, and some pupils had ice cream. Gordon sometimes ate meat, and Gordon sometimes ate fish.
The examples hence exhibit an entailment pattern for existential quantification over individuals and times which is analogous to free choice effects. Some other, more subtle intuitions could be explicated as follows: The speaker wants to talk about some specific people, things or occasions. The specific persons, things or occasions that she wants to talk about exemplify a general property which is exemplified by the disjuncts. In the pupil-example, the speaker could announce her intentions as follows: ‘Among the ways that children can amuse themselves in a zoo, let me now name those which have to do with food.’ In the Gordon-example: ‘Among the occasions of Gordon eating, let me talk about those which violate his strict vows as a vegetarian.’ Let me now try to delineate the pragmatics of free choice effects with explicit existentials. I characterized them as ‘specific’ so far. Several observations suggest that, in fact, a partitive use of the existential seems crucial in order to elicit free choice effects. This is evidenced by several kinds of examples where the respective word order patterns do not support partitive readings very well, or where the context definitively excludes such a reading. In order to come to a richer spectrum of data, I will consider both English and German data in this section. The free choice effects for German explicit existentials exactly mirror the English effects, but due to greater freedom in word order, we can look at a broader spectrum of data. As a first example, let us passive sentences. Examples like (3.5) do not easily give rise to free choice effects. The resulting sentences have a very odd ring to them. It seems extremely hard to pinpoint the pragmatic side messages, and hence the intended overall information conveyed by (3.5). Here, as well as below, we see sentences with unlicensed and hard-to-license uses of disjunctions. There is as yet no technical term for the kind of interpretive puzzlement that befalls the hearer in view of sentences like (3.5), (3.7), or (3.8). (3.5)
a. Beer or pizza was ordered by some pupils. b. Bier oder Pizza wurde von manchen Schülern bestellt.
Whatever (3.5) may mean, we do not understand a free choice effect easily. If we compare (3.5) with other, good partitive uses of existentials,
48
Presupposition and Implicature in Compositional Semantics
we observe that the quantifier does not precede the property that is attributed to the NP referents. In common partitive uses of quantifiers, however, the stressed quantifier should precede the property in question (see e.g. Jäger, 1996, Eckardt 1998). We can hence conlude that free choice effects arise only when the explicit existential precedes the disjunctive property. More minimal pairs which illustrate word order effects are those in (3.6) which rest on verb pairs like get/bring, and (3.7) in German where changes in word order can turn an odd disjunction into a good disjunction. (3.6) (3.7)
(3.8)
√
Some children get their presents from the Christkindl or Santa Claus. # The Christkindl or Santa Claus bring the presents to some children. √ Manche Kinder bekommen die Geschenke some kids get the presents vom Christkind oder dem Nikolaus from-the Christkind or the Santa Claus. # Vom Christkind oder dem Nikolaus bekommen from-the Christkind or the Santa Claus get manche Kinder die Geschenke. some children the presents √ Manchen Kindern bringt das Christkind oder der Nikolaus die Geschenke. to-some children brings the Christkind or the Santa Claus the presents # Das Christkind oder der Nikolaus bringt manchen Kindern die Geschenke. The Christkind or the Santa Claus brings to-some children the presents.
Note that ‘specific use’ of an explicit existential does not exclude generic uses. The following example seems to be talking about kinds of guests rather than a specific group of persons present. The example could elaborate a remark like ‘you know how hard it is to run a pizzeria which meets all customers’ ideas of Italian food’. (3.9)
There are some guests who want pizza or beer. There are OTHER guests who want bruscetti or saltimbocca.
What seems to be crucial is that the disjuncts are understood as representative for a more overarching property, one which the speaker does not want to use explicitely (‘underclass food’ in contrast to ‘refined food’). In other cases, the property might not even have a good name, like in the Gordon-example in (3.2). ‘exhibiting non-vegetarian eating behaviour’
Licensing or 49
seems an extremely clumsy paraphrase for the property of ‘eating meat or fish’. The above example morevoer shows that resentational sentences with there is do not impede free choice effects. However, if we take (3.9) as an utterance in a different kind of situation (and with a different kind of message in the speaker’s mind), we can test that the partitive use of some is necessary in order to derive free choice effects. In the scenario in (3.10) where a partitive use is definitively excluded, or will be understood as epistemically licensed: ‘and I don’t know which’: (3.10)
The owner of the local pizza hut is shaken awake by his wife at 3 am: ‘Get up, man! There are some guests at the door who want pizza or beer!’ (undertone: I didn’t quite understand which.)
In sum, a partitive use of the explicit existential seems crucial for free choice effects. Sentences with word orders which impede partitive use, as well as utterances in situations where partitive uses are blocked do not give rise to free choice effects. Free choice effects for explicit existentials can be captured by the following, simple kind of analysis. Let me stress that the exact nature of specificity in (i) might require further investigation. (i) partitive/specific use of an explicit existential [[ Some N ]] = P . ∃X( [[N]](X) ∧ P(X) ) Presupposition: The hearer understands that the speaker has an intended instantiation A for the existential X in mind. A is a true subset of a group B of known objects with property [[N]]: A ⊂ B. Notational comment: I assume that properties P are defined for atomic and plural objects. Predication over plural objects (for non-collective properties) is defined as in (ii). (ii) plural predication: A property P holds for some given plurality X, P(X) iff for all x ≤ X: P(x) (iii) Pragmatics of or in plural predication over group of objects A: 1. Speaker uses (P or Q)(A) and has sufficient knowledge of the situation. 2. Speaker did not use the simpler P(A), nor the simpler Q(A), and hence violated the maxim of manner: be brief. 3. He must have reason to do so. The reason could be the maxim of quality: The speaker seems to believe that neither P(A) nor Q(A) are true.
50
Presupposition and Implicature in Compositional Semantics
4. Hence, speaker knows that for some x1 , x2 ≤ A : P(x1 ) ∧ ¬Q(x1 ), ¬P(x2 ) ∧ Q(x2 ) The present analysis of free choice effects is essentially pragmatic. Before moving on to the modal case, I will briefly review some related data that offer evidence about the semantics–pragmatics interface. There is an ongoing debate about the interaction of pragmatic inferencing with semantic composition. Specifically, work like Chierchia (2004) or Sauerland (2004, forthcoming) raise the possibility that the pragmatic evaluation of subconstituents might precede the completion of semantic composition. Explicit existentials with or show interesting effects in the antecedent of conditionals. Consider the sentence pattern in (3.11). (3.11)
If any/some N do A or B, then S.
It offers the opportunity to study the semantics–pragmatics interface. Specifically, if free choice effects arise in the antecedent of a conditional, the present account will require pragmatic inferencing before semantic composition has been completed. If, however, free choice effects do not arise in this constellation, this would lend support to the traditional sequential picture of semantic composition preceding pragmatic inferencing. The actual data are intricate. In the bulk of examples, it seems impossible to understand free choice effects with nominal existential quantifiers plus disjunction in the if -clause. Consider the sentence in (3.12). (3.12)
If any pupils take drugs or steal jewellery, then the teacher will be fired.
The sentence intuitively seems equivalent to the conjunction in (3.13). This equivalence is predicted by classical logic. (3.13)
If any pupils take drugs then the teacher will be fired, and if any pupils steal jewellery, then the teacher will be fired.
Note that or is in a downward-entailing context in (3.12). We would therefore expect that the sentence (3.12) has a reading which rests on Boolean disjunction, and shows no further implicatures. According to my intuition, this expectation is borne out. In the present example, it is very implausible to understand that only the conjoined vices will get
Licensing or 51
the teacher sacked, while he may stay if only one or the other kind of criminal activity has been reported. (3.14)
If some pupils take drugs, and some pupils steal jewellery, then (and only then) the teacher will be fired.3
If we take sentence (3.12) as a typical exemplar of the scheme (3.11), we will conclude that antecedents of conditionals block free choice effects. This supports a pragmatic analysis of the effect, plus the traditional view of a strictily sequential modus operandi of semantics and pragmatics. This observation is challenged by similar modal examples like in (3.15), discussed in Zimmermann (2000). He points out that free choice effects can arise in (3.15).4 (3.15)
If Mr X might be in Chelsea or Hyde Park, then we can as well give up. understood: Neither possibility alone would be reason to give up, but the joined possibilities set Mr X in such a strong stragetic position as to be invincible.
At this point, we could either conclude that the analysis of free choice effects with modals should be substantially different from the one for explicit existentials (contra the aim of my chapter). Or we could resort to some unnamed difference between the pragmatics for modals in contrast to explicit existentials (not very attractive). Or we can look for more examples with explicit existentials in if-clauses. Indeed, it seems that with a suitable preceding context, we can reinforce free choice effects. The sentence in (3.16) plausibly is a free choice example, as wittnessed in (3.17) and (3.18). (3.16) (3.17) (3.18)
If Gordon sometimes drinks beer or wine, we could offer him a good bottle of Bordeaux as a present. If Gordon sometimes drinks wine, then we could offer him a good bottle of Bordeaux. Not implied: If Gordon sometimes drinks beer, then we could offer him a good bottle of Bordeaux.
We seem to read ‘Gordon sometimes drinks beer or wine’ as ‘Gordon appreciates mild alcohol abuse’. The implication certainly does not distribute over the disjuncts in the antecedent. Beer drinking is not a good indication that someone will appreciate expensive Bordeaux wines. With
52
Presupposition and Implicature in Compositional Semantics
a bit of goodwill, the same effect can even be achieved with nominal quantifiers. (3.19)
If some visitors started talking or fell asleep during the concert, then the performance must have been awful.
Again, it is plausible to understand the disjunction in (3.19) as an exemplaric list of the property of ‘not concentrating on the presentation’. If this is in fact the speaker’s point, he would insist that neither sleeping nor wispering audience alone is sufficient to indicate a poor concert. Yet, if several indicators coincide, certain conclusions about the quality of the concert are legitimate. The data are puzzling so far. Free choice effects seem to arise in some, but not all embedded uses in if -clauses. Does this mean that pragmatic inferencing can sometimes, but not always, take place during semantic composition? A closer look at the examples in question might help to resolve the paradox. To my intuition, the conditionals above (including the Zimmermann examples) are not the law-like uses of conditionals like in ‘if it rains, then the street gets wet’. The conditionals in question seem to take up an asserted fact of previous utterances (‘It seems that Mr X might be in Chelsea or Hyde Park’). In such a plain assertion, plain Boolean disjunction would not be licensed and pragmatic strengthening has to take place in order to motivate the use of or. If I am right in this intuition, then the antecedent takes up a fact that was estabished in discourse, and the consequent names the conclusion the speaker wants to draw. A case without disjunction is given in (3.20). (3.20)
A: I do not feel well today. B: Too bad. If you don’t feel well, you won’t want to try my homebaked cream cake I guess.
This kind of conditionals can be used in discourse in order to present an assertion of the speaker as a natural consequence of previous assertions. What is important here is that the content of the if -clause is taken up anaphorically from the discourse. Let me therefore call these conditionals anaphoric conditionals. Upon closer inspection, it turns out that free choice effects in the antecedent of a conditional can be systematically provoked in anaphoric conditionals. If that is true, the role of pragmatics is quite different than in an all-new utterance. While the details of a copying mechanism of literal content plus implicatures are still highly unexplored, we may expect
Licensing or 53
that the free choice effects are copied effects from an antecedent root sentence rather than pragmatic inferencing dove-tailed with semantic composition. The analysis so far suggests that pragmatics indeed follows semantic evaluation. If no appropriate antecedent sentence (plus implicatures) is available, free choice effects in the antecedent of conditionals are very hard to understand. This holds true at least for free choice effects for explicit existentials. Let me now turn to the modal case.
4
Modal existentials
In this section, I will pursue the question whether the analysis for free choice effects for nominal and adverbial quantifiers can be generalized to the case of modal existential quantification. Let me briefly repeat the essential characteristics of the former: • The existential quantifier was used partitively, refering to a subset of a set of objects that is salient in context. • The existential quantifier was used in a specific sense: The hearer will understand that the speaker has a certain set of referents in mind. • The disjoint property is loosely understood as the case-by-case exemplification of a more general property.5 Evidently, these characteristics can not be directly translated into characteristics of modal existentials. First, there is no clear notion of partitive modal quantification. It is unclear (to me) how one would present a set of worlds as salient, and which modal statements could convey meanings like: ‘In some of these worlds, Jane takes coffee, but there are others left.’ Likewise, it is very difficult to verify whether the speaker had some specific set of worlds in mind. We can not point at worlds and ask: did you mean this one? or that one? or yet another one? The last characteristic will turn out to be easiest to rediscover in the modal domain. In summary, it seems likely that a transfer of the analysis in Section 3 will have to rely on slightly different versions of these characteristics of free choice modal existentials. I will first discuss deontic modals (Section 4.1) and then turn to epistemic possibility (in 4.2). Epistemic possibility will be closely compared to Zimmermann 2000 where this is taken to be the basic case. 4.1
Deontic possibility
Deontic may, in the simplest case, conveys an existential statement about the set of worlds which are deontic alternatives for some subject (as seen from the real world w∗ ). Semantic modelling usually assumes that the
54
Presupposition and Implicature in Compositional Semantics
existential statement states the existence of one world at least. A sentence like (4.1) states that there is a deontic alternative where Jane takes coffee. (4.1)
Jane may take coffee.
However, as Klinedinst (2005) points out, it would be more realistic to assume that such a statement asserts the existence of a plurality of worlds of a certain kind. He argues that every permission sentence leaves many things unmentioned that Jane may do or not do, as well as irrelevant side aspects. Hence, there is not only one but many deontic alternatives w (for Jane at w∗ ) which exemplify her taking coffee. The semantic analysis of modal existentials does not lose its strength if we allow quantification over pluralities of worlds: there are some deontic alternatives for Jane where Jane takes coffee. Next, let me discuss which part of a person’s deontic alternatives is described with a free choice disjunction. Two scenarios come to mind. In some cases, the preceding context makes it clear that the full space of deontic alternatives is covered. The subject has several options but has to chose one of them. (4.2) offers an example. (4.2)
a. You must choose a password. b. You may choose your old one, or a new one.
Such examples are very close in meaning to imperatives like you must choose your old one or a new one. The fact that each disjunct denotes a genuine alternative is derived like in Section 2 above, and the pertinent conjunction follows. We will come back to such cases below. Usually, however, permission sentences are true permissions in that the subject also has the option to do nothing. (4.3)
Jane may take tea or coffee.
This sentence, intuitively, has a threefold message. First, Jane has permission to take tea and she has permission to take coffee. Second, it is her choice which one to take. And third, she may as well choose not to take either one. In free choice permission sentences, • the speaker decides to talk specifically about all and only those deontic alternatives which exemplify one among several possible actions that the subject may take in one specific choice • the possible actions are losely understood as excluding each other (which is, however, not strictly necessary)
Licensing or 55
• logical independence is not sufficient for mutual exclusion; the possible actions must be tied together as being all possible outcomes of one decision of the subject. The last observation is important. Remember that disjoint lists of mutually independent privileges do not give rise to free choice effects. I repeat the crucial example (1.10) for convenience. (1.10)
What new privileges does Judy gain when she’s 18 years old? # Judy may drive a car, or marry without her parents’ consent or vote in elections.
In sum, the speaker has indeed a specific subset of the set of all deontic alternatives (of the subject) in mind. The subset is the one which exemplifies all possible decisions that the subject is allowed to take in one given case. A sentence like (4.3) expresses: ‘The speaker wants to inform you about Jane’s range of deontic alternatives as far as the choice of hot restaurative beverages is concerned.’ While we will turn to more complex examples below, I will take this paraphrase as the base line for the semantic analysis of modal existential quantification if licensing for ‘or’ is required. I have no conclusive views on other ‘neutral’ cases. The following points generalise the analysis of Section 2 to deontic modals. (i) may S refers to some specific intended subset W of all deontic alternatives Wdeont for some subject a at world w∗ . It is the subset W which exemplifies all possible decisions that the subject is allowed to take in one given case.6 (ii) Plural predication over worlds: A property P holds for some given plurality W of worlds, P(W) iff for all w ≤ W: P(w).7 (iii) Pragmatics of or in plural predication: 1. Speaker uses (A or B)(W) and has sufficient knowledge of the situation. 2. Speaker did not use the simpler A(W), nor the simpler B(W), and hence violated the maxim of manner: be brief. 3. He must have reason to do so. The reason could be the maxim of quality: The speaker seems to believe that neither A(W) nor B(W) are true. 4. Hence, speaker knows that for some w1 , w2 ∈ W: A(w1 )∧¬B(w1 ), ¬A(w2 ) ∧ B(w2 )
56
Presupposition and Implicature in Compositional Semantics
If we apply this analysis to sentence (4.3), it yields the following. (4.5)
a. ∃W( [TAKE(JANE, TEA) ∨ TAKE(JANE, COFFEE) ](W) ) where speaker has a certain set W of alternatives in mind. b. . . .that is, for all w∈W: [TAKE(JANE, TEA)∨TAKE(JANE, COFFEE)](w) c. As speaker uses (A or B)(W) instead of the simpler A(W), or B(W), he implicates that the simpler statements would not hold true for the set of worlds W he has in mind. Hence, there are w1 , w2 ∈ W such that TAKE(JANE, TEA)(w1 )∧¬TAKE(JANE, COFFEE) (w1 ) and ¬TAKE(JANE, TEA)(w2 )∧TAKE(JANE, COFFEE) (w2 ). In other words: Jane may take tea. And: Jane may take coffee.
This shows that the desired free choice effects follow from the analysis of this basic case. Let us come back to those cases where may A or B is tantamount to must A or B. Geurts (2005) points out that the preceding discourse can specify the quantificational domain. I repeat the crucial example in (4.2). (4.2)
a. You must choose a password. b. You may choose your old one, or a new one.
The first sentence characterizes the deontic alternatives in total. In this context, it is completely clear that the may quantification in b. refers to the range of deontic worlds described in a. In this case, an analysis as in (i)–(iii) is fully convincing, as the set of alternatives that the speaker must have in mind is the set of all those where the addressee adheres to the obligation in a. This totality of worlds W is characterized by the disjunctive property in (4.6), and as above the free choice conjunction in (4.7) is implicated. (4.6) (4.7)
[ CHOOSE(YOU, OLDPASSWORD) ∨ CHOOSE (YOU, NEWPASSWORD) ](W) ∃w(CHOOSE(YOU, OLDPASSWORD) (w))∧∃w(CHOOSE(YOU, NEWPASSWORD) (w))
Note that the sentence ‘You must choose your old password, or a new password.’ is logically equivalent to (4.2.b) in the given context. An explanation of the ‘friendlier’ undertone of the may statement remains to be developed. Even though the range of options that the speaker has in mind in modal existentials is more difficult to pinpoint than for explicit existentials, a closer look at the data can reveal some conventions. Let us
Licensing or 57
turn back to example (4.3). There is some tacit suggestion that the disjunct in (4.3) exhausts the options of Jane as far as hot beverages with restorive quality are concerned. The speaker restricts attention to these cases and remains tacit about cold drinks, food, and any other activity. The speaker would not be uncooperative if it turned out that Jane has more unrelated options in addition: the choice between soft drinks, the choice between wine/beer, the choice between vegetarian and nonvegetarian lunch etc. The speaker has just that set of deontic alternatives in mind that exemplify the hot beverages choice. To see this point more clearly, imagine an air flight where the stewardess approaches passengers and offers: (4.8)
You may take coffee or ham sandwich.
In this context (passenger and stewardess both informed about conventional refreshments) the utterance strongly suggests that the stewardess has run out of any other food or drink. She is cooperatively listing the remaining options and leaves for you to choose. If there were in fact more kinds of drink or food available, the passenger would be justly annoyed and have the feeling that the stewardess did not exhaust the full range of deontic alternatives that sentence (4.8) suggests she has in mind. Further examples can serve to illustrate this intuition: (4.9)
Samantha may take up her studies at Stanford or Harvard. possible intended deontic alternatives: worlds that exemplify all possible universities S. may go to worlds that exemplify all possible ivy league universities S. may go to
Hence, the utterer of (4.9) would not be uncooperative if it turned out that Samantha also has the possibility to go to Mayor Chesterton College at Sheperd’s Hill, say. In that case, the speaker would just have the second set of deontic alternatives in mind. Matters are different in (4.10). (4.10)
Frederick may take up his studies at Stanford or at Major Chesterton College at Sheperd’s Hill. possible intended deontic alternatives: worlds that exemplify all possible universities F. may go to
In (4.10) the two disjuncts that describe Frederick’s options are disparate enough not to exemplify any sub-class of universities. Hence, we’ll tend to understand the list in (4.10) as exhaustive list of Frederick’s options.
58
Presupposition and Implicature in Compositional Semantics
Let me finally briefly mention modal disjunctions like the following, as discussed in Geurts (2005): (4.11) (4.12)
You may do A, or you must do B. You must do A, or you must do B.
I think that these require a more refined analysis of deontic space and discourse. An example like the schematic (4.11) suggests that the first A option is more agreeable than the second, B option. Hence, (4.13) seems a natural instantiation for (4.11) while (4.14) is not (according to my own preferences): (4.13) (4.14)
You may do the shopping, or you must clean the toilet. You may clean the toilet, or you must do the shopping.
The pattern in (4.12), in contrast, has been disputed as illogical (e.g. Schulz 2003 as reported in Geurts 2005). How can there be, logically thinking, an alternative of two equally binding obligations? (Note that (4.12) is not understood as an epistemic alternative here ‘. . . and I don’t remember which’.) I think that the criticism is well taken, even though the examples are valid and existant. Yet, the contribution of or seems to be dynamic-temporal and beyond the range of phenomena that I want to address here. Consider a classical instantiation of (4.12): (4.15)
(4.16)
Your money, or your Life! (Geld oder Leben!) (You must pass me your money, or you must die.) Hands up, or I’ll shoot you! (You must take your hands up, or you must die) You have to take away your car now. Or you’ll have to pay a fine of 50 Euro.
As before, the options are again ordered according to the degree of unpleasantness of the requirements. But secondly, there is a clear temporal dimension in the contribution of or in such examples. You must do A. And if you have not done A in due time, B will follow. Hence, the disjunction of obligations like in (4.12) does not present an alternative of obligations now but a sequence of ever more unpleasant obligations in the future. (This is the essence of all threats for punishment). I will not consider such uses of or in more detail here.
Licensing or 59
4.2
Epistemic possibility
Let us take a natural example of epistemic free choice as our starting point. We observe our friend John late at night, pale, leaning at a lamp post. We conclude: (4.17)
John might be drunk or ill.
In cases like this, there is one clear issue, pieces of evidence or some fact and we present a list of possibilities that we hold compatible with this fact. At this point, it might be useful to recapitulate Zimmermann’s acceptability conditions on lists. Relating the listed propositions p1 , . . . pn to the epistemic alternatives of the speaker Hw,c it is assumed that i. pi ∩ Hw,c = Ø for 1 ≤ i ≤ n (each disjunct is a genuine alternative) ii. there is no set M such that M ∩ Hw,c = Ø and M ∩ pi = Ø for 1 ≤ i ≤ n. (the disjuncts together exhaust the epistemic space) iii. the disjunctions do not overlap, in the sense of a suitable topology. (independence; a violable pragmatic requirement) Condition ii. amounts to the assumption that the speaker designs the list of alternatives with the intention to exhaust his epistemic space. In order to remain realistic, Zimmermann allows for a tacit elsewhere disjunct in order to account for open ended lists of irrelevant cases. This assumption could be rephrased as follows: (A) The speaker intends to make a modal existential statement about his full epistemic space; or: that part of the epistemic space where the relevant, realistic options that he wants to list are exemplified.8 The intention in uses like (4.17) can be paraphrased as follows: ‘I want to talk about the set W of all reasonable epistemic alternatives – even though I use specific existential quantification to address them – and I attribute the property [ DRUNK(JOHN)∨ILL(JOHN) ](W) to this set of worlds.’ Epistemic alternatives which might be excluded from this statement are those remote possibilities the speaker has bad eyesight, it is in fact a lifesize doll of John that is leaning at the lamp post, etc. Note that the resulting constellation is parallel to the permission sentence in (4.5): ‘You must choose a password. You may choose your old password or a new one’. Like in that earlier case, (4.17) could be replaced by John must be drunk or ill. The quantification expressed, and the set of worlds refered
60
Presupposition and Implicature in Compositional Semantics
to, therefore do not exhibit the standard neutral case of existential quantification. However, (4.5) and (4.17) can be captured as one limit case of specific existential quantification where some are practically all. Like in Section 4.1, the genuineness of each disjunct follows by pragmatic inference. This, in turn, again entails the free choice conjunction. (4.18)
John might be drunk. And: John might be ill.
One very nice illustration of the tension between (a) the desire to exhaust all possibilities, (b) the fact that this might require an endless list, and (c) the systematic neglect of certain epistemic possibilities, is exhibited by the German children’s book Schachtelmonster (Monster in a Box). A little boy finds a cardboard box and explores the possibility that there might be some Schachtelmonsters (cardboard box monsters) in that box. The hidden premise of the reflections in (4.19) is ‘assume that there are monsters in that box’ (i.e. let’s devise this list of disjuncts with that set of epistemic alternatives in mind where there are actually monsters in the box). (4.19)
Vielleicht sind sie gelb mit blauen Tupfen, oder blau mit gelben, perhaps are they yellow with blue spots or blue with yellow oder vielleicht sind es Mumienmonster, oder ein paar or perhaps they are mummy-monsters or a pair Fledermausmonster, oder ganz viele . . . bat-monsters or a-whole many ‘Perhaps they are yellow with blue spots, or blue with yellow spots, or perhaps they are mummy monsters, or some bat monsters, or a terrible lot of them . . .’
The possibilities listed here do not cover the whole epistemic space of the little boy (because, luckily, there is still the possibility that there are no monsters in the box at all). The disjuncts listed do not seem to cover the full range of possible monsters either. Who can think of all possible monsters there could be? The epistemic alternatives rather seem to address the set of worlds which exemplify types of monsters the little boy can think of. Finally, further explorations will have to settle the extent to which or in epistemic statements can sometimes share the dynamic quality of the disjunctions at the end of Section 4.1. It might be plausible to assume that or in an open list refers dynamically to sequences of epistemic states of the speaker: Or, if the previous possibilities I could think
Licensing or 61
of are not true, this new set of possibilities is another option. I will leave such uses of contemplative or untreated here. At the beginning of the section, I suggested that modal free choice examples arise most easily in contexts where the speaker addresses one clear issue, pieces of evidence or fact. He lists, so to speak, all resulting possibilities that are compatible with a certain kind of evidence, fact, observation etc. (It is perhaps not an accident that Zimmermann 2000 rests entirely on examples from a detective mystery game.) To what degree is such a context not only ‘natural’ but even necessary? To check this, consider an example where the speaker in fact reports unrelated possibilities. (4.20)
(On the ride to a wedding, the family discusses what the ceremony might be like.) The bride might wear a fancy white dress, or there might be children with flowers, or a band might be playing, or there might be a big cream cake.
A disjunction like this has an odd ring to it. If we understand it as disjoint possibilities – and a suitable prosodic pattern may enhance this – then the speculations seem to imply that under normal circumstances, only one of these possibilities is expected to hold. (The couple could have just enough money to buy either a dress, or flowers, or hire a band, or buy a cake.) Compare a list of possibilities as in (4.21). (4.21)
The bride might wear a fancy white dress and there might be children with flowers and a band might be playing and there might be a big cream cake.
(4.21) is a good way to speculate about a wedding in its various, unconnected aspects. In this sense, it does not follow from (4.20). Zimmermann 2000 could, of course, capture this fact by appeal to condition (iii), the mutual non-overlap of alternatives. (4.20) illustrates that non-overlap is not a requirement about logical non-redundancy (as his examples could suggest) but about mutual exclusiveness.9 In sum, the analysis of epistemic free choice examples mirrors the previous cases in the following manner. (i) might S expresses existential quantification over epistemic alternatives. The speaker has in mind a subset W of all epistemic alternatives
62
Presupposition and Implicature in Compositional Semantics
Wepist (for speaker, in context c at world w∗ ). It is the subset W which exemplifies all possible conclusions or options that the speaker can think of in view of a certain piece of evidence, facts, situations, or reflection. (ii) Plural predication over worlds: A property P holds for some given plurality W of worlds, P(W) iff for all w ≤ W: P(w). (iii) Pragmatics of or in plural predication: (a) Speaker uses (A or B)(W) and has sufficient knowledge of the situation. (b) Speaker did not use the simpler A(W), nor the simpler B(W), and hence violated the maxim of manner: be brief. (c) He must have reason to do so. The reason could be the maxim of quality: The speaker seems to believe that neither A(W) nor B(W) are true. (d) Hence, speaker knows that for some w1 , w2 ∈ W : A(w1 ) ∧ ¬B(w1 ), ¬A(w2 ) ∧ B(w2 ) The only assumption that one could find problematic, (A), has been shown to mirror analogous assumptions that are inherent in earlier work, and in that sense, the present proposal is as good or as bad as these. However, the present analysis for epistemic free choice has been developed as an instantiation of a more general pattern that arguably also covers deontic free choice examples, and extensional free choice examples. Moreover, the analysis rests on the assumption that the word or denotes Boolean disjunction, and that all other effects are due to pragmatic inferencing, if necessary enhanced by reinterpretation of other parts of the sentence (here: existential quantification). This leaves the option of devising individual analyses for different kinds of implicatures for different kinds of disjunctive sentences. The final section will briefly but non-exhaustively highlight this issue.
5
Interactions between different licensers
The present approach presents a modular semantic-pragmatic analysis of free choice effects of or. The connective or uniformly denotes boolean disjunction. The word or (hence) offers a systematic means to convey a weaker property (A or B) instead of two stronger properties (A, B). This weakening requires pragmatic licensing: why say less, in more words? In many uses, we may have simple and perspicuous reasons to use weaker
Licensing or 63
properties instead of stronger ones, e.g.: downward entailing contexts: antecedents of conditionals, restrictors of universal quantification, negation with certain implicatures: Scope of universal quantification (every guest received a rose or a hat) and necessity operators (epistemic, deontic, etc.) Other uses would, in their literal meaning, trade a weaker statement for a stronger one. In such uses, speakers/hearers can use conventional reinterpretations of other parts of the sentence in order to come to a content where or is pragmatically meaningful. This option was elaborated for explicit and modal existential quantification. Finally, if no other explicit reason to use or is given in the sentence, epistemic might serves as a default licenser, as in (5.1). It was demonstrated in the previous section how we can derive the conjoined possibilities in (5.2).
(5.1) (5.2)
John is a buddhist or a hindu ( . . . and I don’t remember which) John might be a buddhist. And John might be a hindu.
I argued that epistemic uncertainty can not be the only way to license or. Many uses of or patently lack any flavour of epistemic uncertainty or insufficient knowledge. The modular analysis of or can explain why or has very different flavours in different contexts of use. Consider the following list of examples. None of these seems to indicate uncertainties on the side of the speaker. (Of course, all examples can be interpreted as involving epistemic alternatives, but this will not be the only, nor the most natural interpretation.)
(5.3) (5.4) (5.5)
(5.6)
All applicants are buddhists or hindus (implicature: and both possibilities occur) If you get an A or a B in the exam, I’ll take you out for dinner (implies: either precondition is sufficient) You must take a written exam or present three times in class. (implicature: which describes all your possible options; either one is a life option.) Nobody was bored or annoyed. Gordon never eats meat or fish. Gordon eats neither meat nor fish.
64
Presupposition and Implicature in Compositional Semantics
All the examples in (5.3) to (5.6) have in common that a plain boolean analysis of or, together with a simple application of the maxims of manner and quality, are sufficient to describe the information content of the examples in question. If the strategy pursued in the present chapter is correct, then a simple boolean analysis of or is likewise sufficient in other cases, specifically those that give rise to free choice effects.10 The observed logical strengthenings arise, I claimed, by a reinterpretation of existential quantification rather than by a nonstandard analysis of or. This will lead to the prediction that such reinterpretations will become superfluous if other licensers are present in the sentence. I will discuss two constellations, negation and conditionals as licensers. Simons (2005b) points out that the negation of a disjoint permission behaves strangely. She states that ‘when we embed or under both a modal and negation, it stops misbehaving and starts acting like a well-behaved Boolean operator’. The following example is discussed in detail: (5.7)
Jane may not sing or dance.
Simons diagnoses the following readings of (5.7): (5.8) (5.9)
Either Jane is not permitted to sing, or Jane is not permitted to dance. Jane isn’t permitted to do either, sing or dance (‘neither nor reading’)
Simons uses a modal re-interpretation of or (close to Zimmermann (2000)) to attribute the pertinent free-choice reading to the unnegated sentence (5.10). (5.10)
Jane may sing or dance. ♦( jane sing ) ∧ ♦( jane dance )
Negation then leads to reading (5.8). (5.11)
¬(♦( jane sing ) ∧ ♦( jane dance ) ) ≡ ¬ ( jane sing ) ∨ ¬ ( jane dance )
Yet, as argued in Simons (2005b), the more salient reading in (5.9) remains unanalysable in Zimmermann (2000). Simons herself can capture the second reading because her symmetry condition, posited on the disjuncts, is satisfied in negated uses. (The shared property of both disjuncts is to have empty intersection with jane’s deontic space.) While
Licensing or 65
this technically leads to the correct results, the basis for Simons’ condition remains unanalysed, which turns it into an austere – and easily violable – stipulation. Under the present account, we’d say that a specific interpretation of existential quantification (may) is essentially driven by the need to license the use of or as a longer way to say less. An analysis of (5.7) can proceed as follows: The material in the scope of the modal operator amounts to the disjunction11 (5.12)
jane sing ∨ jane dance.
The modal applies, and we can assume that it does so under the standard existential analysis. (5.13)
♦ (jane sing ∨ jane dance)
Finally, the result is combined with negation. (5.14)
¬ ♦ (jane sing ∨ jane dance)
The disjunction in (5.14) is in a downward entailing context, and hence the resulting statement is logically stronger than the competing simpler forms. (5.15)
¬ ♦ (jane sing), ¬ ♦ (jane dance)
The use of disjunction is pragmatically licensed here because it leads to a stronger statement in a simpler way than by using two separate sentences ( Jane may not sing, and Jane may not dance). In this case, the boolean interpretation of or does not need any further interpretative or pragmatic processes to make sense. The epistemic reading (5.8) comes about in a different manner. In the present approach, the free choice effects for deontic may . . . or . . . are derived without any reference to epistemic modality. In whatever way negation, modality and conjunction distribute out, this will not bring us from deontic to epistemic modality. However, I assume that epistemic possibility is the default licensing mechanism for or. The epistemic disjunction will be derived directly, by assigning wide scope to or and licensing it epistemically. (Note that wide scope or is not in a downward entailing context.)
66
Presupposition and Implicature in Compositional Semantics
(5.16)
Possibly, Jane may not sing or Jane may not dance. ‘It is not allowed for Jane to sing, or it is not allowed for Jane to dance, and I don’t know which.’
If we analyse epistemic ‘possibly’ as shown in Section 4, we can also derive the corresponding conjunction. (5.17)
Possibly, Jane may not sing, and possibly, Jane may not dance.
Going back to the preceding accounts, note that Simons (2005b) and Zimmermann (2000) (as in Simons 2005b) derive the disjunction in (5.11) by boolean distribution of negation over two conjuncts. The connective in (5.11) (as copied from Simons (2005b)) is simple Boolean or without any further modal force. Hence, (5.17) can not be derived from (5.11). Of course, both authors could resort to a Gricean argument here, but this would be at odds with the general program pursued in both papers, namely to derive free choice effects in the use of or as part of the literal content. Note that there is a third scoping possibility for or in sentence (5.7), leading to the logical structure ¬ (♦ (jane sing) ∨ ♦(jane dance)). Under the present analysis all logical operators and quantifiers get their standard interpretation unless more is required for pragmatic reasons, which is not the case here. (Negation of two disjuncts is stronger than negation of either disjunct alone). Therefore this option is logically equivalent to (5.14). It seems a natural assumption that the scope of or determines the pragmatic repair strategies and implicatures – if necessary – which warrant its use. Specifically, epistemic possibility appears to occur only with wide scope or. Exceptions from this observation can arise in contexts where a previous wide scope epistemic modality gets copied into an embedded context from previous discourse.12 This happens in the second kind of example that I will look into, or in the antecedent of conditional clauses. Antecedents of conditionals provide a good reason to use or. Consider the conditional in (5.18). It is intended as equivalent to the conjoined conditionals in (5.19). The two sentences are indeed logically equivalent under a Boolean analysis of or. The example is taken from Zimmermann (2000:275). (5.18)
If Mr X is in Regents Park or in Bloomsbury, he cannot take a boat.
Licensing or 67
(5.19)
If Mr X is in Regents Park, he cannot take a boat, and if Mr X is in Bloomsbury, he cannot take a boat.
Consequently, there is no further need to re-interpret any part of the sentence in order to come to a statement that warrants the use of disjunction. Specifically, note that an epistemic interpretation of or, while still possible, is at best marginal in such an example. (5.20)
If Mr X is in Regents Park or in Bloomsbury, he cannot take a boat. If Mr X is in Regents Park, he cannot take a boat, or if Mr X is in Bloomsbury, he cannot take a boat, and I don’t remember which.
This observation suggests that the different ways to make sense of or are not mutually exclusive. The presence of one does not lead to a principled unavailability of the other (even though ‘innocent’ speakers of English tend to refute (5.20) on stylistic grounds). The epistemic use of or as paraphrased in (5.20) requires the following steps in the semantic evaluation: (i) Disjunction is interpreted with highest scope above the conditional, (ii) the resulting literal meaning contains a disjunction which lacks motivation, (iii) the hearer grants a modal possibly in the specific sense discussed in section three and (iv) derives the net information of literal content plus implicatures.
6
Summary
In the present chapter, I proposed a modular semantic-pragmatic analysis of free choice effects for or. The account rests on the following assumptions: • Uses of or are tendentially pragmatically unmotivated, specifically where the speaker would use more words to convey less information. The hearer will infer some reason why the speaker used or. • Uses of or can be motivated by implicatures of the literal content of the sentence, by re-interpretation of the original sentence (plus implicatures), and by tacit epistemic modality. • Existential quantification can be re-interpreted as specific existential quantification. This possibility is available in general. It is exploited in motivating uses of or. • The semantics of specific existential quantification allows to derive the free choice effects which have been observed for or. Most importantly,
68
Presupposition and Implicature in Compositional Semantics
it can explain free choice effects in all quantificational domains. Hence, the approach extends naturally to free choice effects for explicit existentials, data which have not received much attention in the literature.13 I argued that core examples like those discussed in the chapter can be captured maintaining a strict sequential order between semantic evaluation and pragmatic inferencing. It should be kept in mind, however, that more elaborate cases of embeddings could require a more refined picture. A sentence like (5.1) has no anaphoric quality, and yet free choice effects in the embedded context are possible. (5.1)
The detective believes that Mr X is in Regents Park or in Bloomsbury.
The proposed analysis hence delineates a new broad range of data which can help us to get a better understanding of the systematic interactions at the semantics-pragmatics interface.
Notes 1. The online version of his SuB 10 talk came to my attention while I was elaborating the core ideas of the chapter. 2. I want to thank an anonymous reviewer whose comments helped me to sharpen and clarify an earlier version of the chapter. 3. I am aware of the less-than arbitrary switch between any and some, but can not spot its exact relevance so far. 4. For all those who are not acquianted with the original paper, it might help to know that the sentence relies on a situation where the interlocutors are engaged in the German board game Scotland Yard in which a single player, Mr X, making hidden moves in London has to be hunted by a team of cooperating detectives. From time to time, Mr X gives clues about his movements which lead to hypotheses about his position. 5. This last observation is less easily explicable than the first two. However, the free choice effects arise most saliently when the preceding context leads the hearer to expect that this-or-that general property will be addressed next (e.g. non-vegetarian behaviour; amusing-by-buying-food and sweets; showing signs of boredom, etc.) 6. There may be more than one set of worlds that serves this purpose. This does not matter for the core of the analysis. It will be sufficient (see below) to assume that the speaker wants to talk about any one subset of deontic alternatives which is large enough to exemplify all options. The pragmatics of or would then implicate that any such subset can only be truthfully described by resorting to a disjoint property. Conjunctions of permissions will follow.
Licensing or 69 7. We could assume here that we are talking about atomic parts of the plurality of worlds, i.e. single possible worlds. Hence, a simpler notation might be w ∈ W. 8. As before, we can weaken this requirement to ‘some sufficiently large part of the epistemic space where all options are exemplified’. 9. At least under normal circumstances. The offer ‘tea or coffee?’ can always be exploited by the request ‘both, please!’ 10. Excluding the dynamic or in Section 4. 11. I follow Simons here in taking the scoping facts of negation relative to must and may for granted. Hence, may not is analysed as ‘negated permission’, while must not denotes ‘obligation not to . . . ’. 12. I do not have any independent proposals to make about scoping of or. In particular, I appreciate Simons’ scoping mechanism for or which allows scopetaking without any syntactic movement operations (dubious or not). My claim is that the scoping mechanism does not require further modifications of the semantics of or which are intended to yield free choice effects. The scoping mechanism in Simons is very strong, as it can potentially give any scope to or. The proposal that I defend here may lead to contentful restrictions to this scoping mechanism. It might turn out that a narrowly defined set of or licensers restricts the possible scopes of or. A systematic exploration remains to be conducted. 13. Klinedinst (2005) poses a welcome exception to this claim. The present chapter shares his perspective and the core analysis. However, the material which is accessible at present does not discuss specificity effects in the explicit domain, nor the interaction of several different ways to make pragmatic sense of or.
References Aloni, Maria (2003a). ‘On choice-offering imperatives’. In P. Dekker and R. van Rooij (eds): Proceedings of the Fourteenth Amsterdam Colloquium, 2003. ILLC, University of Amsterdam. Aloni, Maria (2003b). ‘Free Choice in Modal Contexts’. In M. Weisgerber (ed.): Proceedings of SuB 7. (Arbeitspapier 114, FB Sprachwissenschaft, Konstanz). University of Konstanz. URL: http://www.ub.uni-konstanz.de/kops/volltexte/ 2003/1103/ Chierchia, Gennaro (2004). ‘Scalarity Implicatures, polarity phenomena, and the syntax/pragmatics interface’. In: A. Beletti (ed.): Structures and Beyond. Oxford: Oxford University Press. Eckardt, Regine (1998). ‘Focus and nominal quantifiers’. In P. Bosch, R. van der Sandt (eds): The Focus Book. Cambridge: Cambridge University Press. Geurts, Bart (2005). ‘Entertaining Alternatives. Disjunctions as Modals’. In Natural Language Semantics 13: 383–410. Jäger, Gerhard (1996). ‘Topics in Dynamic Semantics’. PhD thesis, HumboldtUniversity Berlin, published as CIS-Bericht Nr. 96–92, Centrum für Informations- und Sprachverarbeitung LMU München. Kamp, Hans (1973). ‘Free choice permission’. In Proceedings of the Aristotelian Society, N.S. 74: 57–74.
70
Presupposition and Implicature in Compositional Semantics
Kamp, Hans (1979). ‘Semantics versus Pragmatics’. In F. Guenthner, S.J. Schmidt (eds): Formal Semantics and Pragmatics of Natural Languages, Dordrecht, Reidel: 255–87. Klinedinst, Nathan (2005). ‘Freedom from Authority’. Talk presented at SuB 10, Berlin. Kratzer, Angelika and Junko Shimoyama (2002). ‘Indeterminate Pronouns: the View from Japanese’. In Proceedings of the Third Tokyo Conference on Psycholinguistics, 1–25. Longer version at Semantics Archive: http://semanticsarchive.net/. Levinson, Stephen (2000). Presumptive Meanings. Boston: MIT Press. Sauerland, Uli (2004). ‘Scalar implicatures in complex sentences’. Linguistics and Philosophy 27: 367–91. Sauerland, Uli (forthcoming). ‘On embedded implicatures’. Journal of Cognitive Science. Schulz, Katrin (2002). ‘You may read it now or later: A case study on the paradox of free choice permission’. Masters Thesis, ILLC, Amsterdam. Simons, Mandy (2005a). ‘Dividing Things Up. The Semantics of “or” and the Modal/“or” interaction’. In Natural Language Semantics 13: 271–316. Simons, Mandy (2005b). ‘Disjunction and Symmetry’. In E. Georgala, J. Howell (eds): Proceedings of SALT 15. Zimmermann, Thomas Ede (2000). ‘Free Choice Disjunction and Epistemic Possibility’. In Natural Language Semantics 8: 255–90.
4 Free Choice and the Theory of Scalar Implicatures∗ Danny Fox MIT
This chapter will be concerned with the conjunctive interpretation of a family of disjunctive constructions. The relevant conjunctive interpretation, sometimes referred to as a ‘free choice effect,’ (FC) is attested when a disjunctive sentence is embedded under an existential modal operator. I will provide evidence that the relevant generalization extends (with some caveats) to all constructions in which a disjunctive sentence appears under the scope of an existential quantifier, as well as to seemingly unrelated constructions in which conjunction appears under the scope of negation and a universal quantifier. Alonso-Ovalle (2005), following Kratzer and Shimoyama (2002), has presented important evidence arguing that free choice effects should be derived by the system that accounts for Scalar Implicatures (SIs). However, we will see that deriving a free choice implicature is not a simple matter within standard approaches to implicature computation. More specifically, FC directly contradicts neo-Gricean attempts to deal with another observation about disjunction due to Chierchia (2004): Sauerland 2004, Spector 2006. In response to this predicament, I will argue for a system that derives SIs within the linguistic system, though in a somewhat different manner from Chierchia (2004). Specifically, I will argue for a covert exhaustivity operator with meaning somewhat akin to that of only (in the spirit of Chierchia (2004), but more directly following suggestions by Groenendijk and Stokhof (1984), Krifka (1995), Landman (1998), and van Rooij (2002)). We will see that all of our observations about FC, as well as Chierchia’s observations about disjunction, follow from a novel (though fairly natural) approach to the meaning of the exhaustivity operator. It is often claimed that the neo-Gricean account of SIs follows from basic truisms about the nature of communication. However, as is well 71
72
Presupposition and Implicature in Compositional Semantics
known, one assumption is crucial, and far from trivial, namely the assumption that Grice’s Maxim of Quantity should be stated with reference to a formally defined set of alternatives. There is clearly no escape from formally defined alternatives. However, if the perspective argued for here is correct, access to these alternatives should be limited to grammar. A quantity maxim which is not contaminated by syntactic stipulations (together with appropriately placed syntactic stipulations, i.e., within grammar) derives better empirical results.
1
Some background on scalar implicatures
Consider a simple disjunctive sentence such as that in (1). When we hear such a sentence we draw a variety of inferences. (1)
Sue talked to John or Fred.
First, we conclude that (if the speaker is correct) Sue talked to John or to Fred, a conclusion, in and of itself, consistent with the possibility that Sue talked to both (Basic Inference). However, we typically also conclude (again assuming that the speaker’s utterance is correct) that this latter possibility was not attested (Scalar Implicature, SI). Finally, we infer that the speaker’s beliefs don’t determine which person (i.e. John or Fred) Sue talked to (Ignorance Inferences).1 (2)
Inferences we draw from (1): a. Basic Inference: Sue talked to John or Fred (or both). b. Scalar Implicature, SI: Sue didn’t talk to both John and Fred. c. Ignorance Inferences: The speaker doesn’t know that Sue talked to John. The speaker doesn’t know that Sue talked to Fred.
The nature of the inferences in (2a) and (2c) seems rather straightforward. The Basic Inference, (2a), is derived quite directly from the basic meaning of the sentence. The Ignorance Inferences, (2c), are not as direct, but, nevertheless, receive a fairly natural explanation. They are derived straightforwardly from a general reasoning process about the belief states of speakers, along lines outlined by Grice (1975). The source of the inference in (2b), SI, is, at least in my opinion, less obvious. The standard, neo-Gricean, approach captures this inference
Free Choice and Scalar Implicatures
73
by enriching the set of assumptions that enter into the derivation of Ignorance Inferences, while various competing proposals attribute the inference to a particular enrichment of the basic meaning. Before we see what is at stake, let’s start with a formulation of what might be uncontroversial, namely the account of (2c).2 The basic idea is that communicative principles require speakers to contribute as much as possible to the conversational enterprise. This idea is further elaborated when it is assumed that the goal of certain speech acts is to convey information, and that if all information is to be relevant, more is better. So, assume that two sentences are true and both contribute information that is completely relevant to the topic of conversation. If one contains more information than the other (i.e. is logically stronger),3 use of the more informative one would constitute a greater contribution: (3)
Maxim of Quantity (basic version): If S1 and S2 are both relevant to the topic of conversation and S1 is more informative than S2 , if the speaker believes that both are true, the speaker should utter S1 rather than S2 .
Typically, when (1) is uttered, the information conveyed by each of the disjuncts is relevant. Furthermore, each disjunct is more informative than the entire disjunction. (p entails p or q, but not vice versa.) The fact that the speaker, s, uttered the entire disjunction rather than just a disjunct, therefore, calls for an explanation. If we, the people who interpret the utterance, assume that s obeys the Maxim of Quantity, we conclude, for each disjunct, p, that it is false to claim that s believes that p is true, or if we keep to our convention of using the verb know instead of believe (see note 1), we can state this as a conclusion that s does not know that p is true. If we assume that s believes that her utterance of the disjunction is correct, we derive the Ignorance Inferences. But one logical property of the situation is worth focusing on. When we conclude that s does not believe that p is true, that is, in principle, consistent with two different states of affairs. s might believe that p is false, or, alternatively, she might have no (conclusive) opinion. The reason we infer the latter is that the former would be inconsistent with our other inferences. Under normal circumstances, we infer that s believes that her utterance of p or q is true (Maxim of Quality). If we were to assume that s believes that p is false, we would have to conclude that she believes that q is true. But that would conflict with our inference about q (based on the Maxim of
74
Presupposition and Implicature in Compositional Semantics
Quantity). Hence we must conclude, for each disjunct, that the speaker has no opinion as to whether or not it is true. 1.1
The symmetry problem4
Consider now whether we could extend this line reasoning to account for the SI in (2b). Since we’ve already concluded that the speaker does not know that p is true and that the speaker does not know that q is true, it follows that the speaker does not know that the conjunction p and q is true. This, again, is consistent with two different states of affairs. s might believe that p and q is false, or, alternatively, she might have no (conclusive) opinion. If this time we could exclude the latter possibility, we would derive the SI. The problem is that basically the same line of reasoning we’ve employed above leads us exactly to the opposite conclusion, namely to the exclusion of the possibility that s believes that p and q is false. The idea is fairly simple. The information that p and q is false, if true, would be relevant to the topic of conversation, hence the fact that s did not provide us with this information calls for an explanation.5 Once again, the natural explanation is that s did not have the information, i.e. that s did not know that p and q is false. In other words, instead of an SI, we derive, once again, an Ignorance Inference: we conclude that s does not know that p and q is true, and (exactly by the same type of reasoning) that s does not know that p and q is false, i.e. we conclude that s does not know whether or not p and q is true. As far as I know, a version of this problem was first noticed in Kroch (1972), and stated in its most general form in class notes of Kai von Fintel and Irene Heim. To appreciate the problem in its full generality, consider a general schema for deriving SIs in response to s s utterance of p (of, say, I have 3 children). We start by considering a more informative relevant utterance, p (say, I have 4 children), and reason that if p were true, and if s knew that p were true, the Maxim of Quantity would have forced s to utter p instead of p. We then might reason that it is plausible to assume that s knows whether or not p is true (say, that it is reasonable to assume that s knows how many children she has), and hence that s knows that p is false. The problem, however, is that there is always an equally relevant more informative utterance than p, namely p and not p (in our case, I have exactly 3 children), call it p . By the same reasoning process, if p were true and if s knew that p were true, the communicative principles would have forced s to utter p instead of p.6 Furthermore, if s knows the truth value of p, and of p , then s knows the truth value
Free Choice and Scalar Implicatures
75
of p . So, the same reasoning process leads to the conclusion that s knows that p is false. The assumption that the speaker knows whether or not p is true, thus, leads to a contradiction and must, therefore, be dropped. This problem was dubbed the symmetry problem in class notes of Kai von Fintel and Irene Heim. Whenever p is uttered, and fails to settle the truth value of a relevant proposition, q, there will be two symmetrical ways of settling it, leading necessarily to an Ignorance Inference. Stated somewhat differently, p is equivalent to the following disjunction (p ∧ q) ∨ (p ∧ ¬q), and therefore should lead to Ignorance Inferences parallel to those stated in (2c). 1.2
The neo-Gricean approach
The neo-Griceans respond to this problem by a revision of the Maxim of Quantity. Specifically, they suggest that the maxim doesn’t require speakers to utter the most informative proposition that is relevant to the topic of conversation, but is more limited in scope. The Maxim merely requires speakers to choose the most informative relevant proposition from a formally defined set of alternatives. It does not require speakers to consider all relevant propositions. The common way to work this out, pioneered by Larry Horn (1972), starts out with the postulation of certain sets of lexical items, Scalar Items, and sets of alternatives to which the scalar items belong, which we will call Horn-Sets:7 (4)
Examples of Horn-Sets a. b. c. d.
{or, and} {some, all} {one, two, three, . . .} {can, must}
These sets of lexical alternatives determine the set of (Horn) alternatives for a sentence by a simple algorithm. The set of alternatives for S, Alt(S), is defined as the set of sentences that one can derive from S by successive replacement of Scalar Items with members of their Horn-Set.8 (5)
Alt(S) = {S : S is derivable from S by successive replacement of scalar items with members of their Horn-Set}
76
Presupposition and Implicature in Compositional Semantics
The Maxim of Quantity can now be stated as follows: (6)
Maxim of Quantity (Neo-Gricean version): If S1 and S2 are both relevant to the topic of conversation, S1 is more informative than S2 , and S1 ∈Alt(S2 ), then, if the speaker believes that both are true, the speaker should prefer S1 to S2 .
Consider the sentence in (1). The postulated scalar item in this sentence is disjunction, for which conjunction is lexically specified as the only alternative, (4a). (1), thus, has just one alternative (other than (1) itself): (7)
Alt(1) = {(1), Sue talked to John and Fred }
When s utters (1), his addressee, h (for hearer), typically concludes (on the assumption that s obeys the revised Maxim of Quantity) that s does not know that the conjunctive sentence in Alt(1) is true, since this alternative sentence is more informative than s s utterance, and is typically relevant. If h assumes, further, that s has an opinion as to whether or not the conjunctive sentence is true, h would conclude that s believes that it is false. The neo-Griceans, thus, attribute a general tendency to addressees, namely the tendency to assume that speakers are opinionated. I state Sauerland’s formulation of this assumption in (8). (8)
Opinionated Speaker (OS): When a speaker, s, utters a sentence, S, the addressee, h, assumes, for every sentence S ∈Alt(S), that the beliefs of s determine the truth value of S , unless this assumption about S leads to the conclusion that the beliefs of s are contradictory.
Under the basic version of the Maxim of Quantity in (3), B-MQ, there was no way to maintain the assumption that the speaker is opinionated about any relevant sentence S (not entailed by S). To repeat, B-MQ derived the symmetric results (a) that the speaker does not know that S and S is true, and (b) that the speaker does not know that S and not S is true. This, together with the assumption that the speaker knows that S is true (Quality), derived the conclusion that the speaker is not opinionated about S . By contrast, under the Neo-Gricean version of the Maxim of Quantity in (6), NG-MQ, the assumption that the speaker is opinionated about various sentences (not entailed by S) is innocuous. NG-MQ does not always derive the inference that the speaker does not know that S and S is true. It derives such an inference only when S and S (or some equivalent
Free Choice and Scalar Implicatures
77
sentence) is a member of Alt(S). Under such circumstances, the speaker could be opinionated about S as long as Alt(S) does not have S and not S as a member (nor some equivalent sentence). If S and not S is not a member of Alt(S), NG-MQ does not derive the inference that the speaker does not know that S and not S is true, and the assumption that the speaker believes that S and S is false could be made consistently. To summarize, assume that a speaker s utters the sentence in (1), Sue talked to John or Bill. The addressee, h, assumes that s obeys the Maxim of Quality as well as the revised Maxim of Quantity, NG-MQ. Based on this assumption, h reasons in the following way: 1. Given NG-MQ, there is no X∈Alt(1), such that X is logically stronger than (1), and s thinks that X is true. 2. Alt(1) contains the conjunctive sentence Sue talked to John and Bill, which is logically stronger than s s utterance. Hence, given 1, it’s not the case that s thinks that this conjunctive sentence is true. 3. Given OS, the default assumption is that s has an opinion as to whether Sue talked to John and Bill is true or false. Given 2 (the conclusion that it’s not the case that s thinks that the sentences is true), we can conclude that s thinks that it is false. So, by modifying the set of assumptions that derive Ignorance Inferences (replacing B-MQ with NG-MQ) one can account for the SI in (2b). I would like at this point to discuss a possible alternative that keeps B-MQ in tact but instead enriches the set of syntactic representations available for (1). But it is worth pointing out first that, as things stand right now, our account of the Ignorance Inferences in (2c) is in jeopardy. Specifically, it is incompatible with NG-MQ and our assumption in (4a) about the Horn-Set for disjunction. The account was crucially dependent on the assumption that the Maxim of Quantity would prefer the utterance of a disjunct to the utterance of a disjunction, an assumption incompatible with the way Alt(1) is defined on the basis of (4a). One might respond to this problem with an independent (pragmatic) account for (2c) (Gazdar 1979) or by enriching the Horn-Set for disjunction (Sauerland 2004). The latter will be discussed in greater detail in Section 4. 1.3
An alternative syntactic approach
The alternative syntactic approach that I would like to defend is guided by the intuition that a principle of language use (such as the Maxim of Quantity) should not be sensitive to the formal (and somewhat arbitrary)
78
Presupposition and Implicature in Compositional Semantics
definition of Alt(S).9 If this intuition is correct, B-MQ is to be preferred to NG-MQ. However, as pointed out in Section 1.1, B-MQ derives Ignorance Inferences that contradict attested SIs. Therefore, if B-MQ is correct, something else is needed to derive SIs. More specifically, SIs must be derived from the basic meaning of the relevant sentences; otherwise the symmetry situation would yield unwanted Ignorance Inferences. Following proposals by Groenendijk and Stokhof (1984), Krifka (1995), Landman (1998), van Rooij (2002), and to some extent Chierchia (2004), I would like to suggest that the syntax of natural language has a covert operator which is optionally appended to sentences, and that this operator is responsible for SIs.10 The guiding observation is that there is a systematic way to state the SI of a sentence, using the focus sensitive operator only. Consider the sentence in (9), which has the SI that John didn’t buy 4 houses. (9)
John bought three houses.
This SI could be stated explicitly using the focus sensitive particle only in association with the numeral expression three. (10)
John only bought THREE houses.
This observation extends to all SIs; SIs can always be stated explicitly with the focus sensitive particle only, as long as the relevant scalar items bear pitch accent: (11)
(12)
a. John did some of the homework. b. Implicature: John only did SOME of the homework. For all of the alternatives to ‘some’, d, if the proposition that John did d of the homework is true, then it is entailed by the proposition that John did some of the homework. a. John talked to Mary or Sue. b. Implicature: John only talked to Mary OR Sue. For all of the alternatives to ‘or’, con, if the proposition that John talked to Mary con Sue is true then it is entailed by the proposition that John talked to Mary or Sue.
Sentence that generates SIs usually contain scalar items,11 and in such cases it is always possible to state the SIs explicitly, by appending the
Free Choice and Scalar Implicatures
79
operator only to the sentence and placing focal accent on the relevant scalar item: (13)
The only implicature generalization (OIG): A sentence, S, as a default, licenses the inference/implicature that (the speaker believes) only S’, where S’ is a modification of S with focus on scalar items.
From the neo-Gricean perspective discussed in Section 1.2, it is pretty clear why (9)–(12) should obey the OIG. As is commonly assumed, and as indicated by the italic paraphrases, the role of only is to eliminate alternatives. Furthermore, when focus is placed on scalar items, the relevant alternatives are precisely the Horn-alternatives that NG-MQ refers to. However, the OIG suggests yet another possibility, namely that B-MQ is the right conversational maxim and SIs are derived within the grammar, namely by of a covert exhaustivity operator with a meaning somewhat a kin to that of only. Assume, for the moment, the semantics for only suggested by the paraphrases in (9)–(12). Specifically, assume that only combines with a sentence (the prejacent), p, and a set of alternatives, A, (determined by focus). The result of this combination is a sentence which presupposes that p is true and furthermore asserts that every true member of A is already entailed by p (i.e. that all non-weaker alternatives, all ‘real alternatives’, are false): (14)
[[only]] (A<st,t> )(pst ) = λw: p(w) = 1. ∀q∈NW(p,A): q(w) = 012 NW(p,A) = {q∈A: p does not entail q}
The exhaustivity operator, exh, should mean the same, with one small modification. While with only the requirement that the prejacent be true is a presupposition, with exh this requirement should be part of the assertive component: (15)
[[Exh]] (A<st,t> )(pst )(w) ⇔ p(w) & ∀q∈NW(p,A): ¬q(w)
Assume that natural language has exh as a covert operator. Assume, further, that this operator can append to a sentence, S, thereby yielding a stronger sentence S+ = Exh(Alt(S))(S).13 It is easy to see that such a representation would derive both the ‘basic meanings’ and the SIs of the sentences in (9)–(12). This would allow us to keep to the non-stipulative quantity maxim (B-MQ). The cost lies, of course, in the stipulation of exh. There is a clear trade-off here, one that suggests that no decision will be justifiable
80
Presupposition and Implicature in Compositional Semantics
on a priori grounds. The goal of this chapter is to provide an empirical argument in favor of a theory in which B-MQ is at the heart of pragmatic reasoning and exh is responsible for SIs. If such a theory is correct, we might think of exh as a syntactic device designed (‘by a super-engineer’) to facilitate communication in a pragmatic universe governed by B-MQ. Every conversational situation, C, can be characterized by a set of sentences that are relevant at C, QC (for question). An utterance of S at C will be associated with a set of Ignorance Inferences determined by the set of sentences QS,C ⊆QC , whose truth value is not determined by S. B-MQ will derive the set of Ignorance Inferences that correspond to QS,C , i.e., I-INF(S,C) = {¬Ks ϕ: ϕ ∈QS,C }. Furthermore, if ϕ is relevant to the topic of conversation, it seems that same would be true of ¬ϕ (see note 5). Hence I-INF(S,C) = {¬Ks ϕ: ϕ ∈ QS,C } = {¬Ks ϕ and ¬Ks ¬ϕ : ϕ ∈QS,C } Sometimes the set of Ignorance Inferences will be implausible and this would motivate a new parse of the linguistic stimuli, one that involves an exhaustive operator on top of S, i.e. S+ . If no further stipulations are added, the procedure should be able to apply recursively leading to S++ , S+++ , etc. As we will see, this possibility will empirically distinguish our syntactic perspective from the Neo-Gricean alternative. In the next section, I will present the core empirical phenomena that I will use to motivate exh, namely the conjunctive interpretation of disjunction under existential modal constructions (free choice, FC). Furthermore, I will present an argument, due to Kratzer and Shimoyama (2002) (K&S) and Alonso-Ovalle (2005), that FC should be derived by the system that derives SIs. In Section 3, I will present evidence that FC arises in additional circumstances: when disjunction is embedded under certain other existential quantifiers and when conjunction is embedded under negation and universal quantifiers (under the sequence ¬∀). In Section 4, I will discuss a relevant observation about disjunction due to Chirchia (204), and in Section 5 I will present the Neo-Gricean response to Chierchia’s observation (Sauerland 2004) and its failure to predict FC phenomena. Finally, in Section 6–10, I will propose a resolution based on recursive exhaustification which extends to the phenomena discussed in Sections 2 and 3.
2
The problem of free choice permission
Consider the sentence in (16) when uttered by someone who’s understood to be an authority on the relevant rules and regulations, for example, a parent who is accustomed to specifying limits pertaining to the consumption of sweets.
Free Choice and Scalar Implicatures
(16)
81
You’re allowed to eat the cake or the ice-cream.
In such a context, (17) would be an immediate inference (Kamp 1973). This inference, sometimes referred to as an inference of free choice permission, is not an expected entailment of obvious candidates for the logical form of (16). (17)
You’re allowed to eat the cake and you are allowed to eat the ice-cream.
For example, (17) is not entailed by (18).14 (18)
Plausible LF for (17): allowed [[you eat the cake] or [ you eat the ice-cream]]
What (18) states is that the relevant rules do not prohibit the disjunctive sentence (the complement of allowed), or, in the terms of possible world semantics, that there is a world consistent with the rules in which one of the disjuncts is true. (18) is thus equivalent to (19), which is clearly weaker than (17). (19)
You are allowed to eat the cake or you are allowed to eat the ice-cream.
The problem is to understand how a disjunctive LF such as (18) can be strengthened to yield the conjunctive inference in (17). In other words, we need to understand how a sentence that should receive the modal logic formalization in (20a) – which is equivalent to (20b) – justifies the FC inference in (20c). (20)
a. ♦(p ∨ q) b. ♦p ∨ ♦q (a ≡ b) c. Free Choice: ♦p ∧ ♦q
2.1 Downward entailing operators, evidence that free choice is an implicature K&S studied FC effects that arise when certain indefinite expressions are embedded under existential modals, and presented a fairly strong argument that the effect should be derived by the system that yields SIs.15 The argument, which has been elaborated and extended to disjunctive
82
Presupposition and Implicature in Compositional Semantics
constructions by Alonso-Ovalle (2005), is based on the observation that there are no traces of FC in certain downward entailing contexts. Consider the sentence in (21). If FC were to follow from the basic meaning of (16), we would expect (21) to have a fairly weak meaning; it should be able to assert that no one is both allowed to eat the cake and allowed to eat the ice-cream, (21 a). We would thus predict (21) to be true in a situation in which everyone is allowed to eat one of the two disserts but no one has free-choice, i.e. in situations in which no one is allowed to decide which of the two deserts to eat. Such an interpretation, if available, is extremely dispreferred.16 To derive the natural interpretation, (21 b), we must factor out whatever is responsible for FC. (21) No one is allowed to eat the cake or the ice-cream. (21)’ a. *negation of FC: ¬∃x[♦P(x) ∧ ♦Q(x)] b. negation of standard meaning: ¬∃x♦(P(x) ∨ Q(x)) As pointed out by Alonso-Ovalle (following K&S), the natural interpretation, (21 b), is expected if FC were to be derived as an SI. Although it is not yet clear how to derive FC as an SI, it is clear that if a derivation were available for the basic case, it would, nevertheless, not be available (at least not necessarily) for (21). This is seen most clearly under the neo-Gricean approach to SIs. Under this approach, an SI is derived as a pragmatic strengthening of the basic meaning of a sentence. The meaning in (21 a) is weaker than the basic meaning in (21 b), and, therefore, cannot be derived along neo-Gricean lines. Under the syntactic alternative, the preference for (21 a) would be stated as a preference for stronger interpretations (See Chierchia 2004, 2005). More specifically, assume, contrary to what you might think at this point, that an exhaustive operator can somehow derive the basic FC effect. We might then suggest that exh can only be introduced if the overall result is a stronger proposition. This could be motivated by the observation that as propositions get stronger fewer Ignorance Inferences are derived by B-MQ. We might, thus, suggest that the introduction of exh is subject to an economy condition related to its functional motivation, namely to the elimination of Ignorance Inferences (i.e. a sentence with exh must lead to fewer Ignorance Inferences than its counterpart without exh, see note 37). Alternatively, we might suggest, in line with the neo-Griceans, that exh must be introduced in matrix position.17 Be that as it may, it is reasonable to assume that the preference for (21 b) would be predicted if FC could be derived as an SI, but not otherwise. Quite independently of particular proposals, it is well known that
Free Choice and Scalar Implicatures
83
SIs tend to disappear in downward entailing environments (Gazdar 1979, Chierchia 2004). The fact that FC appears to share this property with SIs seems to be a good incentive to search for a theory that would derive the effect as an SI. 2.2
Deriving an FC implicature: the nature of the problem
The projection properties of FC seem to suggest that the effect should be derived as an SI. But how could we derive such a result? K&S made a very interesting proposal which is the basis for the proposal that I will make in Sections 6–10. But this will require quite a bit of ground work. At this point, it is worth understanding what one might say in order to derive SIs based on the neo-Gricean maxim of quantity (NG-MQ). Quite generally, suppose that ϕ is the basic meaning of a sentence, S, and that our goal is to derive a stronger meaning, ϕ , based on NG-MQ. The result could be achieved if we proposed a set of alternatives of the following sort: ALT(S) := {S, S & not Sϕ }, where the meaning of Sϕ is ϕ . If s was to utter S, the addressee, h, would conclude, based on NG-MQ, that s does not believe that not Sϕ is true. Furthermore, based on the assumption that s is an opinionated speaker, h, would conclude that s believes that Sϕ is true. More specifically, suppose that the alternatives of the sentence in (16), repeated below, are the sentences in (22). This is slightly different from the general scheme for deriving implicatures characterized above, but the basic idea is the same.18 (16) (22)
You’re allowed to eat the cake or the ice-cream. Alternatives needed to derive FC for (16) based on NG-MQ: a. You are allowed to eat the cake or the ice-cream. b. You are allowed to eat the cake but you are not allowed to the ice-cream c. You are allowed to eat the ice cream but you are not allowed to eat the cake.
Based on NG-MQ, we would now derive the SI that (22b) and (22c) are both false, which, together with (22a), yields the FC inference. To see this, assume (22a) is true. Now assume that one of the conjuncts in (17) is false, say that you cannot eat the ice-cream. From this it follows that (22b) is true, contrary to the SI. But of course this is not intended as a serious proposal. It follows from a general algorithm that allows us to derive, on a case by case basis, any
84
Presupposition and Implicature in Compositional Semantics
SI that we would like to, and, hence, does not explain the particular SIs that are actualized (See Saeboe 2004). The obvious way to turn this into a serious proposal is to show that the alternatives in (22) are needed on independent grounds. K&S, and in particular Alonso-Ovalle, propose a more natural set of alternatives, namely the one in (23). (23)
Alternatives proposed by K&S/Alonso-Ovalle: a. You are allowed to eat the cake or the ice-cream. b. You are allowed to eat the cake. c. You are allowed to eat the ice cream.
The set of alternatives in (23), in contrast to the one in (22), is consistent with a general constraint on alternatives proposed in Matsumoto (1995).19 Furthermore, as we will see in Section 4, there is independent evidence for the type of Horn-Sets that would derive (a super-set of) the alternatives in (23). The problem, however, is that NG-MQ can not derive the FC effect on the basis of (23). In fact, as we will see in greater detail in Section 5.2, it derives Ignorance Inferences that directly conflict with FC. K&S suggest, however, that FC should be derived from (23) based on a novel principle, which they call anti-exhaustivity. When h interprets s s utterance of (23a), s needs to understand why it is that s preferred this sentence to any of the alternatives. The standard Neo-Gricean reasoning, which relates to the basic meaning of the alternatives, would lead to the conclusion that s does not know/believe that any of the alternatives is true. K&S, however, suggest that h might reason based on the strong meaning (basic meaning + implicatures) of the alternatives. Specifically, K&S suggest that h would attribute the choice of s to the belief that the strong meanings of (23b) and (23c) are both false. Furthermore they assume that the strong meaning of (23b) and (23c) is the basic meaning of (22b) and (22c) respectively. As pointed out by Aloni and van Rooij (forthcoming), this line of reasoning raises a question pertaining to simple disjunctive sentences, such as (1). We would like to understand why such sentences don’t receive a conjunctive interpretation via an anti-exhaustivity inference of the sort outlined above. If each disjunct is an alternative to a disjunctive sentence, why doesn’t the speaker infer that the exhaustive implicature of each disjunct is false? K&S provide an answer this question by postulating a covert modal operator for any disjunctive sentence. I will not go over this proposal
Free Choice and Scalar Implicatures
85
and the way it might address Aloni and van Rooij’s objection. I would like, instead, to raise another challenge to K&S s basic idea. I think it is important to try to understand how the anti-exhaustive inference fits within a general pragmatic system that derives Ignorance Inferences (as well as SIs). Specifically, I think it is important to understand why NG-MQ does not lead to the Ignorance Inferences in (24) (see Section 4.2 for details). (24)
Predicted inferences of (16), based on (23) and NG-MQ: a. s doesn’t know whether or not you can eat the cake. b. s doesn’t know whether or not you can eat the ice cream.
This is a challenge that this chapter attempts to meet. The idea, in a nut-shell, is to eliminate NG-MQ in favor of the non-stipulative alternative B-MQ. However, understanding how this is to work requires the introduction of a proposal made in Sauerland (2004), which would be extracted from its neo-Gricean setting in order to meet our goals. But before I get there, I would like to introduce an additional challenge. Specifically, I would like to present a few other surprising inferences that are intuitively similar to FC, and should, most likely, be derived by the same system.
3
Other free choice inferences
In this section we will see effects that are very similar in nature to FC, but arise in somewhat different syntactic contexts. These effects will argue for a fairly general explanation of the basic phenomenon, one that is not limited to modal environments or to disjunction. 3.1
FC Under negation and universal modals
Consider the sentence in (25) when uttered by someone who is understood to be an authority on the relevant rules and regulations, for example, a parent who is accustomed to assigning after-dinner chores. (25)
You are not required to both clear the table and do the dishes.
In such a context, (26) would normally be inferred by the addressee. (26)
You are not required to clear the table and you are not required to do the dishes.
86
Presupposition and Implicature in Compositional Semantics
This inference seems very similar to the FC inference drawn in (17) based on (16). To see the similarity, notice that the basic meaning of (25) is predicted to be equivalent to the disjunctive statement that you are allowed to either avoid clearing the table or evade doing the dishes, (27a), and that (26) is equivalent to the conjunction of two possibility statements. (You are allowed to avoid clearing the table and you are allowed to avoid doing the dishes, (27b).) (27)
a. Standard Meaning of (25) ¬(p ∧ q) ≡ ♦¬ (p ∧ q) ≡ ♦(¬p ∨ ¬q) ≡ ♦(¬p) ∨ ♦(¬q) b. Free Choice Inference ♦(¬p) ∧♦(¬q)
Just as in (16), the basic meaning does not explain the inference, and the gap is formally identical. Once again, we have to understand how a sentence that is equivalent to a disjunctive construction can be strengthened to something equivalent to the corresponding conjunction. 3.2
More generally under existential quantifiers
In the basic FC permission sentence in (16), disjunction appears in the scope of the existential modal allowed. Furthermore, as is well-known, FC extends to all constructions in which or is in the scope of an existential modal: (28)
a. The book might be on the desk or in the drawer. (= The book might be on the desk and it might be in the drawer) b. He is a very talented man. He can climb Mount Everest or ski the Matterhorn. (= He can climb Mount Everest and he can ski the Matterhorn.)
What has not been discussed in any systematic way is that this type of conjunctive interpretation extends also to some non-modal constructions:20 (29)
a. There is beer in the fridge or the ice-bucket. (= There is beer in the fridge and there is beer in the ice-bucket.) b. Most people walk to the park, but some people take the highway or the scenic route. (Irene Heim, pc attributed to Regine Eckardt, pc) (= Some people take the highway and some people take the scenic route.)
Free Choice and Scalar Implicatures
87
c. This course is very difficult. In the past, some students waited 3 semesters to complete it or never finished it at all. (Irene Heim, pc) (= Some students waited 3 semesters to complete the course and some students never finished it at all.) It is thus tempting to suggest that conjunctive interpretations for disjunction are available whenever disjunction is in the scope of an existential quantifier (with the domain of quantification, worlds or individuals, immaterial). However, there are limitations: (30)
a. There is a bottle of beer in the fridge or the ice-bucket. ( = There is a bottle of beer in the fridge and there is a bottle of beer in the ice-bucket.) b. Someone took the highway or the scenic route. ( = Someone took the highway and someone took the scenic route.) c. This course is very difficult. In the past, some student waited 3 semesters to complete it or never finished it at all. ( = some student waited 3 semesters to complete the course and some student never finished it at all.)
As pointed out in Klindinst (2005), the relevant factor seems to be number marking on the indefinite. We might, therefore, suggest the following generalization: (31)
Existential FC: A sentence of the form ∃x [P(x) ∨ Q(x)] can lead to the FC inference, ∃xP(x) ∧ ∃xQ(x), as long as the existential quantifier, ∃x, is not marked by singular morphology.
3.3 More generally, under negation and universal quantifiers In (25) we saw an FC effect arising when conjunction is under the scope of negation and a universal modal (under the sequence, ¬). As illustrated in (32), and stated in (33), the effect arises also when is replaced by an ordinary universal quantifier: (32)
We didn’t give every student of ours both a stipend and a tuition waiver. 1. basic meaning: ¬∀x[P(x) ∧ Q(x)] ≡ ∃x¬ [P(x) ∧ Q(x)] ≡
88
Presupposition and Implicature in Compositional Semantics
∃x [¬P(x) ∨¬Q(x)] ≡ ∃x ¬P(x) ∨∃x¬Q(x) 2. Free Choice: ∃x ¬P(x) ∧∃x¬Q(x) (33)
Conjunctive FC: A sentence of the form ¬∀x[P(x) ∧ Q(x)] can lead to the FC inference ∃x ¬P(x) ∧ ∃x¬Q(x).
In Sections 7–10 we will provide an account of our two generalizations ((31) and (33)) within a general theory of SIs that we will introduce in Section 6, based on a discussion, in Section 5, of Sauerland’s approach to SIs. But before we move on, it is important to rule out an alternative explanation of (25) and (32), in terms of wide scope conjunction. To understand the concern, focus on (25). One might think that this sentence has a logical form in which conjunction takes wide scope over the sequence ¬. If such a logical form were available, the inference in (26) would follow straightforwardly from the basic meaning, and would thus be unrelated to the FC effects that are distributed according to (31). But wide scope conjunction is not a probable explanation. One argument against such an explanation is based on the sentences in (34). If conjunction could take scope over the sequence ¬ in (25) (and over ¬∀ in (32)), we would expect it to be able to outscope negation in (34), an expectation that is not born out.21 (34)
a. I didn’t talk to both John and Bill. b. We didn’t give both a stipend and a tuition waiver to every student.
What I think we learn from (34) is that conjunction can appear to outscope negation only when a universal quantifier intervenes.22 This is expected if conjunction never outscopes negation, and the generalization in (33) is real. Another argument against wide scope conjunction comes from an additional inference we draw from sentences such as (25) and (32). In both cases we draw the inference that the alternative sentence with disjunction instead of conjunction is false. That is, we would tend to draw (35a) as an inference from (25), and (35b) from (32). These inferences are not expected if conjunction receives wide scope, but, as we will see later on, are expected if the phenomenon is derived along with other FC effects. (35)
a. You are required to clear the table or do the dishes. b. We gave every student of ours a stipend or a tuition waiver.
Free Choice and Scalar Implicatures
4
89
Chierchia’s puzzle
The account of FC that I will develop will be based on a modification of a proposal made in Sauerland (2004) to deal with a puzzle discovered in Chierchia (2004).23 To understand the puzzle in greater detail, consider first (36) and its implicature that (36) is false. (36) John did some of homework. (36)’ John did all of the homework. As outlined in Section 1.2, this implicature can be derived by NG-MQ under the assumption that some and all are members of the same Horn-Set, (4b),24 from which it follows that (36) is an alternative to (36). NG-MQ, together with the assumption of an opinionated speaker, lead to the conclusion that the speaker believes that (36) is false. Consider next what happens when (36) is embedded as one of two disjuncts: (37)
John did the reading or some of homework.
This type of embedding was presented by Chierchia (2004) as a challenge to the neo-Gricean derivation of implicatures. As Chierchia points out, (37) should be an alternative to (37), and it would therefore seem that (with the assumption that the speaker is opinionated) we should derive the implicature that (the speaker believes that) (37) is false. (37)
John did the reading or all of homework.
This implicature, however, is clearly too strong. If a disjunctive sentence is false, then each of the disjuncts is false. When (37) is uttered, we do derive the inference that the second disjunct of (37) is false. However, we clearly do not derive a similar inference for the first disjunct (which is also the first disjunct of (36)). Chierchia’s challenge for the Neo-Griceans is to avoid the implicature that the first disjunct of (36) is false while at the same time to derive the implicature that the stronger alternative to the second disjunct is false: (38)
Let U be an utterance of p or q where q has a stronger alternative, q . a. Problem 1: to avoid the implicature of ¬p b. Problem 2: to derive the implicature of ¬q’
90
Presupposition and Implicature in Compositional Semantics
Chierchia provides an account for the relevant generalization based on a recursive definition of strengthened meanings. I will not discuss his account, since I can’t figure out how to extend it to FC. Instead, I will discuss the neo-Gricean alternative, which also fails to account for FC, but, which can, nevertheless, be modified in order to provide a syntactic (non-Gricean) alternative that successfully extends to FC.25
5
Sauerland’s proposal26
As pointed out at the end of Section 1.2, the Horn-Set for disjunction in (4a) ({or, and}) cannot account for the Ignorance Inferences that are attested when a simple disjunctive sentence such as (1) is uttered. Sauerland suggests a remedy for this problem which also resolves Chierchia’s puzzle. To derive the appropriate Ignorance Inferences for (1), Sauerland suggests that the alternatives for a disjunctive statement include each of the disjuncts in addition to the corresponding conjunction: p
(39) Alt(p∨q) = p∨q
p∧q q
These alternatives, which are plotted to represent logical strength,27 derive (based on NG-MQ) the following inferences with respect to a speaker, s, who utters p or q, inferences which Sauerland calls Primary (or weak) Implicatures, PIs: (40)
PIs for p or q (based on NG-MQ) a. s does not believe that p is true. b. s does not believe that q is true. c. s does not believe that p and q is true. both a and b.
Already follows from
Given that s is assumed to believe that her utterance of p or q is true (Quality), we derive the Ignorance Inferences discussed in Section 1, that is, for each disjunct, we derive the inference that the speaker does not know whether or not it is true. To derive SIs, the principle of an Opinionated Speaker is employed, (8): (8)
Opinionated Speaker (OS): When a speaker, s, utters a sentence, S, the addressee, h, assumes, for every sentence S ∈Alt(S), that the
Free Choice and Scalar Implicatures
91
beliefs of s determine the truth value of S , unless this assumption would lead to the conclusion that the beliefs of s are contradictory. This principle asks us to scan the set of alternatives that are stronger than S, and to identify those for which the assumption that the speaker is opinionated is consistent with our prior inferences based on Quality and NG-MQ. For each such alternative, the speaker is assumed to be opinionated, and given the relevant PI, a stronger inference is derived, namely that the speaker believes that the relevant alternative is false, an inference which Sauerland calls a Secondary Implicature (an SI, conveniently). As mentioned above, (40a,b) together with Quality, lead to ignorance with respect to p and to q. Hence, p and q is the only alternative for which the assumption that the speaker is opinionated is consistent with prior inferences. Therefore, only one SI is derived based on OS, namely the inference that the speaker believes that p and q is false: (41)
SI for p or q (based on OS) s believes that p and q is false.
Sauerland, thus, derives the following definition for the two relevant sets of implicatures: (42)
When a speaker s utters a sentence A, the following implicatures are derived: a. PIs = {¬Bs (A’): A’∈ ALT(A) and A’ is stronger than A} b. SIs = {Bs (¬A’): A’∈ ALT(A), A’ is stronger than A, and Bs (A)∧∩PI ∧ Bs (¬A’) is not contradictory}
Based on these definitions, a PI is derived for every alternative stronger than the assertion and an SI for a subset of the stronger alternatives for which an Ignorance Inference hasn’t already been derived (based on NG-MQ and Quality):28 (43)
Implicatures for p ∨q: p Alt(p∨q) = p∨q
p∧q q
a. PIs: ¬Bs (p), ¬Bs (q) b. SI: Bs ¬(p∧q)
The rest, ¬Bs (p∧q), follows
92
Presupposition and Implicature in Compositional Semantics
Sauerland shows that this rather principled approach solves Chierchia’s puzzle once the lexical alternatives that derive the sentential alternatives in (39) are specified. The basic intuition is fairly straightforward. An utterance of p or q derives Ignorance Inferences that are inconsistent with the assumption that the speaker is opinionated about p, thereby solving problem (38a). Problem (38b) is solved as well, but seeing this requires precision about the relevant lexical alternatives and the way they determine sentential alternatives for complex disjunctions, such as (37). The starting point is the observation that in order to derive (39) the alternatives for disjunction must contain two lexical entries that are never attested:29
(44)
Horn-Set(or) = {or, L, R, and}, where pLq = p and pRq = q.
These four alternatives, when combined with the alternatives for some ({some, all}), yield 8 alternatives to (37), based on (5) above:30
(45)
Alt(r or sh) =
a. r ∨ sh b. r L sh ≡ r c. r R sh ≡ sh d. r ∧ sh e. r ∨ ah f. r L ah ≡ r g. r R ah ≡ ah h. r ∧ ah
To see what PIs and SIs are derived, it is useful to plot the alternatives in a way that indicates relative strength. But it is already easy to see how the two problems in (38) are solved. To repeat, problem (38a) is solved based on the observation that the speaker cannot believe that r is false if a PI ensures that she does not believe that sh is true and Quality ensures that she believes that r or sh is true. Problem (38b) is solved based on the observation that ah is a member of the alternative set (alternative g), and that an SI can be derived for this alternative (consistent with prior
Free Choice and Scalar Implicatures
93
inferences): (46)
Implicatures for r ∨ sh: ALT(r ∨ sh) = r r∨ah
r∧ah ah
r∨sh
r∧sh sh
PI = ¬Bs (r∨ah), ¬Bs (sh), SI = Bs (¬ah), Bs (¬(r∧sh))
(the rest follow) (the rest, Bs ¬(r∧ah), follows)
5.1 Advantage of Sauerland’s proposal: embedding under universal quantifiers The Horn-Set in (44) plays two independent roles for Sauerland. It provides NG-MQ with the alternatives needed to derive the Ignorance Inferences for p ∨ q. These inferences explain (within the Neo-Gricean paradigm) the lack of certain SIs when scalar items are embedded within one of the disjuncts (problem (38)a). Furthermore, given (5), we can generate alternatives for complex disjunctive sentences (e.g. q for the sentence in (38)) that derive otherwise surprising SIs (problem (38b)). However, the system makes a further prediction. Specifically, it predicts that in certain contexts the two basic alternatives p and q will generate SIs rather than Ignorance Inferences. The relevant case involves embedding of disjunction under an upward monotone operator O such that O(p ∨ q) does not entail the disjunctive sentence O(p) ∨ O(q). For such an operator, the following is not contradictory. (47)
O(p∨q) ∧¬O(p) ∧¬O(q) ∧¬O(p∧q)
Hence, if s utters O(p ∨ q), an SI would be generated for each of the stronger alternatives (O(p), O(q), and O(p∧q)). Evidence that this prediction is correct comes from (48) and (49), which naturally yield the implicatures in (a) and (b).31
94
(48)
Presupposition and Implicature in Compositional Semantics
You’re required to talk to Mary or Sue. Implicatures: a. You’re not required to talk to Mary. b. You’re not required to talk to Sue.
(49)
Every friend of mine has a boy friend or a girl friend. Implicatures: a. It’s not true that every friend of mine has a boy friend. b. It’s not true that every friend of mine has a girl friend.
These facts follow straightforwardly from the Sauerland scale:
(50)
Alt(∀x(P(x)∨Q(x)) = ∀x(P(x)∨Q(x))
∀xP(x) ∀x(P(x)∧Q(x)) ∀xQ(x)
PIs = ¬Bs (∀xP(x)), ¬Bs (∀xQ(x)) (the rest, ¬Bs ∀x(P(x)∧Q(x)), follows) SIs = Bs (¬∀xP(x)), Bs (¬∀xQ(x)) (the rest, Bs ¬∀x(P(x)∧Q(x)), follows) 5.2
But . . . what about FC?
Sauerland’s system makes yet another prediction about disjunction embedding, a prediction which is in direct conflict with FC. If disjunction is embedded under an upward monotone operator O such that O(p∨q) entails the disjunctive sentence O(p) ∨ O(q), the system predicts Ignorance Inferences with respect to O(p) and O(q). The reasoning is exactly identical to the basic case of unembedded disjunction: there is no way to assume that the speaker is opinionated about one of the alternatives O(p) and O(q) without contradicting the Primary Implicature that the speaker does not know that the other disjunct is true (given Quality). This does not seem to be the correct prediction for existential modals and plural existential DPs (generalization (31)). These operators, under their basic meaning, are both commutative with respect to disjunction (♦(p∨q) ≡ (♦p∨♦q); ∃x(P(x)∨Q(x)) ≡ ∃xP(x)∨∃xQ(x)). Hence Ignorance Inferences are predicted.
Free Choice and Scalar Implicatures
(51)
95
You may eat the cake or the ice-cream.
Alt(51) = ◊(p∨q)
◊p ◊q
◊(p∧q)
PIs = ¬Bs (♦p), ¬Bs (♦q), ¬Bs ♦(p∧q) SIs = Bs ¬♦(p∧q) We’ve seen good arguments that FC should be derived as an implicature. However, under Sauerland’s system we derive Primary Implicatures (in bold) that contradict FC. The situation is quite interesting. The alternatives that K&S and Alonso-Ovalle appeal to in order to derive FC derive contradictory Ignorance Inferences in Sauerland’s system. If K&S are right, Sauerland’s system needs to change. However, K&S’s insight, if correct, needs to be embedded in a general system for implicature computation, one that can account for Chierchia’s puzzle, as well as for the emergence of Ignorance Inferences.
6
An alternative perspective
There are two problems with Sauerland’s system. On the one hand, it derives Ignorance Inferences that directly contradict the attested FC effect. On the other hand, it does not provide the basis for antiexhaustivity, which, if K&S are correct, is at the heart of FC. I will argue that the first problem teaches us that, contrary to the neo-Gricean assumption, Primary Implicatures do not serve the foundation for the computation of SIs. Instead, SIs are derived in the syntactic/semantic component via an exhaustive operator, as suggested in Section 1.3. Once a semantic representation is chosen, Ignorance Inferences are computed by the pragmatic system, based on the non-stipulative maxim of quantity (B-MQ). Without an exhaustive operator, incorrect Ignorance Inferences are computed in FC environments. However, once we modify the meaning of exh, based on Sauerland’s insights, the inferences can be avoided by a sequence of two exhaustive operators, which yield, in effect, antiexhaustivity, thereby solving the second problem. Furthermore, it turns out that FC is predicted in all the environments discussed in Section 3. 6.1
A new lexical entry for the exhaustive operator
Let’s start by reviewing our lexical entry for only and exh from Section 1.3. These entries ((14) and (15), repeated below) derive strong meanings that
96
Presupposition and Implicature in Compositional Semantics
are in most cases equivalent to the basic meaning conjoined with the SIs derived by the neo-Gricean system. (52)
a. [[only]] (A<st,t> )(pst ) = λ w: p(w) = 1. ∀q∈NW(p,A): q(w) = 0 NW(p,A) = {q∈A: p does not entail q} b. [[Exh]] (A<st,t> )(pst )(w)⇔ p(w) & ∀q∈NW(p,A): ¬q(w)
However, predictions are sometimes different when the alternatives are not totally ordered by entailment. In particular, for the sets of alternatives that Sauerland has postulated, the lexical entries in (52) can derive contradictory results. To see this, consider the following dialogue: (53)
A: John talked to Mary or Sue. B: Do you think he might have spoken to both of them? A: No, he only spoke to Mary OR Sue.
Under (52)a, A’s final sentence should presuppose that the prejacent, John spoke to Mary or Sue, is true and that this is not the case for any of the (non-weaker) alternatives. Thus, if Sauerland is right about the lexical alternatives for disjunction, the two alternatives in (54) would both have to be false for the utterance to be true, which would, of course, contradict the presupposition. (54)
a. John talked to Mary. b. John talked to Sue.
This is a wrong result, which means that if Sauerland is right about the alternatives for disjunction, the lexical entries in (52) probably need to be revised.32 A revision of this sort is also needed based on much older observations due to Groenendijk and Stokhof (1984): (55)
a. Who did John talk to? Only Mary or SUE b. Who did John talk to? Only Some GIRL Groenendijk and Stokhof (1984)
Let’s focus on (55a). If (52a) is correct, the answer to the question should assert that every alternative not entailed by the prejacent, John talked to Mary or Sue, is false. This time the set of alternatives consists (most likely) of every proposition of the form John talked to x based on the denotation
Free Choice and Scalar Implicatures
97
of the question, and the fact that the whole DP Mary or Sue is focused. In other words, the answer to the question in (55a) should entail the propositions that John didn’t talk to Mary and that he didn’t talk to Sue, which should contradict the presupposed prejacent. Groenendijk and Stokhof (1984), who noticed the problem, suggested a modification to the standard lexical entry for only, which was accommodated in Spector (2006) to the syntax we are assuming (based on Schultz and van Rooij (2006)): (56)
a. [[only]] (A<st,t> )(pst ) = λw: p(w) = 1. Minimal(w)(A)(p) b. [[Exh]](A<st,t> )(pst )(w) ⇔ p(w) & Minimal(w)(A)(p) Minimal(w)(A)(p) ⇔ ¬∃w’p(w’) = 1(Aw ⊂ Aw ) Aω = {p∈A: p(ω) = 1}
As pointed out by Spector (again, based on van Rooij and Schultz), this lexical entry can solve Chierchia’s problem. However, it yields results that contradict FC (see note 41). For this reason, I would like to suggest an alternative, one that is linked in a very direct way to Sauerland’s proposal. What we learn from Groenendijk and Stokhof is that there is something in the meaning of only ‘designed’ to avoid contradictions: only takes a set of alternatives A and a prejacent p, and attempts to exclude as many propositions from A in a way that would be consistent with the requirement that the prejacent be true. I would like to suggest that the basic algorithm is Sauerland’s, i.e. that propositions from A are excluded as long as their exclusion does not lead (given p) to the inclusion of some other proposition in A: (57)
a. [[only]] (A<st,t> )(pst ) = λw: p(w) = 1. ∀q∈NW(p,A) [q is innocently excludable given A→q(w) = 0] b. [[Exh]] (A<st,t> )(pst )(w)⇔ p(w) & ∀q∈NW(p,A) [q is innocently excludable given A→ ¬q(w)]
q is innocently excludable given A if ¬∃q’∈ NW(p,A) [p∧¬q⇒ q’] To see how this is supposed to work, consider an utterance of the disjunction p or q. Consider first what happens without an exhaustive operator, under the basic syntactic representation. Under such a representation, the sentence would assert that the disjunction is true and would be consistent with the truth of the conjunction (inclusive or). By
98
Presupposition and Implicature in Compositional Semantics
B-MQ this would yield a variety of Ignorance Inferences, which might be implausible in a particular context, and if so, would motivate the introduction of an exhaustive operator, Exh(Alt(p or q))( p or q). Under this alternative parse the sentence would assert that the prejacent p or q is true and that every innocently excludable alternative is false. Assuming the Sauerland alternatives, we derive the simple ExOR meaning. None of the disjuncts is innocently excludable, since the exclusion of one will lead to the inclusion of the other, given the prejacent. Once again, we will circle the innocently excludable alternatives.
(58)
p Alt(p∨q) = p∨q
p∧q q
Excluding p will necessarily include q while excluding q will necessarily include p. p∧q is thus the only proposition in NW(p∨q, Alt(p∨q)) that can be innocently excluded given the set of alternatives in (58). Thus, it is the only proposition that is excluded and the derived meaning is the familiar ExOR. Before moving to FC, I would like to show how the lexical entries in (57b) replicates Sauerland results. But even before that, I would like to point out that Sauerland’s algorithm is not totally contradiction free, and that his assumptions should therefore be modified slightly. This modification would motivate a corresponding modification in (57). Consider the question answer pair in (59) from Groenendijk and Stokhof. Assume that the alternatives for A is the Hamblin denotation of Q, Alt((59)A), in (60). (59) (60)
Q: Who did Fred talk to? A: Some GIRL Alt((59)A) = {that Fred talked to x: x is a person or a set of people}
Assume also that there are more than two girls in the domain of quantification. If all these assumptions are correct, it would be possible (by
Free Choice and Scalar Implicatures
99
Sauerland’s algorithm) to introduce an SI of the form Bs ¬ϕ, for every ϕ in Alt((59)A). Each SI of this sort is consistent with the set of PIs and Quality. However, once all the SIs are collected, the result is contradictory. The problem extends to the lexical entry for exh in (57) (and for only, if we look back at (55b)). Every member of Alt((59)A) is innocently excludable. Hence, if we were to append exh to (59)A, the result would be contradictory. One way to deal with this problem is to assume that the set of alternatives is always closed under disjunction (see Spector 2005, as well as note 29). An alternative, which is available when exh is assumed, is to eliminate additional elements from the set of innocently excludable propositions for a prejacent, p, given a set of alternatives A, I-E(p,A):
(61)
a. [[only]] (A<st,> )(pst ) = λw: p(w) = 1. ∀q∈I-E(p,A)→ q(w) = 0 b. [[Exh]] (A<st,t> )(pst )(w)⇔ p(w) & ∀q∈ I-E(p,A)→ ¬ q(w) I-E(p,A) = ∩{A’⊆A: A’ is a maximal set in A, s.t., A’¬ ∪{p} is consistent} A¬ = {¬ p: p∈A}
To see if a proposition q is innocently excludable, we must look at every maximal set of propositions in A such that its exclusion is consistent with the prejacent. Every such set could be excluded consistently as long as nothing else in A is excluded. Hence the only propositions that could be excluded non-arbitrarily are those that are in every one of these sets (the innocently excludable alternatives). Every proposition which is not in every such set would be an arbitrary exclusion, since the choice to exclude it, will force us to include a proposition from one of the other maximal exclusions (if the result is to be consistent), and the choice between alternative exclusion appears arbitrary.33 To see what results are derived by this lexical entry, it is probably best to go through the various cases we’ve discussed. Let’s first see how we would exhaustify p∨q given the Sauerland alternatives. The first step would be to identify the maximal consistent exclusions given the prejacent p∨q. If p is excluded, q must be true and vice versa. Hence, one maximal exclusion is {p, p∧q}, and the other is {q, p∧q}. The intersection is p∧q, hence, Exh(Alt(p∨q))(p∨q) = (p∨q) ∧¬(p∧q) = p∇q.
100 Presupposition and Implicature in Compositional Semantics
(62)
p p∧q
Alt(p∧q) = p∨q q
We circle (with dotted-lines) the maximal exclusions consistent with the prejacent, and we circle the intersection, the set of innocently excludable alternatives, with a completed line. Exh(Alt(p∨q))(p∨q) = (p∨q) ∧¬(p∧q) = p∇q. Consider now the exhaustification of (59)A under the assumption that the set of alternatives is the Hamblin-denotation of the question, (60). Assume that there are three girls in the domain of quantification, Mary, Sue, and Jane, and that there are no non-girls.34 Every maximal exclusion will include every member of the Hamblin-set but one of the following: (m) Fred talked to Mary, (s) Fred talked to Sue, and (j) Fred talked to Jane. So the intersection of all-maximal exclusions, the set of innocentlyexcludable alternatives, is the set of propositions of the form Fred talked to X, where X is a plurality of girls: (63)
Alt((59)A) = {Fred talked to x: x a person or a set of people} =
m
m&s
s
m&j
j
m&s&j
s&j
Exh(Alt((59)A))((59)A) = Fred talked to some girl ∧¬(m& s) ∧ ¬ (m& j) ∧¬ (s& j) = Fred talked to exactly one girl. 6.2
Replicating Sauerland’s results
Consider again Chierchia’s sentence r ∨ sh and its Sauerland alternatives. To see which alternative can be innocently excluded we have to identify the maximal (consistent) exclusions. The set of innocently excludable alternative is the intersection. The reader can consult the diagram in (64) to see that Sauerland’s results are replicated.
Free Choice and Scalar Implicatures
(64)
101
Alt(r ∨ sh) =
r r∨ah r∨sh
r∧ah ah r∧sh
sh
ah, r ∧ sh, r∧ah are the proposition in Alt(ss∨b) that can be innocently excluded given the set of alternatives: Exh(Alt(r∨sh))(r∨sh) = (r∨sh)∧¬ ah ∧¬(r∧sh) Consider next embedding under universal quantifiers. As discussed in Section 5.1, such embedding allows for the consistent exclusion of all the Sauerland-alternatives (other than the prejacent). Hence, there is only one maximal exclusion, which is excluded by exh:35 (65)
You’re required to talk to Mary or Sue. Implicatures: a. You’re not required to talk to Mary. b. You’re not required to talk to Sue.
(66)
Every friend of mine has a boy friend or a girl friend. a. It’s not true that every friend of mine has a boy friend. b. It’s not true that every friend of mine has a girl friend. ∀xP(x)
(67) Alt(∀x(P(x)∨Q(x)) = ∀x(P(x)∨Q(x))
∀x(P(x)∧Q(x)) ∀xQ(x)
Exh (Alt(∀x(P(x)∨Q(x)))(∀x(P(x)∨Q(x))) = ∀x(P(x)∨Q(x)) ∧¬∀xP(x) ∧¬∀xQ(x)
So the exhaustive operator as defined in (61)b, based on what’s needed for only, (61)a, derives the same results as Sauerland’s system (with the
102
Presupposition and Implicature in Compositional Semantics
exception of cases such as (59) for which Sauerland’s system can derive contradictory implicatures). This is not surprising. The set of innocently excludable proposition is (modulo (59)) precisely the set of propositions for which Sauerland’s system yields SIs – for which SIs can be introduced innocently. However, there is an important architectural difference between the two systems, one that relates to the division of labor between syntax/ semantics and pragmatics. Under Sauerland’s neo-Gricean system, NG-MQ (and the PIs that it generates) is the underlying source of SIs. Under the syntactic alternative that we are considering, SIs have a syntactic source, and can serve to avoid Ignorance Inferences, which are computed post syntactically based on B-MQ. This architectural difference has empirical ramifications for FC. We’ve already seen that Sauerland’s system predicts Ignorance Inferences that contradict FC. We will see that under our syntactic alternative, the problem can be avoided by recursive exhaustification.
7
Recursive exhaustification and FC36
Suppose that exh is a covert operator which can append to any sentence. It is reasonable to assume that, in parsing (or producing) a sentence, exh will be used whenever the result fairs better than its counterpart without exh. One way in which a sentence with exh would be better than its exh-less counterpart is if the latter generates implausible Ignorance Inferences based on B-MQ. We thus predict the following recursive parsing strategy: (68)
7.1
Recursive Parsing-Strategy: If a sentence S has an undesirable Ignorance Inference, parse it as Exh(Alt(S))(S).37 Simple disjunction
Consider the disjunctive sentence in (69) (69)
I ate the cake or the ice-cream.
If this sentence is parsed without an exhaustive operator, B-MQ will generate the Ignorance Inference that the speaker doesn’t know what she ate (only that it included the cake or the ice-cream or both). This inference might seem implausible, and the hearer might therefore prefer the following parse, where C is the set of Sauerland-alternatives to the disjunctive sentence.
Free Choice and Scalar Implicatures
(70)
103
Exh(C)(I ate the cake or the ice-cream)
As we’ve seen already, the meaning of (70) is the ExOR meaning of (69). This meaning will now generate (given B-MQ) the Ignorance Inference that the speaker doesn’t know what she ate (only that, whatever it was, it included the cake or the ice-cream but not both). This, again, might seem implausible, and the hearer might employ the parsing strategy, again: (71)
Exh(C’)[Exh(C)(I ate the cake or the ice-cream)] where C’=Alt[Exh(C)(I ate the cake or the ice-cream)] = {Exh(C)(p) : p∈C}38
However, (71) ends up equivalent to (70). And further application of the parsing strategy is not helpful either. It will, thus, follow that there is no way to avoid the (sometimes undesirable) Ignorance Inference. To see this, we need to compute the set of alternatives, C’: (72)
C’ = {1. Exh(C) (p∨q), 2. Exh(C)(p), 3. Exh(C)(q), 4. Exh(C)(p∧q)} 1. Exh(C) (p∨q) = (p∨q) ∧¬ (p∧q) = (p ∧¬q)∨(q ∧¬p) 2. Exh(C)(p)= p∧¬q 3. Exh(C)(q)= q∧¬p 4. Exh(C)(p∧q) = p∧q (can be ignored since already excluded by the prejacent) Exh(C) (p∨q) = Exh(C)(p)∨Exh(C)(q)
Two simple observations are worth making. The first alternative, the prejacent of (71), is equivalent to the disjunction of the second and third alternative, and the fourth alternative is already excluded by the prejacent, and hence can be ignored. The relevant alternatives are thus the following: 2.Exh(C)(p)
(73)
C= Alt(Exh(p∨q)) = Exh(C)(p) ∨ Exh(C)(q) 3. Exh(C)(q) Exh(C)[Exh(C)(p ∨ q)] = Exh(C) (p ∨ q) = (p ∨ q) ∧ ¬ (p ∧ q)
If 2 is excluded, 3 must be true, and vice versa. Hence, meaning does not change with a second level of exhaustification, nor will it change when
104 Presupposition and Implicature in Compositional Semantics
exh is appended yet another time.39 There is thus no way to avoid what might be an undesirable Ignorance Inference. 7.2
The basic free choice effect
Consider now (74). (74)
You may eat the cake or the ice-cream.
Without an exhaustive operator, this sentence will generate the Ignorance Inference that the speaker doesn’t know what one is allowed to eat (only that the allowed things include the cake or the ice-cream or both). This might seem implausible, and the hearer might opt for another parse: (75)
Exh(C)(You may eat the cake or the ice-cream)
Given the Sauerland alternatives for disjunction, the set of alternatives, C, would be the following:40 (76)
Alt(74)
C = ◊(p ∨ q)
◊p ◊q
◊(p∧q) Notice ◊(p ∨ q) ⇔ ◊p ∨ ◊q but (crucially) ◊(p∧q) <≠> ◊p ∧ ◊q
♦(p∧q) is the only proposition in Alt(♦(p∨q)) that can be innocently excluded given the set of alternatives (excluding ♦p will necessarily include ♦q while excluding ♦q will necessarily include ♦p). Hence, the meaning of (75) in our modal logic formalization is ♦(p∨q)∧¬♦(p∧q). Crucially (75) is consistent with the free choice possibility, ♦p∧♦q, though it, of course, does not assert free choice.41 This new meaning will now generate the Ignorance Inference that the speaker doesn’t know what one is allowed to eat (only that the allowed things include the cake or the ice-cream but not both). This might seem implausible, and the hearer might employ the parsing strategy again: (77)
Exh(C’)[Exh(C)(You may eat the cake or the ice-cream)] where C’={Exh(C)(p): p∈C}
Free Choice and Scalar Implicatures
105
This time, second exhaustification has consequences. To see this, we need to compute the meanings of the various alternatives: (78)
C’= {1. Exh(C) (♦(p∨q)), 2. Exh(C)(♦p), 3. Exh(C)(♦q), 4. Exh(C) (♦(p∧q))} 1. Exh(C) (♦(p∨q)) = ♦(p∨q) ∧¬♦(p∧q), crucially = (♦p ∧¬♦q)∨ (♦q∧¬♦p) 2. Exh(C)(♦p)= ♦p ∧¬♦q 3. Exh(C)(♦q)= ♦q ∧¬♦p 4. Exh(C) ♦(p∧q) = ♦(p∧q) (can be ignored since already excluded by the prejacent)
C9 = ◊(p∨q) ∧¬ ◊(p ∧ q)
◊p ∧¬ ◊q ◊p ∧¬ ◊q
There are now two propositions in C’ that can be innocently (and nontrivially) excluded. (Excluding Exh(C)(♦p) will not necessarily include Exh(C)(♦q), and excluding Exh(C)(♦q) will not necessarily include Exh(C)(♦p).) Hence, (79)
Exc(C’)(Exh(C) (♦(p∨q))) = ♦(p∨q))∧¬♦(p∧q) ¬(♦p ∧¬♦q) ¬(♦q ∧¬♦p) = ♦(p) ∧ ♦(q) and ¬♦(p∧q)
and and
We thus derive the FC effect for cases in which disjunction is embedded under existential modals.42
8
Other existential quantifiers
The key to the distinction between disjunction embedded under an existential modal and unembedded disjunction is that in the latter case the strongest alternative ♦(p ∧ q) is stronger than the conjunction of the two other alternatives ♦p and ♦q. Hence, the first layer of exhaustification is consistent with the later exclusion of Exh(C)♦p and Exh(C)♦q.43 This answers Aloni and van Rooij’s (2005) objection
106
Presupposition and Implicature in Compositional Semantics
to Kratzer and Shimoyama (Section 2.2), and extends to account for embedding under existential quantifiers: (80)
a. There is beer in the fridge or the ice-bucket. b. People sometimes take the highway or the scenic route (Irene Heim, pc attributing Regine Eckardt, pc) c. This course is very difficult. Last year, some students waited 3 semesters to complete it or never finished it at all. (Irene Heim, pc)
Here, too, first exhaustification will be fairly weak ∃x(Px ∨ Qx)∧¬∃x(Px ∧ Qx) consistent with later exclusion of Exh(C)∃xPx and Exh(C)∃xQx, the cumulative effect of which entails ∃xPx ∧ ∃xQx.
9
Singular indefinites
At the moment the system makes wrong predictions for embedding under singular indefinites: (81)
a. There is a bottle of beer in the fridge or the ice-bucket. b. This course is very difficult. Some student waited 3 semesters to complete it or never finished it at all.
Right now, an FC effect is expected for this case as well. However, the expectation changes once an independently needed difference between plural and singular indefinites is factored in. Consider the sentences in (82). These sentences have the indicated implicature that the alternative assertion involving quantification over plural individuals is false: (82)
a. There is a bottle of beer in the fridge. Implicature: there aren’t two bottles of beer in the fridge. b. Some student talked to Mary Implicature: It’s not true that two students talked to Mary.
This implicature leads to the conclusion that a singular indefinite is a scalar item, with a plural (or dual) indefinite as an alternative: (83)
Horn-Set(Some NP-sing) = {Some NP-sing (henceforth ∃1), two NPs (henceforth ∃2)}
Free Choice and Scalar Implicatures
107
With this Horn-Set, exh would, of course, derive the correct implicature for (82). But, interestingly, we also explain the lack of FC in (81). To see this consider (81a), and the set of alternatives derived by (5), (84). The alternative-set includes alternatives of the sort we’ve considered in (76) (upper diamond of (84)). But it also includes alternatives generated by replacing ∃1 with ∃2 (lower diamond of (84)). (84)
Alt((82)a) ∃1xP(x) ∃1x(P(x)∧Q(x)) ∃1xQ(x) ∃2xP(x)
C = ∃1x(P(x) ∨ Q(x))
∃2x(P(x) ∧ Q(x))
∃2x(P(x)∨Q(x)) ∃2xQ(x)
The set of innocently-excludable alternatives contains ∃1x (P(x)∧Q(x)) as well as ∃2x(P(x)∧Q(x)). Hence the exhaustification of (82a) is the following:44 (85)
Exh(C)( ∃1x(P(x)∨ Q(x))) = ∃1x(P(x)∨ Q(x)) & ¬∃1x(P(x)∧Q(x)) & ¬∃2x(P(x)∨ Q(x)) ⇒ ¬(∃1xP(x)∧∃1x Q(x))
This representation already contradicts the FC effect, a situation which, of course, cannot change by further exhaustification. We thus derive the generalization in (31). It is of course important to return to our computation of basic FC and make sure that nothing changes when universal quantifiers are introduced as alternatives to existential quantifiers (4b,d) (see note 40). I leave this as a task for the interested reader, though an equivalent computation will be carried out in (87) and (88), below, and the appendix will
108 Presupposition and Implicature in Compositional Semantics
contain a theorem that will make all of our results transparent with fewer computations.
10
Other FC effects
Multiple exhaustification also accounts for the generalization in (33), i.e., it generates FC effects for the sequences ¬∧ and ¬∀∧ (introduced in Sections 3.1 and 3.3.) I will illustrate this for ¬∧, and allow the reader to verify that nothing changes when is replaced with ∀. Consider (25), repeated below as (86), with its FC inference, which (to repeat) is not predicted by the basic meaning. (86)
You are not required to both clear the table and do the dishes. 1. basic meaning: ¬(p∧q) ≡ ♦¬ (p∧q) ≡ ♦(¬p∨¬q) ≡ ♦(¬p) ∨♦(¬q) 2. Free Choice: ♦(¬p) ∧♦(¬q)
Once again, FC will follow after two layers of exhaustivity are computed. Let’s start with the first layer, which we compute based on the alternatives generated by Sauerland’s Horn-Set {∧, L, R, ∧} and the traditional Horn-Set {♦, }, (8d).
(87)
Alt((86))
¬ p ¬ (p ∨ q) ¬ q ¬◊p C = ¬ (p ∧ q))
¬ ◊(p ∨ q)
¬ ◊(p ∧ q) ¬◊p
Exh(C) (¬ (p ∧ q)) = ¬ (p ∧ q)) & (p ∨ q) & ◊(p ∧ q)
Free Choice and Scalar Implicatures
109
If we decide to add another layer of exhaustification, we get the following result: (88)
Exh(C’)[Exh(C)(¬(p ∧ q)] = ¬(p ∧ q)) & (p ∨q) & ♦(p ∧ q) & ¬(¬p ∧ q) & ¬(¬q ∧ p) This yields the FC effect, based on the following equivalences: ¬(p ∧ q)) ≡ ♦¬p ∨ ♦¬q ¬(¬p ∧ q) ≡ ¬(♦¬p ∧ ¬♦¬q) ¬(¬q ∧ p) ≡ ¬(♦¬q ∧ ¬♦¬p)
11
Remaining issues
In Sections 6–10 we have seen how our two generalizations about the distribution of FC ((31) and (33)) can be derived based on recursive exhaustification under a Sauerland-inspired meaning for exh. But before concluding, I would like to discuss two apparent predictions of the account which are somewhat problematic. 11.1 ¬♦(p ∧ q) The lack of a conjunctive interpretation for p ∨ q was derived in Section 7 on the basis of the observation that the first layer of exhaustification excludes p∧q, an exclusion which cannot be overridden at the second level of exhaustification. The situation changes in FC environments, by the introduction of appropriate operators. When p∨q is embedded under an existential quantifier, e.g. ♦ (p ∨ q), the first level of exhaustification excludes ♦ (p ∧ q), a relatively weak exclusion, i.e. consistent with ♦ p ∧ ♦q. Hence it is possible (at the second level of exhaustification) to innocently exclude the exhaustive interpretation of ♦p and of ♦q. This proposal makes a clear prediction, or at least so it seems. Specifically, it predicts that FC will always be accompanied by the anticonjunctive inference ¬♦(p ∧ q). However, Simons (2005) claims that
110
Presupposition and Implicature in Compositional Semantics
this prediction is false. Specifically, she presents (89) as a sentence that can produce FC while lacking an anti-conjunctive inference. (89)
Jane may sing or dance. (Simons 2005) Possible Reading: Jane may sing and Jane may dance, compatible with permission to do both.
I think I agree with Simons judgments, and the speakers I’ve consulted with also seem to agree. Interestingly, people seem to have a different feelings when asked about the FC effect that arises for the sequence ¬∀∧. Consider, again, (86). (86)
You are not required to both clear the table and do the dishes.
It seems quite hard to get rid of the inference that you are required to either clear the table or do the dishes. More specifically, although judgments of this sort are notoriously difficult, there doesn’t seem to be an interpretation which involves FC (i.e. entails that each chore is such that you are allowed to avoid it), which does not, at the same time, entail that at least one of the chores is requited.45 I will try to explain these facts on the assumption that in (89) each of the disjuncts could be exhaustified separately. Assume, as we have assumed above, that the alternatives for Exh could be determined (at least in the absence of scalar items) based on focus. (89) might now receive the following parse, where C and C are determined based on scalar items (as outlined above), and C1 , C2 are determined based on the focus value of the relevant prejacent. (89)
Exh(C )(Exh(C )(♦(Exh(C1 )(Jane sing) or Exh(C2 )(Jane dance)))).
Assume, further, that sing and dance are focused so that C1 = C2 . To simplify the exposition (but with no loss of generality) let’s assume that C1,2 has only two members {p = Jane sing, q = Jane dance}, with the following result: p! := Exh(C1 )(p) = p ∧ ¬q q! := Exh(C2 )(q) = q ∧ ¬p As we see in (90), we derive Simons’ reading.
Free Choice and Scalar Implicatures
111
(90)
Exh(C )(Exh(C )(♦(p!∨q!))) = ♦(p!∨q!) ∧¬♦(p!∧ q!) ∧ ♦(p!) ∧♦(q!) = ♦(p!∨q!) ∧ ♦(p!) ∧ ♦(q!)
11.2
The scope of disjunction
The analysis of FC crucially depends on the assumption that in the relevant sentences disjunction receives narrow scope relative to the relevant existential quantifier. This assumption is corroborated by the contrast in (91). (91)
a. We may either eat the cake or the ice-cream. (+FC) b. Either we may eat the cake or the ice-cream. (−FC)
Larson (1985) has pointed out that either marks the scope of disjunction in constructions such as (91). The fact that FC is present only for (91a) thus corroborates the scopal assumptions made in this chapter. However, as pointed out in Zimmerman (2002), FC seems to be available in (92). (92)
You may eat the cake or you may eat the ice-cream.
I leave this as an unresolved problem, noting that the behavior with indefinites is different. (See Simons 2005, Alonso-Ovalle (2005), and Klindinst (2005) for relevant discussion.)46 (93)
12
a. Some students waited 3 semester to complete this course or never finished it at all. (+FC) b. Some students either waited 3 semester to complete this course or never finished it at all. (+FC) c. (Either) Some students waited 3 semester to complete this course or some students never finished it at all. (−FC)
Conclusion
In this chapter I’ve argued that Free Choice effects arise in two seemingly unrelated contexts: when disjunction is embedded under non-singular existential quantifiers, (31), and when conjunction is embedded under a universal quantifier which is, itself, c-commanded by negation, (33). Both FC effects follow from a method for exhaustification inspired by Sauerland’s solution for Chierchia’s puzzle, a method in which the notion of an innocently excludable alternative plays a central role.
112
Presupposition and Implicature in Compositional Semantics
However, the proposal can work only if Sauerland’s basic idea is removed from its Neo-Gricean setting. The reason for this is rather plain. Under the Neo-Gricean assumptions, SIs are derived as strengthenings of inferences that follow from NG-MQ, a maxim which would derive Ignorance Inferences based on the symmetric alternatives generated by disjunction. Hence the Neo-Gricean assumptions derive Ignorance Inferences that conflict with the empirically attested Free Choice effects. A necessary conclusion, given the alternatives for disjunction, is that NG-MQ cannot be correct, and that there can be no primary implicatures which are computed on the basis of the relevant alternatives. The conclusion, itself, is an immediate consequence of a system in which pragmatic reasoning is based on all relevant alternative assertions (B-MQ), i.e. a pragmatic system which can only derive Ignorance Inferences. If B-MQ is correct, SIs must be derived within grammar, as argued for on independent grounds in Chierchia (2004) and Fox (2004). The grammatical mechanism needed for FC seems to be an exhaustive operator, along the lines of Groenendijk and Stokhof (1984), and Krifka (1995), which can apply recursively to a single linguistic expression (based on a Sauerland-inspired lexical entry). If this is correct, it might be useful to ask questions about possible external/functional motivations for exh. I hinted at the possibility that exh is needed to solve a communication problem that will arise very often in a pragmatic universe governed by B-MQ.
Free Choice and Scalar Implicatures
113
Appendix In this appendix, I prove a rather simple theorem which should allow the reader to understand the results described in this chapter with fewer computations. I define an FC interpretation which I call AnEx (for Antiexhaustivity), and prove that this interpretation, if consistent, is the result of the 2nd layer of exhaustification. Let C be a set of propositions with p∈C. I = I-E(p,C) = Ø I’= (C\ I\{p}) = Ø AnEx = ∩{¬ExhC (q):q∈I’}∩ ExhC (p) Claim If AnEx = Ø (is consistent), Exh2 C (p) = AnEx. Proof ExhC (p) entails ¬q, for all q∈I Hence, ExhC (p) entails ¬ ExhC (q), for all q∈I Hence, AnEx entails ¬ExhC (q), for all q∈I
(by definition of Exh) (¬q entails ¬ ExhC (q)) (AnEx has ExhC (p) as a conjunct) Hence, AnEx = ∩{¬ExhC (q):q∈I’}∩{¬ExhC (q):q∈I}∩ExhC (p) = ∩{¬ExhC (q): q∈ C\ {p}} ∩ ExhC (p).
Hence, If AnEx is consistent, I-E(ExhC (p), C’) = C’ \{ExhC (p)} (where C’:= {ExhC (q):q∈C}) Hence, Exh2 C (p) = ∩{¬ExhC (q):q∈ C’ \{ExhC (p)}} ∩ ExhC (p) = AnEx (by definition of exh)
Notes ∗ This work is very much inspired by earlier proposals of Kratzer and Shimoyama,
Alonso-Ovalle, Gennaro Chierchia, an in particular by Kai von Fintel’s class presentation of’ Kratzer and Shimoyama. Special thanks go to Gennaro Chierchia, Kai von Fintel, and Irene Heim. I’ve also benefited from discussions with Luis Alonso-Ovalle, Jon Gajewski, Nathan Klindinst, Ezra Keshet, Angelika Kratrer, Fred Landsman, Philippe Schlenker, Benjamin Spector, and from comments and questions at the 2005 LSA summer institute (MIT and Harvard), and at colloquia at Umass Amherst, SUNY Stony Brook, Tel-Aviv University, and Tuebingen University.
114
Presupposition and Implicature in Compositional Semantics
1. Following Gazdar (1979) and Sauerland (2004), I will sometimes use the verb know to describe Ignorance Inferences. This choice is problematic because of factivity inferences associated with know, which are clearly inaccurate. However, it’s not clear that there is a better choice: believe is problematic because of neg-raising. When I will find factivity particularly disturbing (or neg-raising sufficiently innocuous), I will favor belief-talk. My choices will, however, be far from systematic. The reader should bear all of this in mind and ignore factivity inferences associated with know, as well as the neg-raising property of believe. 2. Although I think that there is agreement that (2c) ought to be derived from principles of communication, there have been conflicting proposals concerning the precise derivation. As we will see below, the complications could be argued to follow from the neo-Gricean perspective on SIs. See our discussion in p. 12 of Gazdar (1979) and Sauerland (2004). 3. The relevant notion of informativity for a pragmatic account should probably be that of contextual-strength (i.e. logical strength given contextual presuppositions). For the consequence of this distinction to the theory of SIs, with a potential argument against (neo-)Gricean accounts, see Fox (2004), and Magri (2005). 4. The discussion in this section relies heavily on the introduction to pragmatics taught by Kai von Fintel and Irene Heim at MIT. 5. The assumptions made here about ‘relevance’ are the following: 1. If p and q are both relevant, so is ‘p and q.’ 2. if p is relevant, so is ‘not p.’ (To say that p is relevant is to say that the question Is p true or false? relevant.) 6. It is sometimes suggested that Grice’s Maxim of Manner (M) could be used to explain s s avoidance of p. Such a suggestion requires an ordering of linguistic expressions by which p would be more optimal than p from M’s perspective. For arguments against obvious orderings (various measures of complexity), see Matsumoto (1995), as well as Fox and Hackl (forthcoming). 7. Usually called Horn-Scales for bad reasons, as discussed in Sauerland (2004). 8. This definition, which comes from Sauerland (2004), is implicit in much earlier work, and is of course very similar to the definition of alternative sets in Rooth (1985). 9. This consideration would be weakened significantly if one could make sense of Alt(S) from the perspective of a general theory of language use. For efforts along these lines, See Spector (2006). 10. For a collection of some of the arguments in favor of this covert operator, see Fox (2004) and Chierchia (2005). 11. This statement is not always true, nor is it predicted to be. Scalar Items generate alternatives, but alternatives could be specified in other ways as well: by pitch-accent or by an explicit question. See the discussion of (59) below, as well as notes 29, 32, and 46. 12. λχ:ψ(χ).φ is a function defined only for objects of which ψ is true (convention from Heim and Kratzer 1998). 13. This assumption is most natural if Alt(S) is the focus value of S, which could follow if scalar items are assumed to be inherently focused, see Krifka (1995) for a possible implementation.
Free Choice and Scalar Implicatures
115
14. In (18) the subject is reconstructed into both disjuncts, and to is omitted. This is done for expository purposes and, of course, does not affect meaning. 15. Although this chapter focuses on disjunction, the basic proposal can derive FC for the relevant indefinites as long as we assume (along the lines of Chierchia (2005)) that the relevant indefinites have the following alternatives: ALT(irgendein NP)= {irgendein NP’: NP’⊆ NP}∪{all NP’: NP’⊆ NP}. 16. I think that the interpretation is available in contexts that, more generally, allow for ‘intrusive’ implicatures of the relevant sort (See Cohen 1971, Horn 1989, Levinson 2000, Recanati 2003, among others): i. None of my students did SOME of the homework, They all did ALL of it. ii. No one is allowed to eat the cake OR the ice-cream. Everyone will be told what to eat.
17.
18. 19.
20.
21.
22.
23. 24. 25.
However, Kamp(1973) points out that it is easier for FC to ‘intrude’ into the antecedent of a conditional: If you are allowed to eat the cake or the ice-cream, you are pretty lucky. If K&S are correct, the FC interpretation of the antecedent would require an analysis involving an embedded implicature, as pointed out by Alonso-Ovalle, but without the ‘meta-linguistic’ feel that is sometimes associated with such intrusion. Unfortunately, I have nothing interesting to add. The latter possibility seems less plausible given the accumulation of evidence for ‘intrusive’ or ‘embedded’ implicatures (at least in non-downward entailing environments), see Chierchia (2004). The discussion in note 16 is of course, problematic, and more so from the Neo-Gricean perspective. It would instantiate the scheme if b and c were replaced by a single alternative, namely the disjunction of the two. Matsumoto argues that lexical items can be members of the same Horn-Set only if they denote functions of the same monotonicity. ∨ is upward monotone with respect to both arguments, but ∧¬ is downward monotone with respect to its right-hand argument. Skipping ahead to Sauerland’s Horn-set, L and R are upward (as well as downward) monotone with respect to their immaterial arguments. While working on this chapter, I have learned about two new papers about FC that make this same observation: KIindinst (2005) and Eckardt (this volume). If the distributor both is omitted the resulting interpretation is equivalent to wide scope conjunction. The correct account relies most likely on a ‘homogeneity’ presupposition (Fodor 1976, Gajewski 2005). In order to account for the difference between (32) and (34b), we would also have to say that inversion of the surface scope of conjunction and universal quantification is impossible in (34b), a consequence of Scope Economy, a principle I’ve argued for in Fox (2000). See also Lee (1995), and Simons (2002). And that the set does not include the ‘symmetric alternative’ to all, some but not all. Chierchia himself developed an account of FC which is quite similar to the account proposed here and is to some extent independent of his recursive procedure for implicature computation. Specifically, his account, like mine, is based on operators that apply to a prejacent and a set of alternatives. However,
116
26. 27. 28. 29.
30. 31.
32.
33.
34.
35.
Presupposition and Implicature in Compositional Semantics the crucial operator for him is an ‘anti-exhaustivity’ operator, distinct from what might be responsible for implicature computation. Benjamin Spector made the same proposal, in a somewhat different (more generalized) format. A related proposal can be found in Lee (1995). If x is to the left of y with a connecting line, then x is weaker than y. From now on, I will circle those alternatives for which an SI can be derived consistently with Quality and NG-MQ. Spector (2003, 2006) suggests a different perspective. Specifically, he suggests that alternative sets are defined as the closure under ∧ and ∨ of the set of positive answers to a given question. The Sauerland alternatives for John talked to Mary or Bill would, thus, be derived (along with other useful alternatives) if the relevant question was who did John talk to? What FC teaches us, if my proposal is correct, is that there is no closure under ∧. Some of what I say could work if Sauerland’s alternatives were replaced by basic answers to a Hamblin-question closed under ∨. (Conjunction in unembedded cases will be part of the basic Hamblin denotation, derived from quantification over pluralities). One would still have to make sense of second layers of exhaustivity (see Section 11.2, note 46). r := John did the reading; sh := John did some of the homework; ah := John did all of the homework. In Fox (2003) I pointed out this prediction, but was not sure about the empirical facts. I was convinced by conversations with Benjamin Spector and the discussion in Sauerland (2005). Gennaro Chierchia (p.c.) points out that in the dialogue in (53) or might be receiving contrastive focus with conjunction, with the other alternatives (L and R) inactive. This possibility will not be helpful in explaining the avoidance of a contradiction in Groenendijk and Stokhof’s examples in (55). As pointed out to me by Angelika Kratzer and Fred Landman, the proposed mechanism for exhaustification is reminiscent of what is needed for counterfactuals in the premise semantics developed by Veltman (1977) and Kratzer (1981). In particular, the set of propositions that can be added as premises to a counter factual antecedent p is ∩{A⊆C: A is a maximal set in C, s.t., A ∪ {p} is consistent} where C is the set of all true propositions. Without non-girls, the answer is somewhat strange. That’s probably because questions presuppose that at least one answer (in the Hamblin sense) is true, and, thus, without non-girls, the answer just repeats the presupposition. Adding non-girls is thus crucial, but, it is trivial to see that it will not affect the result, in any interesting way; propositions related to non-girls will be excluded and things will be more difficult to draw, but other than that, it’s all the same. In conversation with Gennaro Chierchia, we’ve noticed that things are a little more complicated. As things stand right now, Alt(∀x(P(x)∨Q(x)) contains additional members: ∃x(P(x)∨Q(x)), ∃x(P(x)), and ∃x(Q(x)). The latter two make it impossible to innocently exclude ∀xP(x) and ∀xQ(x). There are various simple ways to correct for this problem. The obvious thing that comes to mind is to define Alt(S) so that it includes only stronger sentences than S. However, this would be a problematic move given data that is not discussed in this chapter. Here’s another possibility: Alt(S) is the smallest set, s.t. (a)
Free Choice and Scalar Implicatures
117
S∈ Alt(S), and (b) If S ∈ Alt(S) and S can be derived from S by replacement of a single scalar item with an alternative, and S does not entail S , S ∈ Alt(S). 36. The core idea was developed during conversations with Ezra Keshet. 37. This should be modified to allow introduction of exh in a non-matrix position. (i) Recursive Parsing-Strategy: If a sentence S has an undesirable Ignorance Inference, try to append exh to some constituent X in S, i.e. modify the parse [S . . . X . . .] as follows: [S . . . Exh(Alt(X))(X) . . .]. We could also incorporate an economy condition of the sort alluded to in Section 2.1: (ii) Condition on exh-insertion: exh can be appended to a constituent X, only if the resulting sentence generates fewer Ignorance Inferences (based on B-MQ). 38. By the algorithm in (5). 39. The theorem in 1 is completely trivial, and the one in 2 (due to Benjamin Spector, p.c.) is less so: 1. Let C be a set of alternatives, Si , such that for each i exhaustification is trivial (i.e., Exh(C)(Si )⇔ Si ), then for each i, 2nd exhaustification is trivial (i.e., Exh(C’)(Exh(C)(Si )) ⇔ Exh(C)(Si ), where C’ = {Exh(C)(S): S∈C}) 2. due to Spector: Let C be a set of finite alternatives, Si , then there is an n∈N, s.t. ∀m>n, Exhn (C)(Si ) = Exhm (C)(Si ) Exhn (C)(Si ) := Exh(C’)Exhn−1 (C)(Si ), where C’ = {Exhn−1 (C)(S): S∈C}, and Exh1 (C)(S) = Exh(C)(S). 40. The set of alternatives is actually larger, including a variant of each alternative in C with a universal modal replacing the existential modal. This does not affect our results as the reader can verify. See the appendix, as well as (84) and (85) where a parallel computation is carried out with the full set of alternatives. 41. This exemplifies the difference between the lexical entry we are considering and the Groenendijk and Stokhof-type alternative in (56). Under (56), (75) would express a stronger proposition ♦(p ∨ q) ∧¬ (♦p∧♦q), which will be inconsistent with FC. 42. As pointed out in Simons (2005), ♦(p ∨ q) sometimes yields FC without the inference that ♦ (p∧q) is false. A solution to this problem will be discussed in Section 11.1. 43. The following is easily verifiable: Let C= {w, s, n, e} be a diamond set of alternatives going stronger from w to e (w is weaker than s, n, and e; s and n are logically independent and weaker than e), where w entails (s∨n). With such alternatives, 2nd exhaustification of w is vacuous (Exh2 (C)(w) ⇔ Exh(C)(w)) iff e ⇔ s&n. Furthermore, when 2nd exhaustification of w is not vacuous, Exh2 (C)(w) ⇔ s∧n∧¬e. 44. This explanation, however, depends on the assumption that the alternatives for ∃1 cannot be inactive when exh associates with ∨. An assumption of this sort was argued for on independent grounds in Chierchia (2005). 45. Note that both is crucial. See note 21.
118
Presupposition and Implicature in Compositional Semantics
46. FC depends on the nature of the alternatives (e.g. E must be stronger than the conjunction of N and S, see note 43, or the appendix for a more general statement). The correlation with scope is predicted on the basis of the algorithm that determines alternatives, (5): at every level of exhaustfication, alternatives are determined on the basis of the structure of the prejacent. If (92) turns out to be indicative (rather than (91) and (93)), i.e. if it turns out that FC is possible even when disjunction has scope over the relevant existential quantifies, it would be possible to capture the facts with a relatively simple modification of the system I’ve proposed. In this chapter I’ve assumed that Alt(S) is determined either contextually or by the algorithm in (5), where the algorithm was crucial for the recursive step (see note 38). However, we could define a special rule for recursive exhaustification which would allow the rule to apply even when alternatives are contextually determined. Suppose that a sentence S is uttered in a context in which Q is the salient set of alternatives. If S were to be exhaustified (relative to Q), the syntactic representation would be Exh(Q)(S). We could now define the 2nd layer of Exhaustification, as follows: Exh2 (S):= Exh(C)[Exh(Q)(S)], where C= {Exh(Q)(ϕ): ϕ ∈Q}. Now (92) could receive an FC interpretation if Q could be the set of sentences of the form You can eat x, where x denotes a singular or plural individuals (perhaps with closure under disjunction, see note 29).
References Aloni, M. & R. van Rooij (forthcoming) ‘Free Choice Items and Alternatives’, Proceedings of KNAW Academy Colloquium: Cognitive Foundations of Interpretation. Alonso-Ovalle, L. (2005) ‘Disjunction in a Modal Space’, handout of a paper presented at NYU Polarity Workshop, http://www.people.umass.edu/luisalo/ alonso_ovalle_NYU_handout.pdf Cohen, L.J. (1971) ‘The Logical Particles of Natural Language’, in Y. Bar-Hillel (ed.), Pragmatics of Natural Language, 56–68. Dordrecht: D. Reidel. Chierchia, G. (2004) ‘Scalar Implicatures, Polarity Phenomena, and the Syntax/ Pragmatics Interface’, in A. Belletti (ed.), Structures and Beyond. Oxford: Oxford University Press. Chierchia, G. (2005) ‘Broaden your Views: Implicatures of Domain Widening and the “Logicality” of Language’, unpublished ms. University of MilanBicocca/Harvard University. Fodor, J. D. (1970) ‘Linguistic Description of Opaque Contexts’, MIT Dissertation. Fox, D. (2000) Economy and Semantic Interpretation, Linguistic Inquiry Monographs, 35, Cambridge, MA: MITWPL and MIT Press. Fox, D. (2003) ‘Implicature Calculation, Only, and Lumping: Another Look at the Puzzle of Disjunction’, handout, Yale University, http://web.mit.edu/linguistics/ www/fox/Yale.pdf. Fox, D. (2004) ‘Back to the Theory of Implicatures, Class 4 of Implicatures and Exhaustivity’, handouts from a class taught at USC, http://mit.edu/linguistics/ www/fox/pdf/class_4.pdf Fox, D. and M. Hacke (forthcoming) ‘The Universal Density of Measurement’, Linguistics and philosophy.
Free Choice and Scalar Implicatures
119
Gajewski, J., (2005) ‘Neg-Raising, Polarity and Presupposition’, MIT Dissertation. Gazdar (1979) Pragmatics: Implicature, Presupposition, and Logical Form. New York: Academic Press. Grice, P. (1975) ‘Logic and Conversation’, in Cole, P. and J. Morgan (eds), Syntax and Semantics 3: Speech Acts. New York: Academic Press. Groenendijk, G., and Stokhof, M. (1984) ‘Studies on the Semantics of Questions and the Pragmatics of Answers’, PhD dissertation, University of Amsterdam. Hamblin, C. (1973) ‘Questions in Montague Grammar’, Foundations of Language 10: 41–53. Heim, I. and A. Kratzer (1998) Semantics in Generative Grammar. Malden, MA: Blackwell. Horn, L. (1969) ‘A Presuppositional Analysis of Only and Even’, CLS 5: 97–108. Horn, L. (1972) ‘On the Semantic Properties of Logical Operators in English’, PhD thesis, UCLA. Horn, L. (1989) A Natural History of Negation. Chicago: University of Chicago Press. Klindinst (2005) ‘Plurals, Modals, and Conjunctive Disjunction’, handout from SuB 10. Kratzer, A. (1981) ‘Partition and Revision: the Semantics of Counterfactuals’, Journal of Philosophical Logic 10: 242–58. Kratzer, A. and Shimoyama, J. (2002) ‘Indeterminate Pronouns: The View from Japanese’, in Y. Otsu (ed.), Proceedings of the Third Tokyo Conference on Psycholinguistics. Tokyo: Hituzi Syobo. Krifka, M. (1995) ‘The Semantics and Pragmatics of Polarity Items’, Linguistic Analysis 25: 209–57. Kamp (1973) ‘Free Choice Permission’, Proceedings of the Aristotelian Society, vol. 74, pp. 57–74. Kroch, A. (1972) ‘Lexical and inferred meanings for some time adverbs’, Quarterly Progress Report of the Research Laboratory of Electronics 104. MIT. Landman, F. (1998) ‘Plurals and Maximalization’, in S. Rothstein (ed.), Events and Grammar. Dordrecht: Kluwer. Larson R. K. (1985) ‘On the Syntax of Disjunction Scope’, Natural Language and Linguistic Theory 3: 217–64. Lee, Y. S. (1995) ‘Scalar Information and Semantics of Focus Operators’, PhD dissertation, U. T. Austin. Levinson, S. (2000) Presumptive Meanings. Cambridge, Mass: MIT Press. June 2003. Magri (2005) ‘Constraints on the Readings of Bare Plural Subjects: Syntax or Semantics’, ms MIT. Matsumoto, Y. (1995) ‘The Conversational Condition on Horn Scales’, Linguistics and Philosophy 18: 21–60. Recanati, F. (2003) ‘Embedded Implicatures’, ms, Institut Jean Nicod, Paris. Rooth, M. (1985) ‘Association with Focus’, PhD dissertation, University of Massachusetts, Amherst. van Rooij, R. (2002) ‘Relevance Only’. Proceedings of Edilog. van Rooij, R., and Schulz, K. (2004) ‘Exhaustive Interpretation of Complex Sentences’, Journal of Logic, Language and Information 13: 491–519.
120
Presupposition and Implicature in Compositional Semantics
Saeboe, Kjell Johan (2004) ‘Optimal Interpretation of Permission Sentences, in Asatiani et al. (eds), Proceedings of the 5th Tbilisi Symposium on Language, Logic and Computation, Amsterdam and Tbilisi, 137–44. Sauerland, U. (2004) ‘Scalar Implicatures in Complex Sentences’, Linguistics and Philosophy, 27: 367–91. Sauerland, U. (2005) ‘The Epistemic Step’, talk presented at Experimental Pragmatics, Cambridge University, Cambridge, UK, April 2005. Schulz, K. and van Rooij (2006) ‘Pragmatic Meaning and Non-monotonic Reasoning: The Case of Exhaustive Interpretation’, Linguistics and Philosophy, 29:2, 205–50. Simons (2000) Issues in the Semantics and Pragmatics of Disjunction, NY: Garland Publishing. Simons (2005) Dividing Things Up: the Semantics of or and the Modal/or Interaction, Natural Language Semantics 13: 271–316. Spector, B. (2003) ‘Scalar Implicatures: Local or Global?’, paper presented at the workshop ‘Polarity, Scalar Phenomena, and Implicatures’, University of Milan – Bicocca. Spector, B. (2006) ‘Aspects de la pragmatique des opérateurs logiques’, PhD thesis, University of Paris 7. Veltman F. (1976) ‘Prejudices, Presuppositions and the Theory of Counterfactuals’, in Groenendijk, J. and Stokhof, M. (eds), Amsterdam Papers in Formal Grammar. Proceedings of the 1st Amsterdam Colloquium, pp. 248–81. University of Amsterdam. Zimmermann, E. (2000) ‘Free Choice Disjunction and Epistemic Possibility’, Natural Language Semantics 8: 255–90.
5 Partial Variables and Specificity Gerhard Jäger University of Bielefeld
In this chapter I propose a novel analysis of the semantics of specific indefinites. Following standard DRT, I assume that indefinites introduce a free variable into the logical representation, but I assume the the descriptive content of an indefinite DP is interpreted as a precondition for the corresponding variable to denote. Formally this is implemented as an extension of classical predicate logic with partial variables – variables that come with a restriction. This leads to a reconception of restricted quantification: the restriction is tied to the variable, not to the quantifier. After an overview over the major existing theories of the scope of indefinites, the central part of the chapter is devoted to develop a model-theoretic semantics for this extension of predicate logic. Finally the chapter argues that the notion of partial variables lends itself to the analysis of other linguistic phenomena as well. Especially presuppositions can be analyzed as restrictions on variables in a natural way.
1
Introduction
This chapter deals with the peculiar scope taking properties of indefinite DPs, which differ massively from other scope bearing elements. The theory that I am going to propose can be seen as a variant of the DRT approach in the version of Heim (1982), according to which the semantic contribution of an indefinite is basically a free variable, while its scope is determined by a non-lexical operation of existential closure. The crucial innovation lies in the treatment of the descriptive material of indefinites. While DRT analyzes it as part of the truth conditions, I will argue that it is to be considered as a precondition for the accompanying variable to denote. Existential closure serves a double 121
122
Presupposition and Implicature in Compositional Semantics
function in my analysis: it binds a variable, and it turns its definedness conditions into truth conditions. While the basic idea of my analysis is taken from DRT, it also adopts crucial features of two other current approaches to the scope of indefinites, Tanya Reinhart’s choice function approach (Reinhart (1997), see also Winter (1997) and Kratzer (1998)), and the analysis of specificity as a presuppositional phenomenon that was independently proposed by several authors in recent years (among others by Cresti (1995); Reniers (1997); van Geenhoven (1998); Krifka (1998); Yeom (1998); Geurts (1999b)). So my proposal can be seen as an attempt to a synthesis. The structure of the chapter is as follows: in Section 2 I will briefly review the empirical phenomena we are going to deal with. Section 3 discusses the most important proposals from the literature. In Section 4 the idea of indefiniteness and partial interpretation will be presented, and I will show how this strategy overcomes the main problems that were discussed in the preceding sections. Section 5 explores the connection between partial variables and presuppositions, and Section 6 summarizes the findings.
2
Specificity and scope
The scope of quantifiers is usually clause bounded. This is illustrated in the following examples: (1)
a. If most experiments succeed, Jones will be happy. b. Jones thanked the student that had conducted experiments.
most
In (la) the quantifier most experiments is embedded in an if-clause. Its scope is restricted to this clause; there is no reading according to which it holds for most experiments x that Jones would be happy if x succeeds. Likewise, the quantifier is embedded in a relative clause in (b) and cannot take scope over the matrix clause. Most quantifiers – actually all strong quantifiers, as well as various weak quantifiers like exactly three students or at most three students – behave in a way similar to most experiments in this respect.1 Singular indefinites like an experiment and plain cardinal quantifiers like three experiments are exceptional though. They can escape scope islands. (2)
a. If an experiment succeeds, Jones will be happy. b. If three experiments succeed, Jones will be happy.
Partial Variables and Specificity 123
(2a) has a reading according to which there is a certain experiment whose success would make Jones happy, and likewise for (b), where there are three specific experiments. The same applies to the example with a relative clause. (3a) has a reading where there is a specific experiment such that Jones thanked the students who had conducted it, and analogously for (b). (3)
a. Jones thanked the students that had conducted an experiment. b. Jones thanked the students that had conducted three experiments.
These facts were first discussed in Fodor and Sag (1982). They proposed that indefinites are ambiguous between a quantificational and a referential reading. According to them, apparent wide scope indefinites are not scope bearing at all, and the illusion of global scope arises through their referential interpretation. However, already Farkas (1981) showed that indefinites can take scope outside their local clause without having global scope, and the same point was made again in Abusch (1994). (4)
a. Every writer overheard the rumor that she didn’t write a book she wrote. (∀ > ∃ > ¬) (from Diesing (1992)) b. Every professor got a headache whenever there was a student he hated in class. (∀ > ∃ > whenever) (from Abusch (1994))
In the most prominent reading of (4a), the indefinite a book she wrote takes narrow scope with respect to every writer, but wide scope with respect to the rumor and the negation. In (b) a student he hated has scope over the matrix clause even though it is embedded in a when-clause, but it is outscoped by every professor. These intermediate scope readings do not depend on the presence of a bound pronoun inside the indefinite in question (as the previous examples might suggest). In (5) we have a reading where a boy is outscoped by every town, but outscopes every girl. (5)
In every town, every girl that a boy was in love with married an Albanian. (∀ > ∃ > ∀ > ∃) (from Cresti (1995))
It has occasionally been suggested (most recently by Dekker (2002); the idea can be traced back to Hintikka (1986), who confines this approach to indefinites modified by certain though) that these readings are referential
124
Presupposition and Implicature in Compositional Semantics
despite appearance to the contrary. According to this view, referential indefinites can refer either to individuals or to Skolem functions, and the latter option leads to the illusion of non-global scope. Applied to (5) this means that a boy refers to a specific function f from towns to boys, and the relevant reading can be paraphrased as In every town t, every girl that f (t) was in love with married an Albanian. The intermediate scope reading is true if and only if there is a boy-valued function f which makes this paraphrase true. However, this strategy does not extend to cases where an indefinite with non-local scope is outscoped by a downward monotonic quantifier.2 Consider (6)
(Some centuries ago, many countries were monarchies, and you could cause a major crisis or even a civil war just by killing the king. Nowadays most countries are republics.) Less than one out of fifty states will cease to function if a particular person is killed.
In the most prominent reading of the last sentence, the indefinite a particular person takes scope over the if-clause, but is outscoped by less than one out of fifty states. Suppose the indefinite has a Skolem function reading. This means that a particular person refers to a Skolem function from states to persons, and the sentence can be paraphrased as Less than one out of fifty states s will cease to function if f(s) is killed. Now suppose that each state would cease to function if the head of the secret police gets killed, while the presidents and prime ministers are entirely unimportant. In this scenario, (6) would be false in the intermediate scope reading. However, there is a Skolem function f , mapping each state to its president, that would make the above paraphrase true. So the Skolem function reading is certainly not equivalent to the intermediate scope reading. In the above example, intuitions are fairly robust that the intermediate scope reading exists. The crucial point here is that specificity cannot be reduced to Skolem function readings in the general case. This is not to deny that indefinites can range over Skolem functions, as for instance in (7)
Most professors recommend a certain book, namely their PhD thesis.
The prominent reading of this sentence can be paraphrased as ‘There is a book valued Skolem function f , and most professors x recommend f (x), and f is the function that maps people to their PhD thesis.’ So Skolem
Partial Variables and Specificity 125
function readings are possible, but this issue is orthogonal to the problem of specificity, and I will not consider it any further in this chapter.3 From this kind of consideration, Chierchia (2001) and Schwarz (2001), inter alia, concluded that the readings in question are genuine scope readings that cannot be reduced to referentiality. At this point a note on terminology is necessary: indefinites with exceptional scoping behavior are sometimes called ‘specific’, and I will follow this convention. Some authors, however, assume that specificity is more than just non-local scope. It has been suggested that the speaker has to have a particular individual ‘in mind’ which the indefinite in question refers to, or that some other relation of cognitive contact holds between the denotation of the indefinite and some agent. I leave these issues open as orthogonal to the problems I am concerned with in this article. For the time being, when I use the term ‘specific’, it simply means ‘having non-local scope’.
3
Solution strategies
The discussion of the last example shows that a standard scoping mechanism like Quantifier Raising or Cooper storage is not sufficient to account for the peculiarities of specific indefinites. Even if the empirical problem pertaining to the collective/distributive distinction could be overcome within such a framework, we would still be left with the conclusion that there are two versions of the scoping mechanism, one of which is clause bounded and the other unrestricted. This would be a conceptually unsatisfactory state of affairs. Alternatively one might conjecture that the scope of specific indefinites is not derived via the standard scoping mechanism but via some entirely different process. This assumption is only conceptually attractive though if this alternative process can be independently motivated and is not just stipulated for the analysis of specificity. The literature contains three families of proposals for unbound scoping mechanisms that are applicable to specificity. In this section I will briefly review them and point out their merits as well as their disadvantages. 3.1
Scope by existential closure
The literature on Discourse Representation Theory (most notably Heim (1982) and Kamp and Reyle (1993)), contains plenty of evidence that indefinites are non-quantificational. According to this theory, indefinites introduce a free variable into the semantic representation. This variable can be bound by some super-ordinate unselective binder (like an adverb of quantification). If the variable is not bound by such an
126
Presupposition and Implicature in Compositional Semantics
operator, a default operation of existential closure applies that amounts to existential quantification over the else free variable. Existential closure in the sense of DRT is an obvious candidate for a mechanism to assign scope to indefinites. However, it leads to incorrect predictions if the indefinite has a non-trivial descriptive content. The following example from Reinhart (1995) illustrates this point. (8)
a. If we invite some philosopher, Max will be offended. b. ∃x((PHILOSOPHER’(x) ∧ INVITE’(WE’, x)) → OFFENDED’(MAX’))
Analyzing (8a) in a DRT-style way without employing any further scoping mechanisms leads to a semantic representation like (8b) for the specific reading of (a), where the existential impact of the indefinite some philosopher takes wide scope while the descriptive content remains in the antecedent of the conditional. As already observed in Heim (1982) for an analogous example, (8b) does not represent the truth conditions of the specific reading of (8a). The former is true if there is one non-philosopher, while (8a) in the wide-scope reading requires the existence of a philosopher x with the property that Max will be offended if we invite x. Since the existence of the non-philosopher Donald Duck is sufficient to verify (8b) but not (8a), this problem is sometimes called the Donald Duck problem in the literature. 3.2
Indefinites and choice functions
To overcome this and related problems, several authors proposed to employ choice functions for the analysis of indefinites (see for instance Reinhart (1992, 1995, 1997), Winter (1997), Kratzer (1998)). To cut a long story short, according to these theories, the semantic counterpart of an indefinite determiner is a variable over a choice function, i.e. a function that maps non-empty sets to one of their elements. This variable is subject to existential closure in a way akin to the treatment of free individual variables in DRT. (8a) would therefore come out as (9)
∃f (CH(f ) ∧ (INVITE’(WE’, f (PHILOSOPHER’)) → OFFENDED’(MAX’)))
The extension of the predicate constant CH is the set of choice function of type e, t, e, i.e. ∀f (CH(f ) ↔ ∀P(∃x.P(x) → P(f (P))))
Partial Variables and Specificity 127
(9) in fact represents the truth conditions of (8a) in an adequate way. Generally speaking, the choice function approach solves several problems in one stroke. Since it uses unselective binding to assign scope to indefinites, it covers the fact that the scope of indefinites is structurally unrestricted. Also, the choice function mechanism makes sure that the existential impact of an indefinite is not unduly divorced from its descriptive content. (The interested reader is referred to Reinhart (1997) for details.) On the other hand, the choice function approach faces at least two serious problems. This has been pointed out at several places, for instance in Reniers (1997), Geurts (2000) and Endriss (2001). First, what happens if the extension of the descriptive content of an indefinite is empty? Consider (10)
If you push a certain red button on your computer, the electricity in the whole building will turn off.4
Suppose I’m just kidding – your computer is not connected to the power supply. There aren’t even any red buttons on your computer. It is, however, possible to switch off the electricity in the building by pushing a certain button in the basement. In this case the sentence is certainly false. The choice function approach predicts it to be true though. The extension of red button at your computer is the empty set. There are no restrictions about what kind of value a choice function returns when applied to the empty set. It could as well be the central switch in the basement, which would make the sentence true. Let us call this problem the empty set problem. The second problem arises if the descriptive content of an indefinite contains a pronoun that is bound by some super-ordinate quantifier. The following example (from Abusch (1994)) can serve to discuss this point. (11)
Every professori rewarded every student who read some book hei had recommended.
According to the choice function approach, the sentence should have a reading which can be represented as (12)
∃f (CH(f ) ∧ ∀x(PROFESSOR’(x) → ∀y(STUDENT’(y)∧ READ’(f (λz.BOOK’(z) ∧ RECOMMEND’(x, z))) → REWARD’(x, y))))
128
Presupposition and Implicature in Compositional Semantics
Suppose two professor, a and b, happened to recommend exactly the same books to their students. Then the expressions λz.BOOK’(z) ∧ RECOMMEND’(a, z) and λz.BOOK’(z) ∧ RECOMMEND’(b, z) denote the same set, and thus the terms f (λz.BOOK’(z) ∧ RECOMMEND’(a, z)) and f (λz.BOOK’(z) ∧ RECOMMEND’(b, z)) denote the same individual. So the reading that is described in (12) can be paraphrased as follows: Every professor has a favorite book. He recommends this book (possibly along with other books), and he awards students that read his favorite book. Furthermore, if two professors recommend the same books, they have the same favorite book. It might be argued that this reading is actually there but hard to detect, because it is logically stronger than the ordinary narrow-scope reading (12). This is not the case anymore though if we use a downward monotonic quantifier in subject position, as in (13)
At most three girls visited a boy that they fancied.
According to the choice function approach, this sentence should have the reading given in (14a), which – ignoring the empty set problem for the moment – is truth-conditionally equivalent to (14b).5 (14)
a. ∃f .CH(f )∧|λx.GIRL’(x)∧VISIT’(x, (f (λy.BOY’(y)∧FANCY’ (x, y))))| ≤ 3 b. |λx.GIRL’(x) ∧ ∀y(BOY’(y) ∧ FANCY’(x, y) → VISIT’(x, y))| ≤ 3
So the prediction is that the sentence has a reading that is synonymous to At most three girls visited every boy that they fancied. Intuitions are fairly solid here that such a reading does not exist. This problem is discussed among others by Kratzer (1998) (who attributes the observation to Kai von Fintel and P. Casalegno) and Winter (1997). Kratzer proposes to admit Skolemized choice functions to solve it. A Kratzer style translation of (14a) would thus come out as (15)
∃f .∀zCH(fz )∧|λx.GIRL’(x)∧VISIT’(x, (fx (λy.BOY’(y)∧FANCY’ (x, y))))| ≤ 3
Partial Variables and Specificity 129
This, however, is also equivalent to (14b) and thus empirically incorrect. This problem has already be noted by Reniers (1997). One might object that the equivalence between (15) and (14b) rests on the assumption that existential quantification over Skolemized choice function is unrestricted, while it should be restricted to functions that are somehow natural. However, example (6) above shows that unrestricted existential quantification with non-local scope does exist. Given this, it would be somewhat ad hoc to admit this option at intermediate scope positions but to exclude it at the global level (an option that is explored by Chierchia (2001)). Intuitively, this bound pronoun problem in connection with the choice function approach is similar to to the Donald Duck problem of the unselective binding approach: in both approaches, the interpretation of the descriptive content of an indefinite is separated from its existential impact, while these two semantic components of indefinites always occur in tandem. 3.3
Specificity as presupposition accommodation
Specificity shares several features with presupposition accommodation, and it is thus tempting to unify these two phenomena. This idea has been suggested at several place in the literature. I am aware of the following sources (but there may be more): Cresti (1995), Reniers (1997), van Geenhoven (1998), Krifka (1998), Yeom (1998), and Geurts (1999b). In this section, I review the main arguments for and against this approach. It turns out that the main stumbling block is the fact that presuppositions appear to be able to be accommodated into the restrictor of operators, while specific indefinites cannot take scope at these positions. Therefore I will re-examine the case for accommodation into restrictors and conclude that the relevant constructions should better be analyzed as instances of presupposition binding by a covert restrictor – an option that is not available for specifics for independent reasons. Consider a classical presupposition trigger like the king. If it occurs in a context where its presupposition ‘There is a king’ is not entailed, this information is accommodated. Accommodation into the global context leads to truth conditions involving a wide scope existential quantification, and likewise intermediate and local accommodation correspond to intermediate and local scope respectively. The mentioned authors agree in the assumption that specific indefinites also trigger an existential presupposition, and their wide/intermediate scope arises via accommodation. The analogy between (undisputed) presupposition triggers and specifics is in fact striking. To start with, if our example trigger the king
130
Presupposition and Implicature in Compositional Semantics
occurs within a scope island, its presupposition is nonetheless preferably accommodated at the global level. The preferred reading of (16a) (if uttered in a context where the existence of a king is not established yet) is as in (b). (16)
a. Every Italian watched a film that showed the king in his childhood. b. = There is a (salient?) kingi and every Italian watched a film that showed himi in hisi childhood.
This is paralleled by the preference of specific indefinites for global scope, as illustrated in (17a), which is preferably interpreted as in (b). (17)
a. Every Italian watched a program that showed a certain diva in her youth. b. = There is a certain divai and every Italian watched a program that showed heri in heri youth.
As discussed in detail in van der Sandt (1992) and Geurts (1999a), the choice between different versions of accommodation is subject to a variety of constraints. Preference for high over low accommodation is the most obvious, but not the strongest one. It is outranked by the so-called ‘trapping constraint’, which is an inviolable constraint. It requires that a bound variable inside a presupposition cannot become unbound via accommodation. Paired with the preference for high accommodation, the prediction is that a presupposition containing a bound variable is preferably accommodated directly below the binder of this variable. This is illustrated in (18). The definite DP her i boyfriend triggers the presupposition shei has a boyfriend. Since heri is bound by each girl, accommodation cannot be global, and hence the preferred reading is the intermediate one in (18b). (18)
a. Each girli claimed that shei had visited heri boyfriend. b. = Each girl has a boyfriend and claimed that she had visited him. c. = There is a boyfriend that each girl claimed to have visited.
The trapping constraint is also operative in connection with specifics. Global scope is blocked in (19) because then shei would not be bound by each girl anymore. Accordingly, the intermediate scope reading is preferred.
Partial Variables and Specificity 131
(19)
a. Each girli claimed that shei had visited a certain boy shei fancied. b. = Each girl fancies a boy and claimed that she had visited him. c. = There is a boy that each girl claimed to have visited.
Last but not least, accommodation is constrained by a variety of pragmatic conditions pertaining to local and global informativity and consistency. For instance, ‘local informativity’ requires that accommodation must not result in a reading where semantic sub-structures become redundant. Consider (20a). The definite the king of France triggers the presupposition that there is a king of France. Accommodating this at the global level (as indicated in (c)), however, results in a structure where the if-clause of the conditional is redundant. Therefore intermediate accommodation as in (b) is the preferred option. (20)
a. If France is a monarchy, then probably the king of France is bald. b. = If France is a monarchy, there is a king of France and he is probably bald. c. = There is a king of France, and if France is a monarchy, he is probably bald.
Again the same applies ceteris paribus to specific indefinites. (21)
a. If John is not a single child, probably a certain sibling of him will inherit his house. b. = If John is not a single child, he has siblings and probably one of them will inherit his house. c. = John has a sibling and if he is not a single child, probably this sibling will inherit his house.
This approach has a lot of appeal. It relates the non-local behavior of specifics to an independently established mechanism for the derivation of non-local scope, and it offers a principled account of the interaction of specificity with pronoun binding. On the other hand though, there is at least one conceptual and one technical problem with the equation ‘specificity = accommodation.’ First and foremost, presupposition accommodation is usually conceived as a repair strategy, which applies if a presupposition cannot be anchored in the local context where the trigger occurs. Even in
132
Presupposition and Implicature in Compositional Semantics
a theory like van der Sandt’s, where accommodation is a fully legitimate interpretation mechanism, it is only applicable if presupposition binding fails. For instance in (22a), the presupposition ‘There is a man’ that is triggered by the man is preferably bound by the indefinite a man in the if-clause (provided the context of utterance does not supply an alternative antecedent). This results in a reading equivalent to (b). Accommodation would lead to reading (c), but this option is not viable here. (22)
a. If a man walks, the man talks. b. = If a mani walks, hei talks. c. = There is a mank , and if a mani walks, hek talks.
If specific indefinites were ordinary presupposition triggers, we would expect that they can be bound as well. This is impossible, however, (23a) cannot mean (b). (23)
a. If a man walks, a (certain) man talks. b. = If a mani walks, hei talks.
This problem is perhaps less serious than it seems. Krifka (1998) suggests that a sentence like (23a) cannot have the reading (23b) precisely because (22) has this interpretation. More generally, he assumes that the option to use a definite DP to express a certain meaning blocks the competing option to use a specific indefinite. Van Geenhoven and Geurts suggest that specifics cannot have an anaphoric reading because they are indefinite and therefore subject to Heim’s ‘Novelty Condition.’ Perhaps, these two alternative explanations are even compatible – if Krifka is right, it might be possible to reduce the Novelty Condition altogether to a blocking effect. I will leave this issue open here. The other problem for the specificity-as-accommodation theories has to do with specifics that occur in the nuclear scope of quantificational or conditional constructions. Take the following example (due to van Geenhoven): (24)
a. Every singeri chose a song hei practiced the day before. ≡ b. Every singeri chose every song hei practiced the day before.
If the indefinite a song he practiced the day before gets a specific construal, it should be accommodated. Global accommodation is excluded because then the bound pronoun he would end up outside the scope
Partial Variables and Specificity 133
of its binder every singer, and this would violate the trapping constraint. According to van der Sandt’s theory, the preferred option should actually be accommodation into the restrictor of the quantifier every singer. This would lead to the reading (24b). However, (a) cannot have this reading. The only possible reading is the one where the indefinite has narrow scope. Krifka argues that accommodation of indefinites into the restrictor of quantificational operators does in fact occur. His example is (25)
a. A green-eyed dog is usually intelligent. b. USUALLY’[ ][INTELLIGENT’(x){GREENEYED_DOG’(x)}] c. USUALLY’x [GREENEYED_DOG’(x)][INTELLIGENT’(x)]
(25b) sketches the semantic interpretation of (a) (in a DRT-like fashion) before accommodation takes place. Presupposed material is indicated by curly brackets. Accommodation of the presupposition of the nuclear scope into the restrictor plus unselective variable binding leads to the representation in (c), which in fact accurately represents the interpretation of (a). Krifka generally assumes that the generic interpretation of indefinites is a consequence of accommodation into the restrictor of some operator. This idea is tempting, because it unifies the two ‘strong’ readings of indefinites – specific and generic – in a natural way. However, I think that this proposal is not really viable in the end. First, (24) shows that accommodation into the restrictor is not always possible, even if the overall laws of presupposition accommodation actually seems to enforce it. Quite generally, strong (i.e. specific or generic) indefinites never wind up in the restrictor of an adnominal quantifier. Genericity is confined to adverbial quantification (including the silent GEN operator). Furthermore, genericity is a clause bound phenomenon. This can be seen from the following examples. (26) (27)
a. b. a. b.
Bill always invites a girl to a restaurant. ≡ ∀x [GIRL’(x)][INVITE’ (BILL’, x)] Bill is always jealous because John invites a girl to a restaurant. = ∀x [GIRL’(x)][BECAUSE’(JEALOUS’(BILL’), INVITE’(JOHN’,x))]
In (26), the adverb always creates a tripartite structure, and the indefinite a girl is generic and supplies the restrictor of the adverb. Accordingly, the truth conditions of (a) are as in (b). In (27), the structure is analogous
134
Presupposition and Implicature in Compositional Semantics
except that a girl is now embedded in an adjunct clause. Here the indefinite from the embedded clause cannot contribute to the restrictor of the matrix clause. The indefinite can only have the specific (i.e. global scope) or the narrow scope reading. If the generic interpretation of indefinites arose via accommodation, such a sensitivity to syntactic structure would be unexpected. I thus conclude that genericity is an independent phenomenon that can neither be related to specificity nor to presupposition accommodation. If this is so, then accommodation of specifics into the restrictor of any operators is excluded. Is this an argument against the subsumption of specifics under presuppositions? Not necessarily. The question whether accommodating presuppositions into the restrictor is possible at all has been the topic of long debates (see for instance the discussion in Beaver (2001)). Some examples clearly suggest that it is possible, see for instance (28a). Its preferred reading is (b), the result of (intermediate) accommodation of the presupposition of the nuclear scope of the quantifier into the restrictor.
(28)
a. All Californians quit smoking. b. All Californians who used to smoke quit smoking.
For other examples, intermediate accommodation is impossible. (29) is due to Beaver:
(29)
a. ???Few of the 15 team members and none of the 5 cheerleaders can drive, but every team member will come to the match with her car. b. Few of the 15 team members and none of the 5 cheerleaders can drive, but every team member that owns a car will come to the match with her car. c. (???)Few of the 15 team members and none of the 5 cheerleaders can drive, but every team has a car and will come to the match with her car.
Applying intermediate accommodation to (a) yields (b), which is absolutely coherent. However, the only possible reading of (a) is (c) – the result of local accommodation – which is contradictory and thus pragmatically deviant.
Partial Variables and Specificity 135
The same point can be made with the following sentence, which is a variant of an example due to Hans Kamp: (30)
a. On Saturdays, every German housewife washes her Porsche. b. Interm. acc: Every German housewife that owns a Porsche washes it on Saturdays. c. Local acc.: Every German housewife owns a Porsche and washes it on Saturdays.
The (a)-sentence is perceived as unambiguous and entailing that every German housewife owns a Porsche. This corresponds to local accommodation, while intermediate accommodation (sketched in (b)) is no possible reading of (a). Let us have another look at (28) now. It is a standard assumption that the domain of a quantifier is not the entire universe, but that it is contextually restricted. So an adequate meaning representation of (28a) would be (31)
∀x [CALIFORNIAN’(x) ∧ C(x)][QUIT_SMOKING’(x){SMOKED’(x)}]
The value of the domain variable C has to be contextually supplied. Suppose it is instantiated as used to smoke. This would lead to the representation in (32a). Now the presupposition of the nuclear scope can be bound by material in the restrictor, which leads to the fully specified meaning (32b) for (28). (32)
a. ∀x [CALIFORNIAN’(x) ∧ SMOKED’(x)][QUIT_SMOKING’(x){SMOKED’(x)}] b. ∀x [CALIFORNIAN’(x) ∧ SMOKED’(x)][QUIT_SMOKING’(x)]
In general terms, what looks like accommodation into the restrictor can also be analyzed as an instance of presupposition binding, provided the implicit domain restriction of the operator in question supplies an antecedent for the presupposition. It is still surprising that this interpretation strategy is possible in some cases as in (28) while in other cases like (29) or (30) it is impossible. I will not resolve this puzzle, but apparently there are various factors involved here. Arguably the implicit restriction of a domain of quantification from Californians to smoking Californians is more natural from a pragmatic point of view than the shift from German housewives to Porsche-owning German housewives.
136
Presupposition and Implicature in Compositional Semantics
Also, information structure plays an important role. Compare (30) to (33)
Every German housewife washes her Porsche on SATURDAYS.
Here her Porsche is backgrounded, and intermediate accommodation appears to be much easier to get. Whatever other factors might be involved here, for the present discussion it is only important that strictly speaking there is no operation of accommodation into the restrictor of operators. What looks like it is really an instance of presupposition binding, and specifics cannot be bound due to the Novelty Condition. Given this, the inability of specifics to be accommodated into the restrictor is entirely regular. So it seems that the analysis of specificity as presupposition accommodation is in fact viable. In the remainder of this chapter I will develop an analysis of specificity which can be seen as an implementation of this approach. It has the advantage of being compositional, and it offers a simple explanation for the asymmetry between presupposition binding and presupposition accommodation that I argued for above: Binding from the restrictor into the nuclear scope of binary operators is possible, but accommodation into the restrictor is not.
4
Partial variables
4.1
The basic idea
In this section I will lay out the basic idea of my own proposal. Consider a simple sentence like (34) which contains an indefinite description. (34)
A cup moved.
Following the choice function approach, I assume that the indefinite a cup is an individual-denoting term with the type e. Intuitively, it simply refers to some cup. On the other hand, I follow DRT in the assumption that an indefinite is like a variable in predicate logic, i.e. its denotation is underdetermined and has to be fixed by the context. This can straightforwardly be implemented by means of an assignment function. So in the semantic representation a cup is to be translated as a variable (the choice of the name of the variable need not concern us here – let us assume that DPs have referential indices to this effect), but with the side condition that the denotation of this variable has to fall under the extension of the
Partial Variables and Specificity 137
predicate cup. If the local assignment function maps this variable to a non-cup, the indefinite fails to denote. This idea is implemented by extending the language of first order logic with the notion of a partial variable. This is a variable that comes with a restriction, i.e. a condition that has to be fulfilled to let the variable have an interpretation. The syntactic convention is simple: Definition 1 (Partial variables – syntax) If x is a variable and ϕ is a formula, then [x|ϕ] is a term (called a partial variable). Partial variables are interpreted via assignment functions, provided the restriction is fulfilled. Definition 2 (Partial variables – semantics (first version)) g(x) iff ϕg = 1 [x|ϕ]g = undefined else Using partial variables thus leads to a semantics with denotation gaps. Unless otherwise stated, the meaning of a complex expression is undefined if one of its subexpressions is undefined. Definedness conditions are thus inherited from partial variables to expressions containing them. There are two exception to this principle – the rules for the two quantifiers. Definition 3 (Semantics of complex expressions (first version)) 1 iff for some a : ϕg[a/x] = 1 ∃xϕg = 0 else 1 iff for all a : ϕg[a/x] = 1 ∀xϕg = 0 else If α is a complex expression but not a partial variable or a quantified formula, and the immediate subexpressions of α are β1 , . . . , βn , then αg is defined iff βi g is defined for all i with 1 ≤ i ≤ n. If it is defined, it is determined by the rules of standard predicate logic. The clauses for quantification are identical to their classical counterparts, but they are conceptually different. If the formula in the scope of existential quantifier is false or undefined for all values of the quantified variable, the formula as a whole is just false. This contrasts with other
138
Presupposition and Implicature in Compositional Semantics
operators, where undefinedness is passed on from sub-constituents to super-constituents. Let us apply these rules to the simple example from above. In the mapping from English sentences to LFs, I follow largely the version of DRT that is developed in chapter 2 of Heim (1982). The main difference is that indefinites are translated as partial variables rather than as open formulas. These variables are bound by a default operation of existential closure. Following the spirit of Reinhart’s version of the choice function approach, I assume that existential closure is selective and can freely apply at every sentential level or at the text level. Put differently, I admit the adjunction of existential quantifiers at every S-node. I represent meanings by means of first order formulas, but the back-and-forth translations between LFs and formulas should be obvious. (35)
a. A cup moved. b. MOVE’([x|CUP (x)]) c. ∃x.MOVE’([x|CUP (x)])
⎧ ⎪ ⎨ 1 iff g(x) ∈ CUP’g ∩ MOVE’g d. MOVE’([x|CUP’(x)])g = 0 iff g(x) ∈ CUP’g − MOVE’g ⎪ ⎩ undefined iff g(x) ∈ / CUP’g 1 iff CUP’g ∩ MOVE’g = ∅ e. ∃x.MOVE ([x|CUP’(x)]) = 0 else (35a) is translated as (b), where the verb corresponds to a predicate and the indefinite subject to a partial variable. The NP-part of the indefinite is mapped to the restriction of the variable. Applying existential closure yields (c). The interpretation of (c) is given in (d). Function application passes the definedness conditions of its argument on to the entire expression, so (b) only has an interpretation if the partial variable in it denotes, i.e. if x is mapped to a cup. If this is the case, the truth conditions of (b) are straightforward. Applying existential quantification to (b) leads to (c) with the interpretation (e). Here undefinedness and falsity are collapsed to falsity, and the bivalent truth conditions are as desired. Suppose (b) is interpreted in a model where there are no cups. Then the interpretation is undefined regardless of the value of x under g. Accordingly, (c) would come out as false under each assignment. There is thus no empty set problem under this approach – existential quantification over an empty set leads to falsity. Let us move on to a more complex example. In (36a), an indefinite is embedded in an if-clause. (The specific reading is perhaps not all that
Partial Variables and Specificity 139
natural here – try to imagine that there is one psychic cup which reacts to ghosts, while the other cups on the table are just ordinary china.) (36)
a. If a cup moved the ghost is present. b. ∃x(MOVE’([x|CUP’(x)]) → GHIP ) c. MOVE’([x|CUP’(x)]) → GHIP’g = ⎧ ⎪ ⎨ 1 iff g(x) ∈ CUP’g &(g(x) ∈ MOVE’g ⇒ GHIP’g = 1) ⎪ ⎩
0 iff g(x) ∈ CUP’g ∩ MOVE’g & GHIP’g = 0
undefined iff g(x) ∈ / CUP’g ⎧ ⎪ ⎨ 1 iff for some a ∈ CUP’g : d. (b)g = a ∈ MOVE’g ⇒ GHIP’g = 1 ⎪ ⎩ 0 else e. ∃x(CUP’(x) ∧ (MOVE’(x) → GHIP’)) Existential closure could be applied either to the if -clause or to the matrix clause. We are interested in the specific reading, which corresponds to the latter option. The translation of (a) under this reading is (b). The definedness condition of the partial variable is passed on to the translation of the local clause as above, and further on to the entire implication corresponding to the conditional construction. The interpretation of the implication is thus as in (c). Applying existential quantification leads to (d). (36b) thus has the same truth conditions as the wide scope restricted existential quantification in (e). However, this effect has been achieved without syntactically moving the restrictor from its in-situ position. This shows that analyzing indefinites as partial variables avoids the Donald Duck problem but nevertheless predicts that the scope of indefinites is syntactically unrestricted.
4.2
Stacked variables
The present approach overcomes the empty set problem and the Donald Duck problem, but how about the bound pronoun problem? A critical example would be (37a). Under wide scope existential closure, its translation is as in (b). (37)
a. Every girl visited a boy she fancied. b. ∃y∀x.GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])
140
Presupposition and Implicature in Compositional Semantics
What are the truth conditions of (37b)? Let us start with the consequent of the implication. (38)
VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎧ ⎪ ⎨ 1 iff g(y) ∈ BOY’g & g(x), g(y) ∈ FANCY’g ∩ VISIT’g =
⎪ ⎩
0 iff g(y) ∈ BOY’g & g(x), g(y) ∈ FANCY’g \VISIT’g undefined else
Accordingly, the implication as a whole has the interpretation (39)
GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎧ 1 iff g(y) ∈ BOY’g & g(x), g(y) ∈ FANCY’g & ⎪ ⎪ ⎪ ⎪ ⎪ g(x) ∈ GIRL’g ⇒ g(x), g(y) ∈ VISIT’g ⎪ ⎨ = 0 iff g(y) ∈ BOY’g & g(x), g(y) ∈ FANCY’g & ⎪ ⎪ ⎪ ⎪ / VISIT’g g(x) ∈ GIRL’g & g(x), g(y) ∈ ⎪ ⎪ ⎩ undefined else
Applying universal quantification brings us back into bivalent semantics. (40)
∀x.GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎧ for all a : g(y) ∈ BOY’g & a, g(y) ∈ FANCY’g & ⎪ 1 iff ⎨ = (a ∈ GIRL’g ⇒ a, g(y) ∈ VISIT’g ) ⎪ ⎩ 0 else
After existentially binding y we get (41)
∃y∀x( ⎧ GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎪ 1 iff for some b, for all a : b ∈ BOY’g & ⎪ ⎪ ⎪ ⎨ a, b ∈ FANCY’g &(a ∈ GIRL’g ⇒ = ⎪ a, b ∈ VISIT’g ) ⎪ ⎪ ⎪ ⎩ 0 else
These are the truth conditions of the classical formula (42a), which is equivalent to (b). (42)
a. ∃y∀x.BOY’(y) ∧ FANCY’(x, y) ∧ (GIRL’(x) → VISIT’(x, y)) b. ∃y.BOY’(y) ∧ ∀xFANCY’(x, y) ∧ ∀x(GIRL’(x) → VISIT’(x, y))
Partial Variables and Specificity 141
It is obvious that this has nothing to do with the intuitive truth conditions of (37). For (42) to be true, there has to be some boy who is fancied by everybody, while no such thing is entailed by (37). The problem here is that the variable x that occurs in the restriction of y is bound at a position where y is still free. However, this occurrence of x depends on y because the former is part of the restriction of the latter. The problem arises because a dependent variable is bound, which ought to be illicit. I will first explain in syntactic terms what goes wrong here, before I present a revised semantics of first order logic with partial variables. Basically, existential quantification turns a definedness condition of the bound variable into an ordinary restriction of the existential quantifier. This can be formulated as a syntactic transformation rule: ∃xϕ([x|ψ]) → ∃x.ψ ∧ ϕ(x) If the descriptive part of a specific indefinite always has the same scope as the corresponding existential quantifier (i.e. if there is no version of the Donald Duck problem whatsoever), this rule must be meaningpreserving. Under an adequate semantics of partial variables, it should thus hold that under all models and assignment functions: (43)
M ∃xϕ([x|ψ])M g = ∃x.ψ ∧ ϕ(x)g
Now reconsider formula (37b): (44)
∃y∀x.GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])
Applying the equality in (43) gives us (45)
∃y.BOY’(y) ∧ FANCY’(x, y) ∧ ∀x.GIRL’(x) → VISIT’(x, y)
Here the first occurrence of x is free. This does in fact correspond to a possible reading of (37), the one where the pronoun she is free. Given the present semantics, the step from (44) to (45) is is not meaning preserving though because there is a variable occurrence that is bound before and free after applying the transformation. To make the step meaning preserving (i.e. to ensure that (43) holds), we have to ensure that a variable occurrence x that occurs inside the restriction of some free variable y cannot be bound by a quantifier outside the restriction of the same occurrence of y.
142 Presupposition and Implicature in Compositional Semantics
We could enforce this by syntactically requiring that quantifiers only bind independent variable occurrences. This would not do the trick, however. After applying the equality in (43), dependent variable occurrence become independent and thus free to be bound. If this is a meaning preserving operation, it must be possible to bind dependent variable occurrences after all. The crucial restriction is that the scope relation between the binders and the dependency relation between the bound elements must be inverse from each other. Schematically, the following structure would represent a possible binding configuration:
Here the binding relations are nested. Crossing binding relations ought to be illicit:
Likewise, binding into the restriction of a free variable is bad:
How can the illicit configurations be avoided? I will use a procedural metaphor. When constructing a formula bottom-up, variable occurrences are arranged in stacks. A single variable creates a new stack with this variable as the only element. If a variable occurrence x is put into the restriction of a variable occurrence y, y is pushed on top of the stack containing x. This may apply recursively, so that stacks may be arbitrarily large. Binding a variable with a quantifier has the effect of popping elements from stacks. The quantifier ∃x removes all occurrences of x that are top elements of a stack from their stacks. Now all variable occurrences that directly depend on a top-level occurrence of x become top-level elements themselves and are thus free to be bound. Crucially, variable occurrences that are not top-level cannot be bound. The technical implementation of this goes as follows. Assignment functions are not just applied to single variables, but to non-empty sequences of variables – the formal counterpart of variable stacks. If the variable x occurs freely and independently in a formula, its interpretation under some assignment g is just g(x). x might as well occur within the restriction of another variable y, as for instance in [y|R(y, x)]. Then the interpretation of the occurrence of x is g(yx). This system works
Partial Variables and Specificity 143
recursively. If you have an expression like [z|S(z, [y|R(y, x)])], then the interpretation of this occurrence of x under g would be g(zyx), etc. So assignment functions do not evaluate single variable occurrences but entire stacks. The interpretation rule for partial variables has to be changed accordingly. I use the following notational convention: if g is an assignment and x a variable, then g x is the assignment with the property that for all non-empty sequences v (46)
g x (v ) = g(xv )
This is the semantic implementation of pushing x on the stack v . This operation is needed to interpret partial variables: Definition 4 (Partial variables – semantics (revised version)) g(x) iff ϕg x = 1 [x|ϕ]g = undefined else In linguistic applications, the restriction ϕ of a variable occurrence x should properly restrict the denotation of x. So x should occur freely in ϕ. However, free occurrences of x in ϕ are now one level lower in the stack than the occurrence of x they restrict. To avoid this side effect, I assume that pushing x on a stack where x is already the top element has no effect on the stack. Quantification turns the restriction of the bound variable into a part of the truth conditions. This means that variables that depend on the quantified variable are turned into independent variables. Hence we have to adjust the definition of quantification as well. We need two more auxiliary notions: • g[a/x] is the assignment which is exactly like g except that g(xn ) = a.6 This ensures that two occurrences of x where the first directly depends on the second refer to the same object. • g{x} is the assignment which is exactly like g except that for all nonempty sequences v that start with a variable = x : (47)
g{x}(xn v ) = g(v )
This implements the notion of popping all top-level occurrences of x from a stack.
144
Presupposition and Implicature in Compositional Semantics
Quantification pops top-level elements from stacks and quantifies over their denotation: Definition 5 (Semantics of Quantification (revised version)) 1 iff for some a : ϕg[a/x]{x} = 1 ∃xϕg = 0 else 1 iff for all a : ϕg[a/x]{x} = 1 ∀xϕg = 0 else It might be helpful to go through the interpretation procedure for (37) under the revised interpretation to see the difference. The formula to be interpreted is still (48)
∃y∀x.GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])
Skipping over some intermediate steps that ought to be obvious, the interpretation of the implication is (49)
GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎧ 1 iff g y (y) ∈ BOY’g & g y (x), g y (y) ∈ FANCY’g & ⎪ ⎪ ⎪ ⎪ ⎪ (g(x) ∈ GIRL’g ⇒ g(x), g(y) ∈ VISIT’g ) ⎪ ⎨ = 0 iff g y (y) ∈ BOY’g & g y (x), g y (y) ∈ FANCY’g & ⎪ ⎪ ⎪ ⎪ / VISIT’g ) (g(x) ∈ GIRL’g & g(x), g(y) ∈ ⎪ ⎪ ⎩ undefined else
Universal quantification over x yields (50)
∀x. ⎧GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g ⎪ for all a : g y (y) ∈ BOY’g & ⎪ 1 iff ⎪ ⎪ ⎨ g y (x), g y (y) ∈ FANCY’g & (a ∈ GIRL’g ⇒ = ⎪ a, g(y) ∈ VISIT’g ) ⎪ ⎪ ⎪ ⎩ 0 else
Here the shift from the assignment function g[a/x] to g[a/x]{ x } has no effect since no variables depend on x. This is different for y : (51)
∃y∀x.GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])g
Partial Variables and Specificity 145
=
⎧ ⎪ 1 iff ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎩
for some b, for all a : b ∈ BOY’g[b/y]{y}y & g[b/y]{y}y (x), b ∈ FANCY’g[b/y]{y}y & (a ∈ GIRL’g[b/y]{y} ⇒ a, b ∈ VISIT’g[b/y]{y}y )
0 else
Applying the definitions for g x (46) and for g{x} (47), we find that for arbitrary assignment functions h and (possibly empty) variable sequences v that do not start with y : h{y}y (y n v ) = h{y}(y n+1 v ) = h(v ) = h{y}(y n v ) This amounts to the fact that hy {y} = h{y} Therefore we can simplify (51) to (52)
∃y∀x. ⎧ GIRL’(x) → VISIT’(x, [y|BOY’(y) ∧ FANCY’(x, y)])_g ⎪ 1 iff for some b, for all a : b ∈ BOY’g & ⎪ ⎪ ⎨ g(x), b ∈ FANCY’g & = ⎪ (a ∈ GIRL’g ⇒ a, b ∈ VISIT’g ) ⎪ ⎪ ⎩ 0 else
These are the truth conditions of the classical formula (53)
∃y.BOY’(y) ∧ FANCY’(x, y) ∧ ∀x.GIRL’(x) → VISIT’(x, y)
This is exactly the result of the syntactic transformation briefly considered above. It represents the (possible) reading where the indefinite is specific and the pronoun in its descriptive part is free. It is NOT possible anymore to derive a reading where the indefinite has wide scope but the pronoun is bound by the universal quantifier. So the bound pronoun problem is solved as well. 4.3
One variable at a time
Still, the theory in its present shape makes wrong predictions when several indefinites interact. (54)
a. It is not the case that a certain boy met a girl. b. ∃x¬(∃y MEET’([x|BOY’(x)], [y|GIRL’(y)])
In (54) we have two indefinites, and we are interested in the reading where the first is existentially bound with wide and the second with
146
Presupposition and Implicature in Compositional Semantics
narrow scope. The interpretation of the atomic formula in the scope of the second quantifier is straightforward: (55)
MEET’[(x|BOY’(x)], [y|GIRL’(y)])g ⎧ 1 iff g x (x) ∈ BOY’g & g y (y) ∈ GIRL’g & ⎪ ⎪ ⎪ ⎪ ⎪ g(x), g(y) ∈ MEET’g ⎪ ⎨ = 0 iff g x (x) ∈ BOY’g & g y (y) ∈ GIRL’g & ⎪ ⎪ ⎪ ⎪ g(x), g(y) ∈ / MEET’g ⎪ ⎪ ⎩ undefined else
Applying existential quantification gives us (56)
∃y⎧MEET’([x|BOY’(x)], [y|GIRL’(y)])g ⎪ for some a : g x (x) ∈ BOY’g & a ∈ GIRL’g & ⎨ 1 iff = g(x), a ∈ MEET’g ⎪ ⎩ 0 else
Existential quantification turns definedness conditions into truth conditions, and it does so in a non-discriminating way. It does not distinguish to which partial variable the different parts of the definedness conditions belong. This has disastrous consequences here. Continuing with negation and a second existential quantification, we get (57)
∃x¬∃y MEET’([x|BOY’(x)], [y|GIRL’(y)])g ⎧ for some b, for no a : b ∈ BOY’g & a ∈ GIRL’g & ⎪ 1 iff ⎨ = b, a ∈ MEET’g ⎪ ⎩ 0 else
This corresponds to the classical (58)
∃x¬∃y.BOY’(x) ∧ GIRL’(y) ∧ MEET’(x, y)
The restriction of the partial variable x remains within the scope of the first quantifier c-commanding it, and this is ∃y here. This is a configuration that is similar to the Donald Duck problem discussed earlier. The formula in (58) is true iff there is an object with the property that if it is a boy, it met no girl. Of course these are not the truth conditions of (54). An adequate semantics for quantification should turn those definedness conditions that belong to the bound variable into truth conditions,
Partial Variables and Specificity 147
while the restrictions belonging to other variables should be passed on as definedness conditions. To implement this intuition, I will use a technique that is inspired by Reinhart (1997). Reinhart assumes that an indefinite with an empty restriction does refer to some special object which behaves in a nonclassical way. In Reinhart’s account, this special object has no properties, so to speak. I will also assume a special object as the referent of such indefinites, but one that is hyper-consistent and has all properties. Let us call this object ‘⊥’. It is not an element of the universe of discourse, so it does not occur in the extension of non-logical predicates. It is a licit referent of a variable though. If g(x) = ⊥, then the partial variable [x|ϕ] will refer to ⊥ under g, no matter whether or not the restriction is true. This means that there is always an assignment function g that fulfills all restrictions on x,namely the one that maps x to ⊥, and the same holds for all variables. This enables us to keep definedness conditions apart that originate from different variables. So if the condition ϕ is true, the referent of [x|ϕ] under g is g(x). If ϕ is false, g(x) is only defined if it is ⊥. There is a third option – ϕ may be undefined because some variable different from x occurring in it has an unfulfilled definedness condition. In this case, [x|ϕ] is undefined under g even if g(x) = ⊥. The modified semantics for first order logic with partial variables is now as follows. Definition 6 (Semantics of partial variables (final version)) ⎧ ⎪ ⎨ g(x) if ϕg x = 1 [x|ϕ]g = ⊥ if ϕg x = 0 and g(x) = ⊥ ⎪ ⎩ undefined else Quantification turns definedness conditions into truth conditions, but it should do so in a discriminating way. The quantifiers should turn all those definedness conditions into truth conditions that are due to restrictions on x, while other definedness conditions remain definedness conditions. Suppose the interpretation of ϕ is undefined under g. If this undefinedness is due to a violation of a restriction on x, then ϕ should be defined under the assignment g[x/⊥]{x} because ⊥ fulfills all restrictions, even inconsistent ones. If, on the other hand, the restrictions on other variables are violated as well, the formula is also undefined under the revised assignment function.
148
Presupposition and Implicature in Compositional Semantics
Furthermore, for an existentially quantified formula to be true requires the existence of an object that verifies the formula in question, and likewise for a universally quantified formula to be false. Here we only care about proper, consistent objects, not about ⊥. Hence quantification should range over the universe without ⊥. This leads to the final version of the semantics for quantification: Definition 7 (Semantics of Quantification (final version)) ⎧ ⎪ 1 if for some a = ⊥ : ϕg[a/x]{x} = 1 ⎪ ⎪ ⎪ ⎨ 0 if ϕ g[x/⊥]{x} is defined, and for all a = ⊥ : ∃xϕg = ⎪ ϕg[a/x]{x} = 1 ⎪ ⎪ ⎪ ⎩ undefined else
∀xϕg =
⎧ ⎪ 1 ⎪ ⎪ ⎪ ⎨
if ϕg[x/⊥]{x} is defined, and for all a = ⊥ :
ϕg[a/x]{x} = 0 ⎪ 0 if for some a = ⊥ : ϕg[a/x]{x} = 0 ⎪ ⎪ ⎪ ⎩ undefined else
Let us see what the new interpretation rule has to say about the previous example. Instead of (56) we have the revised interpretation: (59)
MEET’([x|BOY’(x)], [y|GIRL’(y)])g ⎧ 1 iff g(x) ∈ BOY’ ∪ {⊥} & g(y) ∈ GIRL’ ∪ {⊥} & ⎪ ⎪ ⎪ ⎪ ⎪ g(x), g(y) ∈ MEET’ ⎪ ⎨ = 0 iff g(x) ∈ BOY’ ∪ {⊥} & g(y) ∈ GIRL’ ∪ {⊥} & ⎪ ⎪ ⎪ ⎪ g(x), g(y) ∈ / MEET’ ⎪ ⎪ ⎩ undefined else
After applying existential quantification once, we have: (60)
∃y MEET’([x|BOY’(x)], [y|GIRL’(y)])g ⎧ 1 iff g(x) ∈ BOY’ ∪ {⊥} & ⎪ ⎪ ⎪ ⎪ ⎪ there is an a ∈ GIRL’ : g(x), a ∈ MEET’ ⎪ ⎨ = 0 iff g(x) ∈ BOY’ ∪ {⊥} & ⎪ ⎪ ⎪ ⎪ for no a ∈ GIRL’ : g(x), a ∈ MEET’ ⎪ ⎪ ⎩ undefined else
Partial Variables and Specificity 149
Negation just switches truth values. Existentially quantifying over x then gives us (61)
∃x¬∃y ⎧ MEET’([x|BOY’(x)], [y|GIRL’(y)])g ⎪ 1 iff there is a b ∈ BOY’ : for no a ∈ GIRL’ : ⎪ ⎪ ⎪ ⎨ b, a ∈ MEET’ = ⎪ ⎪ ⎪ 0 iff for no b ∈ BOY’ : for no a ∈ GIRL’ : ⎪ ⎩ b, a ∈ MEET’
These are in fact the (bivalent) truth conditions of the classical formula (62) 4.4
∃x.BOY’(x) ∧ ¬∃y.GIRL’(y) ∧ meet’(x, y) Variable free semantics?
The final semantics for first order logic with partial variables we end up with turns out to be rather complex. We have to enrich the domain of interpretation with an abstract, inconsistent object, and we employ assignment functions that range over sequences of variables rather than over single variables. One might wonder whether this degree of complexity is really necessary. This touches the issue what the status of variables in natural language semantics actually is. Some researchers, especially categorial grammarians, have argued that natural language does not use variables. Rather, variables are a convenient technical device that can ultimately be dispensed with, just like logical representations in a surface compositional model of grammar (see especially the discussion in Szabolcsi (1988) and Jacobson (1999)). The technical devices that I introduced in the last two subsections are arguably tailored to solve problems that a variable free system does not have in the first place. In a variable free system, variable occurrences are replaced by identity functions that function compose with their environment. Having several occurrences of the same variable amounts to merging the argument roles that correspond to them. There is no real counterpart to first order quantification in a variable free system. Rather, quantification is dealt with as in Generalized Quantifier theory, i.e. quantifiers are higher order functions that apply to other functions. There is a natural variable free counterpart to partial variables, namely partial functions. A variable free version of the present theory using partial functions is in fact possible, see Jäger (2001, 2005). In this system,
150
Presupposition and Implicature in Compositional Semantics
the meaning, for instance, of an indefinite like a boy would just be the identity function over the set of boys. The improper object ⊥ was introduced to make sure that the definedness conditions of different variables can be kept apart. In the variable free system, this is automatically the case. Take the example a boy meets a girl. This would be interpreted as a partial function that is defined only over the set of boys. Each boy b is mapped to a partial function fb . fb is a partial function that is defined only over the set of girls, and where defined, it maps each girl g to the truth value of b met g. Now if there are no boys, the clause denotes the empty function, and if there are boys but no girls, it denotes a constant function on the set of boys that always has the empty function as a value. These are different objects, and there is no problem keeping them apart. The treatment of pronouns inside the description of an indefinite (as discussed in the last subsection) is more intricate in the variable free system, and I will not discuss it here. Suffice it to say that the bound pronoun problem does not arise either due to the combinatorics of the system. A proponent of variable free semantics could thus make a case that the complexity introduced in the last two subsections is an artifact of the usage of variables, which should be avoided anyway. At the present point, I remain agnostic as to whether the true theory of natural language semantics actually needs variables. It is undeniable though that working with variables makes life a lot easier because it allows to hide a good deal of the semantic complexity behind a convenient notation. (Even fervent proponents of variable free semantics work with the λ calculus as representation language instead of the variable free combinatory logic, for exactly this reason.) If the variable free view is correct, there is still nothing wrong with using variables as long as one knows how to get rid of them, and the complexity introduced here is just the price one has to pay for a convenient notation. If, on the other hand, variables are really inevitable, the complexity is probably genuine.
5
Partial variables and presuppositions
The analysis of indefinites as partial variables is obviously strongly related to Heim’s treatment, the only difference being that the descriptive part of an indefinite is part of the definedness conditions rather than of the truth conditions. However, the theory developed here can also be considered a variant of the choice function approach. An indefinite DP like a cup is analyzed as [x|CUP’(x)], i.e. as an expression of type e. If the NP part cup is analyzed as a predicate of type e, t like CUP’, the indefinite determiner must have the interpretation λP[x|P(x)], i.e. a function of type e, t, e.
Partial Variables and Specificity 151
Furthermore, this function, if defined, always maps a set to one of its members, i.e. it is a choice function. Unlike classical choice functions, this function is undefined for certain sets, and where defined, it always returns the same value, namely g(x), where g is the local assignment function. So the present theory can be presented as the variant of the choice function approach that results if only partial and constant choice functions are considered. Last but not least, there is a strong affinity between partial variables and presuppositions. Even stronger, I suggest that presuppositions should generally be analyzed as restrictions on partial variables. Presupposition projection works due to the following equivalence under this account: Theorem 1 If [x|ψ] is a free occurrences of x in ϕ and does not occur in the restriction of a variable, then for all models M and assignments g: M ∃xϕ([x|ψ])M g = ∃x.ϕ(x) ∧ ψg
Proof: See Appendix. This law accounts for both varieties of presupposition projection, binding and accommodation. Since Karttunen (1974) it is known that at least in the following two configurations, presuppositions are bound (or, in Karttunen’s terminology, ‘filtered’). I use a simple definite description as an example of a presupposition trigger and only consider the existential presupposition triggered by it for the purpose of illustration. The sentence (63)
The cup broke.
triggers the presupposition that there is a cup. However, in the following complex constructions, this presupposition is filtered/bound. (64)
a. Conjunction First a cup moved, and then the cup broke. b. Conditional If a cup moved, then the cup broke.
I treat the definite description the cup analogously to its indefinite counterpart, except that it does not have to obey the Novelty Condition, and hence it may (and in our examples indeed does) pick up a familiar variable. Given this and the variable binding conventions of DRT, (64a) and (b) are to be translated as (65a) and (b) respectively.
152
(65)
Presupposition and Implicature in Compositional Semantics
a. ∃x.MOVE’([x|CUP’(x)]) ∧ BREAK’([x|CUP’(x)]) b. ∀x.MOVE’([x|CUP’(x)]) → BREAK’([x|CUP’(x)])
Using Theorem 1 twice (and in the case of (b) the inter-definability of the quantifiers), these formulas can be re-written as the equivalent (66)
a. ∃x(CUP’(x) ∧ MOVE’(x) ∧ BREAK’(x)) b. ∀x(CUP’(x) ∧ MOVE’(x) → BREAK’(x))
Neither formula presupposes the existence of cups. (The first one entails it, but its negation does not.) More generally, it is an immediate consequence of Theorem 1 that • If ϕ |= ψ, then – ∃x.ϕ ∧ χ([x|ψ]) ⇔ ∃x.ϕ ∧ χ(x) – ∀x.ϕ → χ([x|ψ]) ⇔ ∀x.ϕ → χ(x) So presupposition binding from the first into the second conjunct, and from the restrictor into the scope of a conditional is correctly predicted to be possible. Technically, presupposition binding means that the variable that hosts the presupposition is familiar in its local context. As van der Sandt’s theory (but unlike other dynamic theories of presuppositions), the present account does not require that a bound presupposition be entailed by its local context. This is empirically justified. In the following example, the presupposition that there is a funny old man, which is triggered by the definite in the second conjunct, is not entailed by its antecedent the teacher, but still binding is possible. (67)
[The teacher]i entered the room, and the students welcomed [the funny old man]i .
Presupposition accommodation amounts to a configuration where the variable hosting the presupposition is novel. Recall that free variables can be bound by applying existential closure at some super-ordinate sentential level (or at the text level). This is tantamount to accommodating the restriction of the variable at this level. Given this, Theorem 1 says that presuppositions can project out of arbitrary embedded contexts. In Karttunen’s terminology, this amounts to the claim that every context is a potential hole for presupposition projection. Filters are binding configurations – following van der Sandt, I assume that binding is
Partial Variables and Specificity 153
ceteris paribus preferred over accommodation. So structurally the example (64b) is ambiguous between binding and accommodation, depending on whether a cup and the cup are translated as the same or as different variables. The translation corresponding to the accommodation configuration would be (68)
∀x.MOVE’([x|CUP’(x)]) → BREAK’([y|CUP’(y)])
Here the variable y is free and has to be bound via existential closure. This can be done either at the level of the succedent clause of the conditional or at the matrix level. Since global accommodation is preferred over local accommodation, the resolved translation is (69)
∃y∀x.MOVE’([x|CUP’(x)]) → BREAK’([y|CUP’(y)])
Due to Theorem 1 and the laws of first order logic, this is equivalent to (70)
∃y.CUP’(y) ∧ ∃x(CUP’(x) ∧ MOVE’(x)) → BREAK’(y)
Since binding is preferred over accommodation, (70) does not represent a pragmatically possible reading of (64b). Specific indefinites are ordinary presupposition triggers, except for the fact that, as indefinites, they must obey the Novelty Condition. Therefore presupposition binding is not possible here, and accommodation is the only option. The natural question to ask is whether all presuppositions can be represented as restrictions of partial variables in a natural way. A detailed answer to this question requires further research; I’ll just add some remarks on it here. Formally, it is always possible to bury a presupposition in a tautological formula, something of the form x = [x|ϕ] which has the only effect of introducing the presupposition ϕ. While this is technically possible, it is rather ad hoc, and one might wonder whether a linguistically more meaningful approach is possible. I assume that many presuppositions that (unlike the presuppositions induced by definite descriptions) cannot naturally be analyzed as restrictions on individual variables should be considered restrictions on event variables. Below are some tentative proposals for a representation of the
154
Presupposition and Implicature in Compositional Semantics
presuppositions induced by aspectual verbs, temporal adverbs, and clefts. (71) (72) (73)
a. Peter quit smoking. b. QUIT_SMOKING’(PETER’, e) ∧ [e |SMOKE’(PETER’, e )] < e a. Peter is tired again. b. TIRED’(PETER’, e) ∧ [e |TIRED’(PETER’, e )] < e a. It was Bill who talked. b. TALK’(BILL’, [e|∃x.TALK’(x, e)])
This approach to the analysis of presuppositions is strongly inspired by van der Sandt’s theory, and the empirical predictions are identical. The difference is a conceptual one: van der Sandt assumes two layers of representation between syntax and model-theoretic interpretation, the underspecified DRS containing unresolved presuppositions, and the resolved DRS. The analysis of presuppositions as partial variables only requires one intermediate level. Presupposition resolution is part of the mapping from syntax to DRSs (or from S-structure to LF, if you like), alongside with quantifier scope assignment and similar disambiguations. It is in fact possible to reformulate the present theory in an entirely surface compositional way if existential closure is treated as a free semantic operation without syntactic counterpart. Since this technical exercise is of minor importance for the issues discussed in this chapter, I refrain from spelling this out here.
6
Conclusion
In this chapter I introduced and motivated an extension of classical first order logic by the construct of partial variables, i.e. variables that only refer if a certain side condition is fulfilled. Quantification was modified such that it turns definedness conditions of the bound variable into a part of the truth conditions. I tried to argue that presuppositions in natural language should be analyzed as definedness conditions on variables. Furthermore, I adopted the idea from the literature that specificity is a special case of presupposition accommodation. So specific indefinites should be translated as partial variables in the semantic representation. I argued that this approach incorporates insights from the DRT approach and the choice function approach to the semantics of indefinites but avoids the empirical pitfalls of these theories.
Partial Variables and Specificity 155
The proposed innovation has consequences for a variety of issues in natural language semantics, and the discussion in this article has necessarily been sketchy at times. There are several points that invite for further research, both empirical and conceptual ones: • Non-specific indefinites: I deliberately remained neutral about whether non-specific indefinites should be analyzed as partial variables as well. Technically this would predict the correct truth-conditions if local accommodation is applied. However, local accommodation is generally the least preferred option, while the non-specific reading of indefinites is usually much easier to get than the specific one. Furthermore, some indefinites like bare plurals only have a non-specific reading. These arguments seem to suggest that we are dealing with a real ambiguity between specific and non-specific indefinites (as Fodor and Sag (1982) and Kratzer (1998) also suggest). This is of course pure stipulation, and the relation between the different brands of indefinites requires further investigation. • Presuppositions: I conjectured that all presuppositions can be considered as restrictions on variables. This was only illustrated with few examples though, where the assumption were rather natural. It remains to be shown that this treatment extends to all classes of presupposition triggers. • Tense: An anonymous reviewer pointed out to me that tense semantics might be another possible application. Following Partee (1973), many authors assume that tenses should be analyzed as variables over times (or time intervals), that can be interpreted deictically or anaphorically. A sentence like (74a) would thus come out as (74b). (74)
a. John walked. b. WALK’(J’)(t) ∧ t < to c. WALK’(J’)([t|t < to ])
In various publications (see for instance von Stechow (2002):14), Arnim von Stechow assumes that past tense sentences are actually undefined if the tense variable does not refer to a time before the speech time. In my notation, the von Stechow-semantics of (74a) would thus come out as (74c). A sentence like (75a), however, is plainly false if John never walked, not undefined, no matter in which context it is uttered. Therefore an appropriate translation should look like (75b), which is equivalent to (75c). The adverbial once upon a time
156
Presupposition and Implicature in Compositional Semantics
can be considered an overt version of existential closure. (75)
a. Once upon a time, John walked. b. ∃t WALK’(J’)([t|t < to ]) c. ∃t(WALK’(J’)(t) ∧ t < to )
• Plural: The interaction of plural semantics with partiality requires further scrutiny. The ability of a plural indefinite to receive a specific interpretation partially depends on whether it is exhaustive and distributive. It remains to be seen whether these restrictions can be accommodated into the present theory in a natural way. • Quantification: Philippe Schlenker (p.c.) suggested to me that partial variables lead to an arguably simpler representation of quantificational structures. A universal quantification as in (76a) could for instance be represented as (b): (76)
a. Every man is mortal. b. ∀xMORTAL’([x|MAN’(x)])
An extension of the present system to generalized quantifiers is straightforward, and the restrictive clause of quantifiers can generally be represented as a restriction of the variable that is bound by the quantifier. This observation might have repercussions for the theory of the syntax-semantics interface. • Compositionality: As pointed out above, treating presuppositions as restrictions on variables does away with one independent level of syntactic representation of van der Sandt’s DRT approach to presuppositions, namely the level of unresolved DRSs. The literature contains several proposals for a compositional reformulation of DRT (see for instance Zeevat (1989) and Asher (1993)). To develop a fully surface compositional theory of specificity and presuppositions, such a compositional DRT must be modified such that accommodation, i.e. selective existential closure, can be incorporated. If successful, one of the main arguments against the DRT approach to presuppositions – its non-compositional architecture – would be invalidated. • Last but not least it is interesting whether the extension of first order logic with partial variables changes its meta-logical properties, and the same question arises for extension thereof like type theory with partial variables. At the present time it is entirely open whether partial variables are just syntactic sugar from a logical point of view or something of higher significance.
Partial Variables and Specificity 157
Acknowledgement First of all I would like to thank Cornelia Endriss, both for getting me interested in the issue of specificity and for ongoing discussions on the subject. Furthermore I thank Christian Ebert, Philippe Schlenker, Yael Sharvit, Peter Staudacher, an anonymous reviewer, and the audiences of related talks at the University of Chicago, the University of Konstanz and the ZAS Berlin for important feedback.
158
Presupposition and Implicature in Compositional Semantics
Appendix Proof of Theorem 1: We need an auxiliary lemma to carry out the proof: Lemma 1 For all assignments g and expressions α, if [x|ϕ] is a free occurrence of x in α and does not occur inside the restriction of a variable, then ⎧ ⎪ ⎨ α(x)g if ϕg x = 1 α([x|ϕ])g = α(x)g if ϕg x = 0 and g(x) = ⊥ ⎪ ⎩ undefined else We prove this via induction over the structure of α. • If α(x) is x or R(. . . , x, . . .), the conclusion follows directly from the semantics of partial variables. • Suppose α(x) is ∃yψ(x). Then y = x because x is free in α by assumption. We distinguish four cases: 1. ϕg x = 1. It follows directly from the definitions that g x = g[y/a]{y}x for arbitrary a (including ⊥). Hence by induction hypothesis, ψ(x)g[y/a]{y} = ψ([x|ϕ])g[y/a]{y} Therefore ψ(x)g[y/⊥]{y} is defined iff ψ([x|ϕ])g[y/⊥]{y} is defined, and this entails that ∃yψ(x)g is defined iff ∃yψ([x|ϕ])g is defined. Given this, it follows that ∃a = ⊥ : ψ(x)g[y/a]{y} = 1 iff ∃a = ⊥ : ψ([x|ϕ])g[y/a]{y} = 1 and hence ∃yψ(x)g = ∃yψ([x|ϕ])g 2. ϕg x = 0 and g(x) = ⊥. Observe that g(x) = g[y/a]{y}(x) and g x = g[y/a]{y}x for arbitrary a. Thus by the same reasoning as in the previous case, ∃yψ(x)g = ∃yψ([x|ϕ])g
Partial Variables and Specificity 159
3. ϕg x = 0 and g(x) = ⊥. As in the previous case, we make use of the fact that g(x) = g[y/a]{y}(x) and g x = g[y/a]{y}x for arbitrary a. Hence by induction hypothesis, ψ([x|ϕ])g[y/⊥]{y} is undefined, and thus ∃yψ([x|ϕ])g is undefined as well. 4. ϕg x is undefined. Since g x = g[y/⊥]{y}x , ψ([x|ϕ])g[y/⊥]{y} is undefined, and thus ∃yψ([x|ϕ])g is undefined as well. • By assumption, α(x) cannot have the form [y|ψ(x)] because x is assumed not to occur in the restriction of another variable. • α(x) has the form ψ(x) → χ, χ → ψ(x), or ¬ψ(x). Suppose ϕg x = 1, or ϕg x = 0 and g(x) = ⊥. Then by induction hypothesis, ψ(x)g = ψ([x|ϕ])g . It follows from the semantics of the propositional connectives that then α(x)g = α([x|ψ])g . If, on the other hand, ϕg x = 0 and g(x) = ⊥, or ϕg x is undefined, then ψ([x|ϕ])g is undefined, and thus α([x|ϕ])g is undefined as well. • All other connectives can be defined in terms of ¬, →, and ∃. This completes the proof of lemma 1. I proceed to the proof of theorem 1. • Suppose ∃ϕ([x|ψ])g = 1. Then there is an a = ⊥ such that ϕ([x|ψ])g[x/a]{x} = 1. Since g[x/a]{x}(x) = a = ⊥, it follows from lemma 1 that ϕ(x)g[x/a]{x} = 1 and ψg[x/a]{x}x = 1. By the definitions, g[x/a]{x}x = g[x/a]{x}, thus, ψg[x/a]{x} = 1, and therefore ϕ(x) ∧ ψg[x/a]{x} = 1. From the semantics of existential quantification, we conclude that ∃x.ϕ(x) ∧ ψg = 1. • Suppose ∃x.ϕ(x) ∧ ψg = 1. By the semantics of ∃, there is an a = ⊥ such that ϕ(x) ∧ ψg[x/a]{x} = 1. Hence ϕ(x)g[x/a]{x} = 1 and ψg[x/a]{x} = 1. Since g[x/a]{x} = g[x/a]{x}x , ψg[x/a]{x}x = 1. By lemma 1, ϕ([x|ψ])g[x/a]{x} = 1, and by the semantics of ∃, ∃ϕ([x|ψ])g = 1. • Suppose ∃ϕ([x|ψ])g = 0. Then ϕ([x|ψ])g[x/⊥]{x} is defined, and there is no a = ⊥ with ϕ([x|ψ])g[x/a]{x} = 1. By lemma 1, ϕ(x)g[x/⊥]{x} and ψg[x/⊥]{x}x are defined. Since g[x/⊥]{x}x = g[x/⊥]{x}, ψg[x/⊥]{x} is defined, hence ϕ(x) ∧ ψg[x/⊥]{x} is defined as well. Now suppose there were a b = ⊥ with ϕ(x) ∧ ψg[x/b]{x} = 1. Then we would have ϕ(x)g[x/b]{x} = 1 and ψg[x/b]{x} = 1 as well. Since g[x/b]{x} = g[x/b]{x}x , we would get ψg[x/b]{x}x = 1, and thus by lemma 1, ϕ([x|ψ])g[x/b]{x} = 1, contra assumption. Hence there is no b = ⊥ with ϕ(x)∧ψg[x/b]{x} = 1, and therefore ∃x.ϕ(x)∧ψg = 0.
160
Presupposition and Implicature in Compositional Semantics
• Finally, suppose ∃x.ϕ(x) ∧ ψg = 0. Then ϕ(x) ∧ ψg[x/⊥]{x} is defined, and thus both ϕ(x)g[x/⊥]{x} and ψg[x/⊥]{x} are defined. By lemma 1, ϕ([x|ψ])g[x/⊥]{x} is also defined. It also follows from the assumption that there is no a = ⊥ with ϕ(x) ∧ ψg[x/a]{x} = 1. Now suppose there were a b = ⊥ with ϕ([x|ψ])g[x/b]{x} = 1. By lemma 1, it would follow that ϕ(x)g[x/b]{x} = 1 and ψg[x/b]{x}1 = 1. Since g[x/b]{x} = g[x/b]{x}x , ϕ(x)g[x/b]{x}x = 1. Therefore ϕ(x) ∧ ψg[x/b]{x} = 1, contra assumption. It follows that there is no b = ⊥ with ϕ([x|ψ])g[x/b]{x} =1. We hence have ∃xϕ([x|ψ])g = 0. This completes the proof.
Notes 1. The restrictions on intra-clausal scope are much more complex, and several classes of scope bearing elements have to be distinguished in this respect – see the discussions in Szabolcsi (1997). Farkas and Giannakidou (1996) show that in certain rather restricted configurations, an embedded universal quantifier can take ‘extra-wide’ scope over an indefinite in the matrix clause. Apart from that, definite descriptions and specific indefinites seem to be the only scopal elements that are not clause-bounded. 2. A similar argument can be made with respect to non-monotonic quantifiers, as an anonymous reviewer correctly points out. 3. See Schwarz (2001) for conclusive arguments that functional readings of indefinites and the peculiar scoping properties of indefinites are independent issues, and neither phenomenon can be reduced to the other. 4. Adjectives like certain and particular tend to make the specific reading of an indefinite more prominent. I have nothing to say why this is so, but I will use that effect nevertheless whenever appropriate to make the reading I am interested in easier to get. 5. Reniers (1997) was presumably the first one to point out that downward monotonic operators are especially problematic for the choice function approach. 6. The notation xn stands for a non-empty sequence of x.
References Abusch, Dorit (1994) ‘The Scope of Indefinites’. Natural Language Semantics 2: 83–135. Asher, Nicholas (1993) Reference to Abstract Objects in Discourse. Kluwer, Dordrecht. Beaver, David (2001) Presupposition and Assertion in Dynamic Semantics. CSLI Publications, Stanford. Chierchia, Gennaro (2001) ‘A Puzzle About Indefinites’. In Carlo Cecchetto, Gennaro Chierchia, and Maria Teresa Guasti (eds), Semantic Interfaces, CSLI Publications, Stanford. Cresti, Diana (1995) Indefinite Topics. PhD thesis, MIT.
Partial Variables and Specificity 161 Dekker, Paul (2002) ‘A Pragmatic View Upon Indefinites’. In Klaus von Heusinger, Ruth Kempson, and Wilfried Meyer-Viol (eds), Proceedings of the ESSLLI 13 Workshop on Choice Functions and Natural Language Semantics, Fachbereich Sprachwissenschaft der Universität Konstanz, Arbeitspapier Nr. 110, 17–34. Diesing, Molly (1992) Indefinites. MIT Press, Cambridge (Mass.). Endriss, Cornelia (2001) The Double Scope of Quantifier Phrases. Master’s thesis, University of Potsdam. Farkas, Donka (1981) ‘Quantifier Scope and Syntactic Islands’. In Papers from the 17th Regional Meeting of the Chicago Linguistic Society, University of Chicago, 59–66. Farkas, Donka and Anastasia Giannakidou (1996) ‘How Clause-bounded Is the Scope of Universals?’ In Teresa Galloway and Justin Spence (eds), Proceedings of SALT VI, CLC Publications, Cornell University, Ithaca, 35–52. Fodor, Janet and Ivan Sag (1982) ‘Referential and Quantificational Indefinites’. Linguistics and Philosophy 5:355–98. Geurts, Bart (1999a) Presuppositions and Pronouns. Elsevier, London. Geurts, Bart (1999b) ‘Specifics’. In Bart Geurts, Manfred Krifka, and Rob van der Sandt (eds), Focus and Presupposition in Multi-Speaker Discourse, ESSLLI’99, University of Utrecht, 99–129. Geurts, Bart (2000) ‘Indefinites and Choice Functions’. Linguistic Inquiry 31:731–8. Heim, Irene (1982) The Semantics of Definite and Indefinite Noun Phrases. PhD thesis, University of Massachusetts, Amherst. Hintikka, Jaakko (1986) ‘The Semantics of a certain’. Linguistic Inquiry 17:331–6. Jacobson, Pauline (1999) ‘Towards a Variable-free Semantics’. Linguistics and Philosophy 22:117–84. Jäger, Gerhard (2001) ‘Indefinites and Sluicing: A Type Logical Approach’. In Robert van Rooij and Martin Stokhof (eds), Proceedings of the 13th Amsterdam Colloquium, ILLC, University of Amsterdam, 114–19. Jäger, Gerhard (2005) Anaphora and Type Logical Grammar. Springer, Dordrecht. Kamp, Hans and Uwe Reyle (1993) From Discourse to Logic: Introduction to Modeltheoretic Semantics of Natural Language, Formal Logic and Discourse Representation Theory. Kluwer, Dordrecht. Karttunen, Lauri (1974) ‘Presuppositions and Linguistic Context’. Theoretical Linguistics 1:181–94. Kratzer, Angelika (1998) ‘Scope or Pseudscope?: Are there Wide Scope Indefinites?’ In Susan Rothstein (ed.), Events and Grammar, Kluwer, Dordrecht, 163–96. Krifka, Manfred (1998) ‘Non-novel Indefinites in Adverbial Quantification’. In Carlota Smith (ed.), Workshop Proceedings: Non-narrative Discourse, Texas Linguistic Forum, Department of Linguistics, University of Texas at Austin. Partee, Barbara (1973) ‘Some Structural Analogies Between Tenses and Pronouns in English’. Journal of Philosophy 70:601–9. Reinhart, Tanya (1992) ‘Wh-in-situ: an Apparent Paradox’. In Paul Dekker (ed.), Proceedings of the Eighth Amsterdam Colloquium, University of Amsterdam, 483–91. Reinhart, Tanya (1995) Interface Strategies. OTS Working Papers. Research Institute for Language and Speech, Utrecht University. Reinhart, Tanya (1997) ‘Quantifier Scope: How Labor is Divided Between QR and Choice Functions’. Linguistics and Philosophy 20:335–97.
162
Presupposition and Implicature in Compositional Semantics
Reniers, Fabien (1997) How to (S)cope with Indefinites. Master’s thesis, University of Utrecht. Schwarz, Bernhard (2001) ‘Two kinds of long-distance indefinites’. In Robert van Rooij and Martin Stokhof (eds), Proceedings of the Thirteenth Amsterdam Colloquium, University of Amsterdam, 192–7. Szabolcsi, Anna (1988) Bound Variables in Syntax. (Are There Any?). Institute of Linguistics, Hungarian Academy of Sciences, Budapest. Szabolcsi, Anna (1997) ‘Strategies for Scope Taking’. In Anna Szabolcsi (ed.), Ways of Scope Taking, Kluwer, Dordrecht, 109–54. van der Sandt, Rob (1992) ‘Presupposition Projection as Anaphora Resolution’. Journal of Semantics 9:333–77. van Geenhoven, Veerle (1998) Semantic Incorporation and Indefinite Descriptions. CSLI Publications, Stanford. von Stechow, Arnim (2002) ‘Temporal Prepositional Phrases with Quantifiers’. Ms, University of Tübingen. Winter, Yoad (1997) ‘Choice Functions and the Scopal Semantics of Indefinites’. Linguistics and Philopsophy 20:399–467. Yeom, Jae-Il (1998) A Presuppositional Analysis of Specific Indefinites. Garland Publishing, New York, London. Zeevat, Henk (1989) ‘A Compositional Approach to Discourse Representation Theory’. Linguistics and Philosophy 12:95–131.
6 Negated Antonyms: Creating and Filling the Gap Manfred Krifka1 Humboldt University, Berlin
1
Double negatives
This chapter proposes a theory of the semantics and pragmatics of certain antonym pairs and their negations, as in the following examples: (1)
a. b. c. d. e. f. g. h.
happy, not happy, unhappy, not unhappy likely, not likely, unlikely, not unlikely intelligent, not intelligent, unintelligent, not unintelligent successful, not successful, unsuccessful, not unsuccessful polite, not polite, impolite, not impolite common, not common, uncommon, not uncommon frequent, not frequent, infrequent, not infrequent many, not many, few, not few
The interpretation of the fourth members of such quadruplets, the so-called ‘double negatives’, have caught the attention of linguists and language critics alike, as detailed in several publications by Horn (1989: 296 ff., 1991, 1993, 2002). Jespersen (1924) describes what is perhaps the most tangible meaning effect in the following way: The two negatives [. . .] do not exactly cancel one another, so that the result [not uncommon, not infrequent] is identical with the simple common, frequent; the longer expression is always weaker: “this is not unknown to me” or “I am not ignorant of this” means ‘I am to some extent aware of it’, etc. (p. 332). According to this, double negatives have the same interpretation as weak positives. There might be additional meaning components associated 163
164
Presupposition and Implicature in Compositional Semantics
with this mitigating effect of double negatives. For example, Fowler (1926, p. 383) observed that this effect makes them suitable for intimating a certain attitude ‘congenial to the English temperatment’ such that ‘it is pleasant to believe that it owes its success with us to a stubborn national dislike of putting things too strongly’. But there are other uses that have been identified. For Orwell (1999 [1946]: p. 357) double negatives had a purely connotational meaning, apparently arising from their added complexity: ‘Banal statements are given an appearance of profundity by means of the not un- formation’ (p. 357). And it has been observed that sometimes double negatives even seem to express a meaning stronger than simple positives, as e.g. certain uses of not bad do. Horn lists a great many convincing examples that show the mitigating use of double negatives quite clearly. A quick search of web sites, blogs and discussion forums reveals this semantic effect in the language of Internet users, as the following randomly selected examples show: (2)
a. For the first time in a long time I woke up feeling good this morning. I felt rested and content and was not unhappy to find that my kids were already awake. b. For the most part he is well informed and not unintelligent. c. It is not unlikely that these early therapsids may have had the beginnings of metabolic development towards the mammalian condition. d. Homeowners Debt: ‘45% of Income not Uncommon’
In all these examples, expressions of the form not un-A could be replaced by rather A, or quite A, without significant change in meaning. Replacing them simply by A results in a more drastic meaning change: (3)
a. I was not unhappy to find that my kids were already awake. b. ≈ I was quite happy to find that my kids were already awake. c. = I was happy to find that my kids were already awake.
But one should be aware that a regular negation reading of double negative construction exists as well, especially in denials, as the following examples show: (4)
a. Now if I look unhappy, I apologize. But I’m not unhappy. b. JIM LEHRER: In other words, if the marines are not talking, it isn’t necessarily because they’re not happy. . . they’re not unhappy. c. i am happy. i am not insanely happy, but i am not unhappy.
Negated Antonyms
165
d. Beauty: I am NOT unintelligent!!! Geek: Okay, you’re not unintelligent, you’re . . . academically un-inclined. A quick survey makes it obvious that the mitigating reading of double negatives occurs with different verbs in different frequency, a phenomenon worth to be studied in greater detail. Also, the two readings appear to differ in the prosodic structure: Whereas for the ‘quite A’ reading the string not un-A is realized within one prosodic phrase, for the ‘¬ un-A’ reading it is typically realized within two separate prosodic phrases: (5)
a. (He is) (not UNhappy about it) ‘He is quite happy about it.’ b. (He is NOT) (UNhappy about it) ‘It is not the case that he is unhappy about it.’
I will leave these issues aside and rather develop an explanation how this interpretation arises in the first place that has advantages over existing explanations.
2
Attempts to explain double negatives
Horn (1991, 1993) explains the downtoning effect of stating not un-A over simply stating A as follows: The speaker avoids A because there might not be sufficient evidence for stating A (violation of Quality), or stating A would violate constraints posed by politeness considerations. This explanation presupposes that A and un-A are contraries, and not just contradictions; otherwise A and not un-A would in fact mean the same, which would result in virtual ungrammaticality (cf. Orwell’s example *a not unblack dog). This is illustrated in the following diagram. (6) unhappy
happy
not unhappy (not unhappy)
The adjectives happy and unhappy each signal continuous regions in the emotional space, reduced here to a linear open-ended scale. The range
166 Presupposition and Implicature in Compositional Semantics
of application leaves a gap in the middle range of neutral emotions, the ‘zone of indifference’ (Sapir 1944). This is characteristic for contraries in general. If not unhappy is the negation of unhappy, then these two expressions are contradictories of each other, and not unhappy can be applied to happy emotions, but also to emotions of the neutral range, as indicated. For expressing happy emotions, the two terms happy and not unhappy stand in competition with each other. Here happy wins on two counts: as the simpler expression of the two, and as the more specific one. Hence by Grice’s maxim of Quantity, as well as by his maxim of Manner, we should expect that not unhappy is pragmatically restricted to signal a neutral state of happiness, as indicated. The problem with this analysis is that not unhappy is in fact not simply used to refer to the zone of indifference, the gap that unhappy and happy leave open. It does not mean what the psychologist Judith M. Bardwick called plateaued in her book The Plateauing Trap of 1986: (7)
When we are plateaued, we are not so much actively unhappy as we are just not happy. We could continue to live as we are, because it’s not awful. But it is also not joyous.
Rather, not unhappy is used to refer to a – typically mild – state of happiness. Example (3a.) does not mean ‘I was indifferent as to whether the kids were awake.’ This is not predicted by the analysis above. There is another attempt to explain the interpretation of the meaning of quadruplets like (1) that should be mentioned here, proposed by Blutner (2004). It makes use of the framework of Bidirectional Optimality Theory, but the explanation of double negatives like not unhappy is essentially the same as with Horn, and has the same problems. While Blutner’s diagrams suggest that he thinks of not being unhappy as a mild state of happiness, this does not follow from his theory in any obvious way. Blutner looked, in addition, at expressions like not happy, whose literal meaning is indicated in the following diagram: (8)
unhappy
happy
not unhappy (not unhappy) not happy
Negated Antonyms
167
It is unclear how not happy should be pragmatically restricted, if it is restricted at all. Blutner suggests that it is restricted to the range of unhappy. This is problematic for the theoretical reason of avoidance of synonymy, as two expressions, unhappy and not happy, would end up with the same range of use. And it actually makes a wrong empirical claim, as unhappy can be used to refer to a more severe state of unhappiness than not happy. This can be seen in examples like the following: (9)
I’m not happy at all, in fact I’m quite unhappy.
∗ I’m quite unhappy, in fact I’m not happy at all.
The intensifier at all forces not happy to be understood in a strengthened sense, and not just as the negation of happy. Yet there is an obvious difference between not happy at all and unhappy. Examples like the following, found on the Internet, support this differential use of unhappy and a pragmatically strengthened not happy. (10)
a. I was very unhappy. I am not happy now. Everything is so sad. [George Eliot, Middlemarch] b. It’s an absolutely horrible feeling to be unhappy, and I don’t even think I was unhappy, just not happy, if you know what I mean.
Alternatively, we might assume that not happy is restricted to the zone of indifference, just as not unhappy, by a similar reasoning as the one that we applied in the case of that expression. But notice that then, again, two expressions would end up pragmatically synonymous. As a matter of fact they aren’t, as can be witnessed by the following example, a coherent interchange. (11)
A: John isn’t happy. B: Well, he’s not unhappy.
What we would like to have is a theory that predicts that both not happy and not unhappy are pragmatically restricted to the zone of indifference, but that not happy occupies the lower (left) region of it, whereas not unhappy settles to the higher (right) area. In the following I will show that this can be done by assuming completely standard steps of pragmatic reasoning if we assume a perhaps not-so-standard theory of the literal meaning of the basic antonyms, unhappy and happy.
168
Presupposition and Implicature in Compositional Semantics
3 Pragmatic strengthening within an epistemic theory of vagueness As mentioned above, antonym pairs like happy and unhappy are typically conceived as contraries that apply to the more extreme ends of a scale and leave a zone of indifference. This accounts for the fact that the two statements I am happy and I am unhappy can intuitively both be false, as when I say I am neither happy nor unhappy. I would like to suggest that happy and unhappy, against this impression, are in fact contradictories that exhaust the whole range of their scale. This means that literally, if I am happy is false then I am unhappy is true, and vice versa; these sentences cannot both be false. It is only after pragmatic strengthening of these expressions that I can say that I am neither happy nor unhappy. This position finds support in the view of vagueness proposed by Williamson (1994). According to his epistemic theory of vagueness, a proposition containing a vague predicate, which is traditionally allowed to be neither true nor false in certain circumstances, in fact always is true or false (provided that its presuppositions are met). Consider the consequences of this view with the classical example of vagueness, the Sorites paradox. If one grain of sand after another is added to a collection of sand grains, there is a precise point at which the sentence This is a heap of sand changes its truth value from false to true. Or, if Louis looses one hair after another, then there is a precise point at which we may truthfully say Louis is bald. Also, the sentence Rembrandt is old is either true or false; according to our understanding of old, it must have changed its truth value at some time – in fact, some second – between the years 1606, the year of Rembrandt’s birth, and 1669, the year of his death. This analysis appears to go against the way how we use language, but Williamson reminds us that the alternatives are even more troubling. For example, if we allow for a zone of indifference in which a person is neither not old nor old, or in which a collection of sand grains neither form a heap nor doesn’t, then we run into the problem of second-order vagueness. This is because we have to reintroduce sharp borders again, like the one between (a) being neither old nor not old, and (b) being old. Using a continuum of truth values has been suggested as a solution, but this turns out to be unsatisfying: There is a sense in which a person definitely can be called old, and the question then arises when exactly this definite state was reached. Due to arguments of this sort, Williamson suggests a theory of vagueness according to which vague sentences literally are always either true
Negated Antonyms
169
or false, but typically language users do not know their precise truth conditions in the so-called zone of indifference. This is the epistemic theory of vagueness, which has been intensily debated ever since it was proposed. I would like to subscribe to this view, adding that speakers might well know that in many cases there is no authority that could determine the precise borders (as is the case for pamphlet and book, which the Unesco Institute of Statistics defines as non-periodic publications of at least 5 but not more than 48 pages, and of at least 49 pages, respectively, exclusive of the cover pages). As speakers in the same community may set the border between A and not A differently, and as even one and the same speaker may set the border differently at different occasions, communicating something as being A, or as being not A, becomes rather pointless if the phenomenon to be classified is close to a potential border between A and not A. Hence such classifications are avoided, as they do not safely communicate what they intend to communicate. If it is not guaranteed that speaker and hearer set the border between being bald and being not bald in the same way, then it is potentially misleading if a speaker calls Louis bald. As the addressee knows this, he will assume, on hearing that Louis is bald, that Louis is not a borderline case of baldness where opinions might differ, but a definite case of apolecia. This means that he can be called bald under any reasonable setting of the border. But this is a matter of use, not of the meaning proper, of such terms as bald, old, or heap. The urge to use expressions that both interlocutors are likely to use in the same way for the classification of phenomena is somewhat reminiscent of the tendency towards stereotypical interpretations that underlies implicatures based on what has been called the R principle by Horn (1984, 1993) and the I principle by Atlas and Levinson (1981) and Levinson (2000). The R/I principle is generally seen as leading to an enrichment of the literal meaning of expressions such that they refer to stereotypical instances. To cite a famous example by McCawley (1978), x kills y means ‘x causes y to die’, but this meaning can be enriched so that x kills y is applied only if x causes y to die in a direct, intentional way. I would like to propose that there is a similar enrichment of the literal meaning of expressions to be applicable to cases in which speaker and addressee can assume that they both apply the expressions in the same way. This restricts the application of vague terms to instances that are sufficiently removed from potential borderline cases, even if the border itself is supposed to be sharp. Both types of restrictions, the one to stereotypical interpretations and the one to regions removed from
170 Presupposition and Implicature in Compositional Semantics
potential borders, serve to make communication safer, as they make it less likely that speaker and addressee misunderstand each other. As a second hypothesis, I would like to suggest that antonym pairs exhaust their semantic space. For example, happy and unhappy completely divide the scale of emotions along the dimension of happiness (at least if we do not account for the differentiation between sad and unhappy, which cannot be captured on a linear scale – see Muehleisen 1997). Similary, likely and unlikely, polite and impolite, frequent and infrequent and other antonyms exhaust their semantic space. It is only because there is no agreed-upon border between the semantic range of such terms that they are effectively used only in cases in which speakers and addressees can assume that they categorize phenomena in the same way. My third hypothesis is similar to one assumed already by McCawley (1978), which was called M principle by Levinson (2000). It states that of two expressions with similar meanings, the simpler one tends to be restricted to a stereotypical, or in general, safer interpretation, whereas the more complex one is used when a non-stereotypical interpretation should be communicated. In short, marked expressions tend to have marked meanings (see also Horn 1984, 1993 for discussion). The M principle can be derived by assuming avoidance of synonymous interpretations, as the more complex expression is used for cases for which the simpler expression is avoided, due to R/I-based implicature. This can be illustrated with McCawley’s minimal pair, Black Bart killed the sheriff and Black Bart caused the sheriff to die, as follows: (12) Lexical meanings:
R/I-based implicature:
M-based implicature:
BB killed the sheriff.
(BB killed the sheriff.)
(BB killed the sheriff.)
BB caused the sheriff to die. BB caused the sheriff to die. (BB cause the sheriff to die.)
This principle can also be derived within Bidirectional Optimality Theory as a balancing of two tendencies: One that prefers the shorter expression over the more complex one, and one that prefers the stereotypical interpretation over the non-stereotypical one. To illustrate this with McCawleys example, assume that we consider two form alternatives and two interpretation alternatives: (13)
F1 : Black Bart killed the sheriff. F2 : Black Bart caused the sheriff. to die
I1 : ‘He killed him in a direct way.’ I2 : ‘He killed him in an indirect way.’
Negated Antonyms
171
As preference orderings we have F1 > F2 and I1 > I2 . The evaluation mechanism of Bidirectional Optimality Theory due to Blutner (2004) and Jäger (2002) (so-called ‘weak optimality’) works as follows: (14)
A form-interpretation pair F, I is optimal iff a. there is no optimal pair F, I such that I > I, b. and there is no optimal pair F , I such that F > F.
Among the four pairs F1 , I1 , F1 , I2 , F2 , I1 , F2 , I2 , the first and the last one are optimal. The first one, F1 , I1 , is clearly optimal as there is no pair at all that is preferred over it. The pairs F2 , I1 and F1 , I2 are not optimal, as the optimal pair F1 , I1 is preferred for reasons of form or interpretation, respectively. Interrestingly, the pair F2 , I2 again is optimal: According to the definition, it is not to be compared with F1 , I1 , as this pair varies in both form AND interpretation. And it is also not to be compared with the pairs F2 , I1 or F1 , I2 , as these pairs are not optimal, as we just have seen. The following diagram illustrates the preference orderings (indicated by arrows) and the optimal pairs (indicated by enclosures): (15)
具F1,I1典 具F2,I1典
具F1,I2典 具F2,I2典
We now return to antonyms and their negations. If antonyms are contradictories, and if negation expresses a contradictory relation as well, the literal meanings of the expressions of a quadruplet relate to each other in the following way: (16) unhappy
not happy
happy
not unhappy
According to the epistemic theory of vagueness, the precise delineation between the two regions is not fixed; there is a whole range of options,
172 Presupposition and Implicature in Compositional Semantics
as illustrated in the following diagram: (17) unhappy
happy
not happy
not unhappy
As a consequence of this uncertainty about the location of the border between happiness and unhappiness, the use of unhappy and happy is pragmatically restricted to those areas for which the interlocutors can assume to be in mutual agreement, to ensure safe communication. This is a special case of an R/I-based implicature: (18) (unhappy)
not happy
(happy)
not unhappy
The diagram indicates a particular way in which happy and unhappy are restricted by R/I-implicature. But of course the borders of the restriction is vague in the sense that even after strengthening, speaker and addressee cannot be sure that they agree on the points at which happiness and unhappiness start. In a sense, the problem of second-order vagueness raises its head again. But notice that now this is not a semantic phenomenon, but a pragmatic one. The borders can, in a sense, be negotiated between speech participants; they are not part of the semantics of the language itself. In the right-hand half of this diagram, we find two competing expressions, happy and not unhappy, where the second one is clearly more complex than the first. This is the configuration that triggers M-implicatures.
Negated Antonyms
173
In the following diagram, some of the ranges of possible interpretations of not unhappy after M-implicatures are indicated. (19) (happy)
(unhappy)
not happy
(not unhappy)
This is precisely the effect that we are trying to reconstruct. Not unhappy is used to express mild states of happiness, and not just the ‘plateau’, the middle ground between happiness and unhappiness. What about unhappy and not happy? These two expressions compete with each other as well, but it is less clear in this case which of them is the simpler one. While unhappy is one complex word, not happy is a syntactically complex expression consisting of two simple words. Thus it is expected that the contrast of use between unhappy and not happy is less clear than the one between happy and not unhappy, and this is what we find. But still we can argue that not happy is more complex than unhappy: This clearly holds as far as phonology is concerned, and we may plausibly defend that syntactic complexity weighs more than morphological complexity, at least for the more frequent derivational affixes. This would predict our somewhat tentative impression that unhappy indicates a stronger feeling of unhappiness than not happy. This situation is indicated in the following diagram: (20) (unhappy)
(happy)
(not happy)
(not unhappy)
Notice that there is no clear border between not happy and not unhappy, and the two expressions are not exhaustive. This is indeed what we find;
174
Presupposition and Implicature in Compositional Semantics
a person can be said to be not happy but also not unhappy at the same time. But notice that our theory predicts in any case that not unhappy can be used to report a higher level of happiness than not happy, and this in also what we find. We have observed above that one can describe one’s emotional state by saying: (21)
I am neither happy nor unhappy.
According to our explanation of the way how these expressions are used, sentences of this type are contradictions. To get a non-contradictory reading, we must assume that unhappy and happy are strengthened locally. The sentence then means, roughly: ‘I am in a state for which neither the statement I am happy nor the statement I am unhappy is appropriate’. While strengthening implicatures have been understood in a way that they only apply at the level of the whole sentence, a number of phenomena are known that have forced researchers to assume implicatures that are triggered more locally (cf. Krifka 1995, Chierchia 2004), and sentences like (21) appear to be one of them. There is one independent argument for the assumption that antonym pairs like happy and unhappy are literally contradictories that receive their interpretations as contraries only via pragmatic strengthening. In contexts in which it is irrelevant where the line between them is precisely drawn, antonyms are indeed exhaustive. This is the case with so-called unconditionals (Zaefferer 1991): (22)
Regardless whether you are a happy person or an unhappy person, you should read this book.
This clearly means that everyone should read the book; it does not exclude people that one would prefer to describe as indifferent, or as only mildly happy or unhappy. I would like to point out that the theory of vague predicates presented here differs in an interesting way from one standard theory of dealing with vagueness, namely supervaluations, cf. Fine (1975). Whereas supervaluation theory assumes that the semantic interpretation of predicates leaves a gap between the positive and the negative extension that can be closed by ever-stricter valuations, the theory proposed here assumes, on the contrary, strict semantic interpretation that can be pragmatically weakened. It appears crucial for the application of pragmatic reasoning
Negated Antonyms
175
like the R/I principle and the M principle to assume a theory of vagueness that works with pragmatic weakening. It is perhaps not necessary to assume the epistemic theory of vagueness, but there seems to be no other plausible pragmatic reason why we should have a restriction of semantically exhaustive predicates to ones that leave gaps in the space of entities to which they are applicable.
4
Conclusion
We have seen that changing a crucial assumption about the lexical meaning of antonyms – that they form contradictories, not contraries – leads to a better understanding of the pragmatics of double negatives, like not unhappy, but also of simple negatives, like not happy. The assumption that antonyms are contradictory itself can be defended as well if we distinguish between their literal meaning and the way in which they are pragmatically restricted to guarantee safe communication. Does this cover all uses of double negatives? Horn (1991, 1993, 2002) has listed seven uses altogether. Some of them clearly cannot be explained by the conspiration of literal meaning and pragmatic reasoning suggested above. There might not be a corresponding A form, as is the case in not unfounded. The speaker might take up with the not un-A construction an un-A form that has been uttered before, perhaps because it is to be denied. A speaker might just want to sound pompous. In this case the use of double negatives may be selected for simply because they are more complex, which could be explained by the handicap principle (Zahavi & Zahavi 1997). Or double negatives might be used because they suggest a more careful exploitation of semantic and pragmatic distinctions of language in general. Or they might be used to imitate others that use them for one of the reasons mentioned. All these uses are rather tangential to our concerns here. Horn also mentiones irony, as in Honey, you’re not uncute. It is obvious how this works: It playfully suggests that the addressee cannot really be called cute, but that the speaker does his very best to report a high state of cuteness while staying honest. Another reason Horn mentions is politeness, which can be explained by the fact that the mitigated interpretation that double negatives afford may help to save face. This includes the face of the speaker, for whom the weaker statement of a double negative provides the possibility of a loophole if what he says is not quite true; cf. Seright (1966). Jespersen (1924) presumably had something similar in mind when he says that double negatives imply an hesitation on the side of the speaker.
176
Presupposition and Implicature in Compositional Semantics
Centuries earlier, Erasmus had claimed that double negatives do not only serve to impute modesty, but also insisted that they may be used to ‘amplify’, that is, to make a statement stronger. Examples of this use of double negatives are It was no small joy to receive your letter, or Not bad! as a comment to an artist performing a salto on a tightrope. At first sight this seems to run contrary to what we have assumed about the nature of double negatives. But there is a natural way by which this use to strengthen a proposition may come about when we assume that basically, double negatives make a weaker statement: If the context actually indicates that A holds, a speaker might suggest by saying not un-A that his standards for saying A are so high that they cannot be attained under normally conceivable circumstances. Getting a comment of this type from a critical mind who does ostensively not succumb to the inflationary tendencies of language should be valued more than cheap praise from someone who is known to exaggerate.
Note 1. Thanks to very helpful suggestions to Benjamin Spector, and to participants at presentations of the content of this chapter at the workshop on ‘Cognitive Foundations of Communication’, Koninklijke Nederlandse Akademie van Wetenschapen, Amsterdam, 2004, and at the workshop ‘Theory and Evidence in Semantics’, Groningen, 2006.
References Atlas, Jay David & Stephen C. Levinson (1981), ‘It-clefts, informativeness, and logical form: radical pragmatics (revised standard version)’, in Peter Cole (ed.), Radical Pragmatics, New York, Academic Press, 1–61. Bardwick, J.M. (1986), The Plateauing Trap: How To Avoid It in Your Career, New York: American Management Association. Blutner, Reinhard (2004), ‘Pragmatics and the lexicon’, in Laurence R. Horn & Gregory Ward (eds), Handbook of pragmatics, Oxford, Blackwell. Chierchia, Gennaro (2004), ‘Scalar implicature, polarity phenomena, and the syntax/pragmatic interface’, in Adriana Belletti (ed.), Structures and Beyond, Oxford University Press, 39–103. Fine, Kit (1975), ‘Vagueness, truth and logic’, Synthese 30, 265–300. Fowler, H. W. (1926), A Dictionary of Modern English Usage, Oxford University Press, reprint 2003. Horn, Laurence R. (1984), ‘Toward a new taxonomy for pragmatic inference: Q-based and R-based implicature’, in D. Schiffrin (ed.), Meaning, Form, and Use in Context: Linguistic Applications, Washington, DC, Georgetown University Press, 11–89. Horn, Laurence R. (1989), A Natural History of Negation. University of Chicago Press.
Negated Antonyms
177
Horn, Laurence R. (1991), ‘Duplex negatio affirmat . . .: the economy of double negation’, Chicago Linguistic Society 27(2), 80–106. Horn, Laurence R. (1993), ‘Economy and redundancy in a dualistic model of natural language’, SKY – Yearbook of the Linguistic Association of Finland. Horn, Laurence R. (2002), ‘The logic of logical double negation’, in Yasuhiko Kato (ed.), Proceedings of the Sohia Symposium on Negation, Tokyo. Jäger, Gerhard (2002), ‘Some notes on the formal properties of bidirectional Optimality Theory’, Journal of Logic, Language and Information 11, 427–51. Jespersen, Otto (1924), The Philosophy of Grammar, Allen and Unwin, London. Krifka, Manfred (1995), ‘The semantics and pragmatics of polarity items’, Linguistic Analysis 25, 209–57. Levinson, Stephen C. (2000), Presumptive Meanings, MIT Press, Cambridge, Mass. McCawley, James D. (1978), ‘Conversational implicature and the lexicon’, in Peter Cole (ed.), Pragmatics, New York, Academic Press. Muehleisen, Victoria Lynn (1997), Antonymy and Semantic Range in English, Diss., Northwestern University, Evanston, Ill. Orwell, George (1999) [originally published in 1946], ‘Politics and the English language’, in Stuart Hirschberg & Terry Hirschberg (eds), Reflections on language, New York, Oxford, Oxford University Press, 500–12. Sapir, Edward (1944), ‘Grading: a study in semantics’, Philosophy of Science 4, 93–116. Seright, O.D. (1966), ‘Double negatives in Standard Modern English’, American Speech 41, 123–6. Van Fraassen, Bas (1966), ‘Singular terms, truth value gaps, and free logic’, Journal of Philosophy 63, 464–95. Williamson, Timothy (1994), Vagueness, Routledge, London. Zaefferer, Dietmar (1991), ‘Conditionals and unconditionals: cross-linguistic and logical aspects’, in Dietmar Zaefferer (ed.), Semantic Universals and Universal Semantics, Berlin / New York, Foris, 210–36. Zahavi, Amotz & Avishag Zahavi (1997), The Handicap Principle, Oxford University Press.
7 A Pragmatic Constraint on Adverbial Quantification1 Orin Percus University of Nantes
Even if we can generate a logical form, principles of use may limit the ways in which we can use it. In this chapter, I motivate one such principle of use, and explore its effects. Much of the discussion involves kinds of sentences that have received attention in the literature on ‘individual-level predicates.’
1 A possible line of argumentation, and reasons to reject it A familiar line of argumentation2 takes this form: it calls attention to the fact that, unlike sentences like those in (1), sentences like those in (2) sound bizarre. It points out that we can explain why they sound bizarre if we assume that they do not admit logical forms of the kind that (1) admits. It concludes that indeed sentences like (2) do not admit logical forms of the kind that (1) admits. (1) (2)
Ingrid is always on the phone. Ingrid was always Swedish.
More specifically, the argument starts from the idea that in the logical form of a sentence containing an adverbial quantifier like always, the quantifier bears an index. The logical form for (1) on this approach is something like (1 a). (And logical forms like these are interpreted along the lines of (1 b).) The argument points out that we can explain why sentences like (2) sound bizarre if we assume, along with some other assumptions, that their possible logical forms contain no item that can be coindexed with the quantifier. It concludes on this basis that indeed (2) contains no item that can be coindexed with the quantifier. 178
A Pragmatic Constraint on Adverbial Quantification
(1 )
179
a. [α [. . . always1 . . . ] [β . . . 1 . . . Ingrid on the phone . . . ] ] b. [[α]]g (w) = 1 as long as for all time intervals t that satisfy in w the contextually salient property of time intervals, [[β]]g 1−>t (w) = 1.
My hope in this chapter is to defeat this line of argumentation. My thesis is that there are principles of grammar that (2) violates even if we analyze it as containing an item that can be coindexed with the quantifier. If this is correct, then just on the basis of the strangeness of (2), we have no evidence for the claim that its logical form lacks these additional indices (or for that matter for the assumptions that derive the bizarreness of (2) given the lack of additional indices). Maybe there is evidence for something like this, but it doesn’t come from sentences like (2). My main goal in this chapter is to motivate independently a principle from which it happens to follow that (2) sounds bizarre. I will demonstrate some of the consequences of this principle, and I will conclude by suggesting that once we recognize this principle, we gain some insight into the question of what interpretations sound natural for sentences with indefinites (I will be concerned here with ‘semantic partition’ effects). This chapter thus should serve as a cautionary note. To the extent that the pragmatic principle here can answer the question of why sentences like (2) sound bizarre and what interpretations sound natural for sentences with indefinites, we should not use the same facts to motivate syntactic stipulations that do the same work. To see why I think that (2) incurs problems that should not be traced to the presence or absence of an item coindexed with the quantifier, consider the sentence in (3). (3)
The student who finished first was always Swedish.
Imagine that we were both present at a series of exams, which took place from Monday through Saturday. We both saw that each time a different person finished first. In fact, the pattern was as in (4a): (4)
We both see: a.
date: M T W Th F S fastest student: Max Ingrid Olof Ingmar Ingvar Ingeborg
You have just seen the students, you don’t know who they were or what their nationalities were, and so I tell you (3) to inform you of their nationalities. There is nothing bizarre about (3) when used in this context. What
180
Presupposition and Implicature in Compositional Semantics
can we conclude from this? If the absence of an item coindexed with the quantifier would render the sentence bad, and in fact the sentence is good, we can conclude that the sentence admits a logical form containing an item coindexed with the quantifier. Now that we have established that (3) admits a logical form containing an item coindexed with the quantifier, imagine a different situation. Just as before, we were both present at the exams, but this time the pattern is different: b.
date: M T W Th F S fastest student: Ingrid Ingrid Ingrid Ingrid Ingrid Ingrid
Since we were both there, we both saw that the same person finished first each time. Still, you don’t know which student or their nationality was. In this context, I cannot use (3) to inform you of the student’s nationality. It sounds bizarre. And, since we have already concluded that the sentence admits a logical form containing an item coindexed with the quantifier, it can’t be that its bizarreness is due to the lack of such a logical form. I think that it is reasonable to expect that the same thing that renders (3) bizarre in this context renders (2) bizarre. That is why I think that (2) incurs problems that have nothing to do with the lack of an item that can be coindexed with the quantifier. A similar point can be made by considering sentences like (5). Unlike the sentence in (6), which sounds fine and therefore must contain an item coindexed with the quantifier, (5) sounds bizarre, at least on an initial reading and encountered in isolation. (5) (6)
John always knows whether Ingrid is Swedish. John always knows whether Ingrid is on the phone.
Now, when it comes to the sentence in (6), it can be argued not only that the quantifier is coindexed with another item, but more specifically that it is coindexed with an item in the matrix clause. (I will sketch how in a moment, but the central observation is that (6) asserts that John is in a certain mental state during each of the relevant time intervals.) On the assumption that the only syntactic difference between (5) and (6) is the embedded clause, this means that (5) too must contain an item in the matrix clause that always can be coindexed with. So it can’t be that the bizarreness of (5) is due to the lack of this kind of item. I think that it is reasonable to expect that the same thing that renders (5) bizarre renders (2) bizarre, and so that is another reason for thinking
A Pragmatic Constraint on Adverbial Quantification
181
that (2) incurs problems that have nothing to do with the lack of an item that can be coindexed with the quantifier.3 Here in rough outline is the argument that sentences like (6) contain an indexed item in the matrix (those already satisfied should feel free to proceed to the next section). The starting assumption is that the possible logical forms for (6) are as in (7b) (7)
a. John always knows whether Ingrid is on the phone. b. [. . . alwaysi . . . ] [β . . . John know whether . . . [γ ]]
where the embedded constituent γ is the same kind of constituent that you get in the logical form of simpler sentences like Ingrid is always on the phone, a constituent that contains an index: (8)
a. Ingrid is always on the phone. b. [. . . alwaysi . . .] [γ . . . i . . . Ingrid on the phone]
On the basis of this, one can argue that to make the right predictions about the semantics of sentences like (7a) and (8a), the constituents β and γ must be interpreted as follows: (9)
[[ β ]]g = λ w. if, in w, Ingrid is on the phone for the duration of g(i) then, in w, John for the duration of g(i) believes Ingrid to be on the phone; and if, in w, Ingrid is not on the phone for the duration of g(i) then, in w, John for the duration of g(i) believes Ingrid not to be on the phone (abbreviated: λw. in w, for the duration of g(i) John knows whether Ingrid is on the phone)
(10)
[[ γ ]]g = λw. in w, Ingrid is on the phone for the duration of g(i)
It is important to note here that, in the semantics of β, g(i) plays a role in determining the duration of John’s beliefs. Now, on common assumptions about semantic composition, to say that the matrix clause in (7b) does not contain an indexed item amounts to saying that the matrix clause material behaves semantically like a function that, given the denotation that you get for γ , will yield the denotation that you get for β. That is, it will behave like a function that, for any arbitrary assignment g and index i such that g(i) yields a
182
Presupposition and Implicature in Compositional Semantics
time interval, will take a proposition of the kind in (10) and give you a proposition of the kind in (9): (11)
The matrix clause material ‘behaves like’ some function F with the characteristic that, for any time interval y, F( λw. in w, Ingrid is on the phone for the duration of y) = λw. in w, for the duration of y John knows whether Ingrid is on the phone.
But at this point problems arise. Not only is it implausible that the matrix clause behaves like this, assuming that the matrix clause behaves like this seems to make wrong predictions. Therefore, we do not want to say that the matrix clause in (7b) lacks an indexed item. The wrong predictions that the proposal in (11) makes are as follows. Parallel to the pair of sentences in (7) and (8), we have pairs of sentences like those in (12a) and (13a), for which we would posit parallel logical forms: (12)
a. John always knows whether at 5pm on the following day Ingrid is on the phone. b. [. . . alwaysj . . .] [β . . . John know whether . . . [δ ]]
(13)
a. At 5pm on the following day Ingrid is always on the phone. b. [. . . alwaysj . . .] [δ . . . j . . . at 5pm on the following day Ingrid on the phone ]
Given the way (13a) is interpreted, the right semantics for δ seems to be: (14)
[[ δ ]]g = λw. in w, Ingrid is on the phone at 5pm on the day after g(j)
Now by assumption to obtain [[β]]g we apply the function F to [[δ]]g . So take an arbitrary assignment g and index j. What will F yield for the proposition in (14)? F will yield: (15)
[[ β ]]g = F([[ δ ]]g ) = λ w. in w, at 5pm on the day following g(j), John knows whether Ingrid is on the phone
A Pragmatic Constraint on Adverbial Quantification
183
Glossing over some steps, the consequence will be that a sentence like (12a) should express that all relevant times are such that John knows one day later at 5pm whether Ingrid is on the phone. But it doesn’t express that. It expresses that all relevant times are such that John knows then whether one day later at 5pm Ingrid is on the phone.
2 A constraint on the use of sentences with adverbial quantifiers 2.1
The basic idea
Let us return to the sentence in (3), repeated in (16). Why can we use it to convey the nationality of the fast students in the first scenario, but not in the second? (16)
The student who finished first was always Swedish.
My diagnosis starts from the idea that (16), roughly speaking, quantifies over time intervals. Just what time intervals it quantifies over depends on what set of time intervals is salient, but, in the scenarios we have chosen, where the set of exam days is salient, it quantifies over all the days in a certain set that extends from Monday to Saturday. It says that, on each of these days, the student who finished first on that day was Swedish. More precisely, the logical form of (16) – the structure that serves as input to semantic interpretation – gets interpreted in such a way that it expresses that, on each of the exam days, the student who finished first on that day was Swedish. But if (16) expresses that on each of the exam days the fastest student was Swedish, why can we use the sentence to convey this on our first scenario but not on our second? I suggest that it has to do with the following fact. A speaker who utters (16) in our second scenario, where we know that the same student finished first on more than one occasion, is quantifying over more things than he needs to in order to make his point. In using the sentence in (16), he is stating that all of the times in one set have a certain character – but he knows that we could have drawn this conclusion if he had informed us that all of the times in a smaller set had that character. Specifically, in using the sentence in (16) he is stating that all members of { Monday, . . . , Saturday } are such that the fastest student on that day was Swedish. But he knows that we could have drawn this conclusion if he had informed us that all of the times in a smaller set have this character. One such smaller set is just { Tuesday }.
184
Presupposition and Implicature in Compositional Semantics
The conjecture that I will pursue, then, is that, at least when it comes to sentences with adverbial quantifiers like always, there is something wrong with quantifying over more things than you need to in order to make your point. I propose that a principle of pragmatics tells us not to use a sentence like (16) – or, more precisely, a logical form like (16)’s – to quantify over the times in one set when we know that the hearer could draw the conclusion that the sentence is true by considering a parallel quantification over a smaller set. A sentence like (16) will sound funny if we think that the speaker who used it violated that principle. Judgments of the sentence in (17) lend support to the idea that it is infelicitous to use sentences with adverbial quantifiers when the result is that we quantify over more things than we need to in order to make our point. The judgments of (17) are parallel to the judgments of (16). In the context of our first scenario, it seems a reasonable sentence for the speaker to utter in order to communicate something (negative) about the nationalities of the fastest students. In the context of our second scenario, it does not. A speaker who used (17) in this context, I suggest, would be claiming that none of the members of { Monday, . . . , Saturday } are such that the fastest student on that day was Swedish. But in the case of the second scenario, the speaker is aware that, to get the addressee to draw this conclusion, he could confine himself to the claim that none of the members of a smaller set – say the singleton { Tuesday } – has the relevant property. (17)
The student who finished first was never Swedish.
In the rest of this section, I will develop a more precise formulation of the pragmatic principle that I have in mind, and in the rest of this chapter I will examine some of its consequences. It is worth noting that I will only concern myself with sentences that contain the quantifiers always and never, and one might well ask how applicable the principle is to sentences that contain different adverbial quantifiers. One remark before we proceed. I claimed here that infelicity arises in cases where a speaker quantifies over the times in a certain set – like the set of exam days – and could convey the same thing with a parallel claim in which the original set is replaced by a subset. At the same time, the specific cases of infelicity that we have encountered are all cases where it is sufficient to consider a singleton subset – like { Tuesday }. Was there a reason to make the more general claim? Yes: cases where it is not sufficient to consider a singleton subset are also bizarre. I think, however, that there is a gradation of acceptability. For instance, when
A Pragmatic Constraint on Adverbial Quantification
185
we consider variations on the scenario that we have been considering, we find that, the smaller the subset of exam days that we can replace the original set by, the stranger a sentence like (16) is. Imagine, for example, a variation on which we see one student finishing first every day from Monday through Wednesday, and a different student finishing first every day from Thursday through Saturday ((18i)). In this new case, we can’t convey what (16) conveys by saying that the fastest student on a single exam day was Swedish, but we can by saying that the fastest student on each of a certain two exam days – say Tuesday and Friday – was Swedish. I think it is still quite strange to utter (16) in this scenario, but perhaps not quite as strange as in the scenario on which the same student finishes first each time. When we alter the scenario still further, so that three students are involved ((18 iv)), (16) gets better still. And so on. (18)
Other scenarios
M T W Th F S i. ii. iii. iv. v.
2.2
a a a a a
a a a a a
a a a b b
# exam days needed for the relevant kind of subset
Judgment
two two two three four
??? ?? ?? ?? ?
b b b a a b b a a b c c c d a
Background assumptions
My discussion of the pragmatic principle will take for granted the familiar picture of semantic interpretation that I have so far implicitly assumed. When we compute the denotation of a sentence’s logical form (LF) with respect to an assignment, what we obtain is a function from worlds to truth values. When we say that a sentence is true, we are saying that we can find an LF for it, and an assignment, such that the actual world is characterized by the function we obtain. My discussion will also rely on some specific assumptions that I will make about the LFs available for sentences like (16) and about the way these LFs are interpreted. These assumptions are intentionally simplified, in a way that will affect the terms of the discussion but hopefully not its essence. First, as far as the syntax of sentences like (16) is concerned, I will be consistent with what I have been assuming until now. I will assume that in the logical forms of these sentences all items have reconstructed to below the VP level, and I will be assuming that there is nothing
186
Presupposition and Implicature in Compositional Semantics
interpretable above always, which is adjoined to VP, so that everything above always can be ignored. (I am thus ignoring tense nodes, and tense information.) As before, I will assume that adverbial quantifiers like always are indexed. The interpretable pieces of (16)’s LF are thus as in (19). (19)
always1 [α
]
When it comes to my assumptions about how LFs like (19) are interpreted, I will be departing slightly from an idea that I alluded to earlier. Earlier (cf. (1’b)), I assumed that the context makes salient a property of time intervals, and that we use this property to determine the intervals that always quantifies over. Now, I will assume that the context makes salient a set of time intervals, and that we interpret quantifiers like always as quantifying over the members of this set. Specifically, an LF like (19) will be interpreted as in (20). This simplification will strongly affect the terms of the coming discussion. (20)
[[
alwaysi [α ] ]]g = λw. for all time intervals t in the contextually salient set S of time intervals, [[α]]g i−>t (w) = 1.
Assumptions like these allow us to draw conclusions about other aspects of the semantics of (16). Consider once again sentence (16) as uttered in the first scenario, on which the student who finishes first is different each time. (21)
Exam date: M Fastest student: a
T b
W c
Th F d e
S f
The fact is that, if we know that the students who finished were all Swedish-born, we would say that the sentence is true. If we assume that the LF of (16) conforms to the format in (19), then we might draw the following conclusions from this. First, the context makes salient a series of exam days ((22a)) – specific time intervals during which the actual exams occurred. Second, in the LF of (16), the denotation of the constituent that combines with always is as in (22b). This way, we will account for that fact. As I indicated in my informal discussion at the beginning of this section, I will assume in what follows that these are the right conclusions to draw.
A Pragmatic Constraint on Adverbial Quantification
(22)
187
a. S = { Monday, Tuesday, Wednesday, Thursday, Friday, Saturday } b. [[ α ]]g = λw. the individual who in w is the student first to finish the exam held in g(1) is Swedish-born in w.
2.3
The constraint
The principle of pragmatics that I propose regulates the sentences that a speaker can make use of by regulating the logical forms that he can make use of. Given my assumptions so far, it is natural to suppose that, when a speaker uses a sentence to make an assertion, he has in mind a particular logical form for the sentence (call it L) and a particular assignment (call it g), and his purpose in uttering the sentence is (among other things) to convey that [[L]]g characterizes the actual world. Let’s assume this. The principle that I have in mind narrows down what a speaker can hope to convey in this way by declaring some logical forms, on some occasions, off limits. Some sentences have only one LF, some have more than one. The principle will effectively ban a speaker from asserting a sentence if, for example, the sentence has only one LF, and the principle blocks the use of this LF. To talk about LFs, I will use the following informal terminology. First of all, consider again the way an LF with always gets interpreted: (23)
[[ alwaysi [α ] ]]g = λw. for all time intervals t in the contextually salient set of time intervals, [[ α ]]g i−>t (w) = 1.
Informally speaking, LFs of the form [alwaysi α] quantify over members of the contextually salient set of time intervals. Let’s call this set the domain of quantification for [[ alwaysi α]]g . Now, for every LF of the form [alwaysi α ] and assignment g, we can imagine an alternative function that differs only with respect to what set is being quantified over: (24)
a. λw. for all time intervals t in set σ 1, [[α]]g i−>t (w) = 1. b. λw. for all time intervals t in set σ 2, [[α]]g i−>t (w) = 1. c. λw. for all time intervals t in set σ 3, [[α]]g i−>t (w) = 1. ...
Let’s call these functions domain-variants of [[ alwaysi α ]]g . (24a), for instance, is a domain-variant of [[ alwaysi α ]]g with domain σ 1. To the extent that LFs with other adverbial quantifiers behave analogously to LFs with always ((25)), we will be able to talk analogously
188
Presupposition and Implicature in Compositional Semantics
about domain-variants of [[QUANTi α]]g , where QUANTi stands for an adverbial quantifier. (25)
[[ QUANTi α ]]g = λw. for proportion δ of the time intervals t in the contextually salient set of time intervals, [[ α ]]g i−>t (w) = 1.
With this in mind, here is a stab at the principle that constrains a speaker’s choice of LF. It is a rule for speakers to follow: (P)
Rule4 : (For any g), Do not use an LF of the form [QUANTi α ] to express that [[QUANTi α]]g holds of the actual world when you can find a domain-variant of [[QUANTi α]]g , , with the following characteristics: i. the domain for is a proper subset of the domain for [[QUANTi α]]g ii. it follows from what the parties to conversation are taking for granted about the actual world that [[QUANT i α]]g holds of the actual world as long as holds of it.5
On its own, of course, this principle does not explain why sentences sound strange. Over and above this, I assume that a hearer can reflect on whether the assertion of a sentence obeys (P) or not, and will find the sentence strange if it does not. For concreteness, I propose (Q) (both aspects of which hopefully derive from more general aspects of the way we judge sentences). (Q)
a. A sentence whose only LF is of the form [QUANTi α] will sound strange if we think that the use of this LF (together with any assignment) violates (P). b. If we know that the speaker can find a that has the relevant characteristics and that has only one element in its domain, then the sentence will sound terrible. Less terrible if two, etc.
Now here is how principle (P) will apply to the use of (16) on our second scenario. By assumption, (16) has only one kind of LF – the one given in (19) – and the denotation of this LF with respect to an arbitrary assignment g is repeated in (26). (The denotation will be the same no matter what assignment the speaker chooses. In what follows I will sometimes
A Pragmatic Constraint on Adverbial Quantification
189
write ‘[[ (19) ]]g ’ to mean this one object that is the denotation of (19) under any assignment.) (26)
a. [[ (19) ]]g = λw. For all days t in S, the individual who in w is the student first to finish the exam held in t is Swedish- born in w. b. S = {Monday, Tuesday, Wednesday, Thursday, Friday, Saturday}
One possible domain-variant of [[ (19) ]]g is the one given in (27). It so happens that the domain of is { Tuesday }, which is a proper subset of the domain of [[ (19) ]]g . So (P) tells us that, if it follows from what the parties to conversation are taking for granted about the actual world that [[ (19) ]]g holds of the actual world as long as does, then the speaker is banned from using (19). (27)
One possible domain-variant: = λw. For all days t in {Tuesday}, the individual who in w is the student first to finish the exam held in t is Swedishborn in w. ( = λw. The individual who in w is the student first to finish the exam held on Tuesday is Swedish-born in w. )
Now, on our second scenario, this does plausibly follow from what the parties to conversation are taking for granted. The important aspect of this scenario was that both parties to conversation were present at all the exams and saw that the same student finished first each time. This makes it likely that the following is being taken for granted: (28)
For all t1, t2 in S, the individual who in w0 is the student first to finish the exam in t1 is the individual who in w0 is the student first to finish the exam in t2.
Given (28), if (w0) = 1 then we can conclude that [[ (19) ]]g (w0) = 1 and if (w0) is not 1 then we can conclude that [[ (19) ]]g (w0) is not 1. So, if (28) is indeed being taken for granted, (P) will prohibit the use of our LF (19). And since by assumption this is the only LF that the sentence in (16) has, (P) will effectively prohibit the use of the sentence in (16). What will be the consequence? If we think that (28) is being taken for granted (or more precisely that the speaker thinks that it is), then we will think that the speaker is violating (P) by using (16). Accordingly, the utterance of (16) will sound strange. And given that the relevant
190 Presupposition and Implicature in Compositional Semantics
domain-variant had only one element in its domain, the utterance of (16) will sound very strange. In what follows, in discussing examples like these, I will sometimes take an expository shortcut: I will draw a diagram that represents the information that the parties to conversation are taking for granted. In the case of the scenario we just considered, the diagram would look like this: (29) M
T
W
Th
F
S
-------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|---|-----------|------------------>
a is first
a is first
a is first
a is first
a is first
a is first
I will only be considering LFs whose denotations are the same with respect to all assignments. To show that (P) blocks a particular LF, I will show that: (i) if we add to the diagram the information that a certain domain-variant characterizes the actual world, then from the information on the diagram we can conclude that the denotation of the LF characterizes the actual world; (ii) if we add to the diagram the information that the domain-variant does not characterize the actual world, then from the information on the diagram we can conclude that the denotation of the LF does not characterize the actual world. For example, if we add to our diagram the information that characterizes the actual world (30)
a. M
T
W
Th
F
S
-------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|---|-----------|------------------>
a is first
a is first
a is Swedish-born
a is first
a is first
a is first
a is first
A Pragmatic Constraint on Adverbial Quantification
191
we now have enough information to conclude that [[ (19)]]g characterizes the actual world. If instead we add the information that does not characterize the actual world b. M
T
W
Th
F
S
-------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|---|-----------|------------------>
a is first
a is first
a is first
a is first
a is first
a is first
a is not Swedish-born
we have enough information in the diagram to conclude that [[ (19) ]]g doesn’t either.
3
Some consequences of this constraint
I will now demonstrate some of the consequences of the pragmatic constraint that I have identified. I am going to go through a number of sentences which have LFs of the form [QUANTi α ], and see what predictions we make about the felicity of these sentences. If the approach that I have been taking is correct, then there are a number of factors that play a role in rendering infelicitous a sentence with an LF of the form [QUANTi α]. One is the denotation of α, because that will play a role in determining the denotation of the whole LF and therefore its possible domain-variants. Another is the specific set of times that the context makes salient, because that is what functions as the domain of [[QUANTi α]]g . Another is what we think is being taken for granted about the world, since this will play a role in determining what we can conclude about the truth of the original sentence given the truth of a domain variant. So demonstrating the predictions that (P) makes will potentially require us to take a stand on all of these factors. We will see, however, that in a large number of cases, just taking a position on one or two of these factors will be enough to make a prediction. In examining the predictions that (P) leads to, there is a further tool that I will make crucial use of. I am going to take it for granted that apart
192
Presupposition and Implicature in Compositional Semantics
from (P) there is an additional constraint on the use of quantifiers like always and never – you are not allowed to use [QUANTi α ] (together with assignment g) if the domain of quantification for [[QUANTi α]]g contains fewer than four members. That is: (31)
Only use [ QUANTi α ] (to express that [[QUANTi α]]g holds of the actual world) if the contextually salient set of time intervals contains more than three members.
The precise way I formulated this constraint depends on the simplified assumptions that I have made, but I believe there is good evidence for some constraint along these general lines.6 Consider for example our old sentence (16) (= (32a)) in some variants of our first scenario. The sentence sounds very odd when there are only two exam days instead of six ((32b)); it still sounds a bit odd (though less odd) when there are three ((32c)); when there are four ((32d)), it doesn’t sound bad at all. (31) captures these facts if, as I have been assuming so far, the set of exam days is the contextually salient set of time intervals. (32)
a. The student who finished first was always Swedish. b. M T ??? c. M T W ? d. M T W Th not bad a
b
a
b
c
a
b
c
d
3.1 Ingrid was always Swedish The first case that I will consider is the case of the sentence we started out with ((33a)). I will assume that the only LFs (33a) has are of the kind we have considered so far, one where always combines with another constituent ((33b)). (33)
a. Ingrid was always Swedish. b. alwaysi [α . . .]
What we assume to be the denotation of this other constituent will obviously play a role in when we predict the LF in (33b), and thus the sentence in (33a), to be usable. I will assume the following about the constituent that always attaches to.7 I will assume that the constituent contains an index but the only purpose of this index is to restrict the domain of the function that we get out of this constituent. (This assumption isn’t innocuous, and we will see that the contribution of the index plays a
A Pragmatic Constraint on Adverbial Quantification
193
role in our reasoning about when (33a) will be felicitous.) The denotation of this constituent with respect to an assignment will be specifically as in (34b). (34)
a. alwaysi [α . . . i ] b. [[ α ]]g = λw: Ingrid is alive in w for the duration of g(i).8 Ingrid is Swedish-born in w.
The result is that the LF of (33a) will have a denotation as in (35), where S is the contextually salient set of time intervals, (35)
[[ (33b) ]]g (w) = 1 as long as for all t in S, Ingrid is alive in w for the duration of t and Ingrid is Swedish-born in w.
and accordingly domain-variants of [[ (33b) ]]g will behave as follows: (w) = 1 as long as for all t in σ, Ingrid is alive in w for the duration of t and Ingrid is Swedish-born in w. Given this, we can derive (by making an additional assumption) that, no matter what exactly the salient time intervals are, a speaker is not permitted to utter (33a). This in turn means that a sentence like (33a) should always sound bizarre. Since it is hard to imagine an utterance of (33a) that sounds sensible, this prediction seems right. The reasoning is as follows. LFs of the kind in (33b) are the only ones that (33a) admits, so if a speaker is prevented from using those he can’t use (33a). Now, it so happens that an independent principle of grammar (I will elaborate briefly below) guarantees that a speaker is only allowed to use (33b) when it is already established that Ingrid is alive at all the intervals in the contextually salient set. So let us just consider what would happen if it is already established that Ingrid is alive at all the intervals in the contextually salient set. Here there are two kinds of cases to consider: cases where there is more than one interval in the contextually salient set, and cases where there is only one. But in both of these cases, we can show, other factors bar the speaker from using (33b). Consider first the case where there is more than one interval in the contextually salient set. In this case, a speaker who uses (33b) will violate (P). To see this, imagine first the information that the parties to conversation
194 Presupposition and Implicature in Compositional Semantics
are taking for granted: (36)
S = { t1 , . . . , tn }
a. t1
t2
t3
t4
t5
…
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
Ingrid is alive
Ingrid is alive
Ingrid is alive
Ingrid is alive
Ingrid is alive
Now consider the domain-variant of [[(33b)]]g whose domain consists only of t1 , the first time interval in the contextually salient set. (Call it again.) characterizes a world w as long as for all t in {t 1 }, Ingrid is alive in w for the duration of t and Ingrid is Swedish-born in w. That is, it characterizes a world w as long as Ingrid is alive in w for the duration of t1 and Ingrid is Swedish-born in w. Suppose we add to the diagram the information that characterizes the actual world. This amounts to adding: b.
Ingrid is Swedish-born
and now from the information in the diagram we can conclude that [[(33b)]]g characterizes the actual world. Suppose instead we add to the diagram that doesn’t characterize the actual world. The only way of doing this consistently with the information already there is to add c.
Ingrid is not Swedish-born
and from this we can conclude that [[(33b)]]g does not characterize the actual world. Now consider by contrast the case where there is only one interval in the contextually salient set. In this case, we will not be able to find a relevant domain-variant of [[(33b)]]g whose domain is a proper subset of [[(33b)]]g ’s domain, so (P) will not be violated. But the principle in (31) that guarantees a domain of more than three intervals will be, so even in this case the speaker will be prevented from using the LF in (33b). The net result: a speaker is barred altogether from using the LF in (33b), and by extension the sentence in (33a).
A Pragmatic Constraint on Adverbial Quantification
195
The independent principle of grammar that I made use of in the course of my reasoning here is one that has been identified in discussions of ‘presupposition projection’ (e.g. Heim, 1983).9 Recall that the denotation of always’ sister comes with a domain condition ((34b)). This principle limits the conditions under which a speaker can use an LF where such a constituent combines with a quantifier: for a speaker to use the LF, it must be taken for granted that a certain property related to the domain condition holds of every item in the set the quantifier ranges over. Here, where the quantifier ranges over time intervals, every one of these time intervals must be one during which Ingrid is alive. (37)
(38)
Consequence of one view of ‘presupposition projection’ and (34b): A speaker is entitled to use (33b) to express that [[ (33b) ]]g holds of w0 only when (38) is taken for granted. For every t in the domain of quantification for [[ (33b) ]]g , Ingrid is alive in w0 for the duration of t.
To conclude this discussion, let me summarize the role (P) played in predicting the unacceptability of (33a). We found that, no matter what set of intervals the context makes salient, any speaker who uses the LF in (33b) while respecting the independent principle of grammar alluded to in (37), and while respecting the principle that the domain of quantification must contain more than three intervals, will wind up violating (P). Since LFs of the kind in (33b) are the only ones that (33a) admits, a speaker will never be able to felicitously utter (33a). 3.2
Napoleon was always dead
The next case I want to consider is (39). Here as before, I will assume that there is just one possible kind of LF, one in which always attaches to another constituent, and that this constituent is interpreted as in (40b). (39) (40)
Napoleon was always dead. a. alwaysi [α . . . i ] b. [[ α ]]g = λw. Napoleon is dead in w for the duration of g(i).
This means that the LF of (39) will have a denotation as in (41), where S is the contextually salient set of time intervals. (41)
[[(40a)]]g = λw. for all t in S, Napoleon is dead in w for the duration of t
196 Presupposition and Implicature in Compositional Semantics
and accordingly domain-variants of [[(40a)]]g will be of the form λw. for all t in σ , Napoleon is dead in w for the duration of t. The fact is that, as with the previous sentence, it is hard to imagine an utterance of (39) that does not sound bizarre. My feeling is that when we try to imagine a speaker uttering the sentence, we have the impression that the speaker who utters it expects that at any moment Napoleon could have come back to life. That a speaker should expect something like this itself seems bizarre. What we can derive on these assumptions about the syntax and semantics of (39) is that (39) will be unusable when it is taken for granted that anyone who is dead at one point is dead at all later points ((42)) – again, irrespective of what exactly those time intervals are that the context makes salient. This is because, no matter what set of intervals the context makes salient, any speaker who uses the LF in (40a) while (42) is taken for granted, and who respects the principle that the domain of quantification must contain more than three intervals, will wind up violating (P). Since LFs of the kind in (40a) are the only ones that (39) admits, as long as (42) is taken for granted, a speaker will never be able to felicitously utter (39). (42)
Fact about the actual world w0: For all x, t, t , if x is dead in w0 at t and t contains no moment that precedes t then x is dead in w0 at t .
To see this, imagine that the contextually salient set of time intervals contains more than three intervals, and again imagine the information that is being taken for granted: (43)
S = { t1 , . . . , tn }
a.
t1
t2
t3
t4
t5
…
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
if N is dead now, he is dead later too
if N is dead now, he is dead later too
if N is dead now, he is dead later too
if N is dead now, he is dead later too
if N is dead now, he is dead later too
A Pragmatic Constraint on Adverbial Quantification
197
Now consider the domain-variant (call it again) whose domain consists only of t1 , the first time interval in the contextually salient set. Suppose we add to the diagram the information that characterizes the actual world: b. t1
t2
t3
t4
t5
…
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
if N is dead now, he is dead later too
if N is if N is dead now, dead now, he is dead he is dead later too later too
if N is dead now, he is dead later too
if N is dead now, he is dead later too
N is dead now
From this we can conclude that [[(40a)]]g characterizes the actual world. Suppose instead we add the information that does not characterize the actual world: c. t1
t2
t3
t4
t5
…
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
if N is dead now, he is dead later too
N is not dead now
if N is if N is dead now, dead now, he is dead he is dead later too later too
if N is dead now, he is dead later too
if N is dead now, he is dead later too
198
Presupposition and Implicature in Compositional Semantics
From this we can conclude that [[(40a)]]g does not characterize the actual world. Here as in the previous case, the conclusion that (39) is unusable depends in part on the idea that the parties to conversation are accepting certain assumptions about the world. The relevant assumption here is that once you are dead, you’re dead. Unlike in the previous case, however, there is no principle that links the acceptance of this assumption to the fact that the speaker has decided to use (39). Therefore, we predict that in cases where it is clear that the parties to conversation are not making this assumption, the sentence might not seem strange. Although we generally do accept that once you are dead, you’re dead, stories about resurrection or about supernatural beings that come back to life suggest that one might choose not to accept this. Here is a case parallel to (39) in which the speaker specifically disavows the assumption that once you are dead, you’re dead. The relevant clause here does not sound so bizarre, so this at least is consistent with the story we have told. (44)
Vampires are supposed to come back to life after midnight, but I checked the coffin every half hour and I can assure you that this vampire was always dead.
A variant of this case that is worth considering is the example in (45), whose LF, I assume, works in a parallel way ((46)-(47)). (45) (46) (47)
Napoleon was always alive. a. alwaysi [α . . . i ] b. [[ α]]g = λ w. Napoleon is alive in w for the duration of g(i). [[(46a)]]g = λ w. for all t in S, Napoleon is alive in w for the duration of t
Here we again derive that the sentence will be unusable when the speaker is taking for granted that anyone who is dead at one point is dead at all later points – irrespective of what exactly those time intervals are that the context makes salient. However, this example is slightly different from the earlier one. In the absence of accepted knowledge about whether Napoleon was born yet at the time of the first interval, the kind of domain-variant that we need to consider in order to demonstrate a (P) violation is different from the kind we needed to consider earlier. It is one with a larger domain. Specifically, its domain must consist of two intervals: the first interval and the last interval of the contextually salient set. Now, the bizarreness of our old example (39) when judged in
A Pragmatic Constraint on Adverbial Quantification
199
the absence of context suggests that in the absence of context we seem by default to take it for granted that once you are dead you’re dead. So, given the proposal thus far, one might expect that this sentence too will sound bizarre when judged in the absence of context – but perhaps a little less bizarre. That seems to be the prediction. 3.3 John always knows whether Ingrid is Swedish In the case of the sentences we just looked at, anytime the parties to conversation are taking for granted the kind of assumptions that we all generally make about the world (for instance, that once you’re dead you’re dead), the speaker will violate (P). And this is true irrespective of what set of time intervals it is that the context makes salient. Accordingly, as long as we have no reason to think that the parties to conversation are departing from these normal assumptions,10 we don’t need to make any additional commitments as to what the salient set of time intervals is in order to reject the sentence as bizarre. At least, that is what follows from the picture I have presented so far. The case of (48) (= (5)) is different. (I will assume as before that (48) has just one kind of LF – given in (49a) – and that the denotation of this LF is as in (50).) (48) (49)
(50)
John always knows whether Ingrid is Swedish. a. alwaysi [α . . . i ] b. [[ α ]]g = λ w. in w, John knows for the duration of g(i) whether Ingrid is Swedish11 [[ (49a) ]]g = λw. for all t in S, in w John knows for the duration of t whether Ingrid is Swedish
We would be able to say the same thing about (48) if we could maintain, for instance, that a normal assumption that we make about the world is that, once you know whether Ingrid is Swedish, you retain that knowledge forever. But I think that is not the kind of assumption that we normally make. We normally imagine that knowledge of this kind tends to fade, especially if you are never called upon to access it (‘I used to know but I forgot’). In this case, if indeed the sentence sounds bizarre to us, that must have something to do with assumptions we are making about the set of time intervals that is being quantified over. Here is the kind of situation in which we predict that a speaker who uses (48) will violate (P). We predict a violation of (P) when it is taken for granted that the distance between intervals is smaller than the amount of time that it would typically take for John’s memory to fade. The idea
200 Presupposition and Implicature in Compositional Semantics
is this: suppose it is taken for granted that, if John knows at interval t2 whether Ingrid is Swedish, then he knows this also at the succeeding interval t3 . S = { t1 , . . . , t5 }
(51) t1
t2
t3
t4
t5
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
John’s memory will not fade between now and the end of t3
Then to determine whether he knows at all of the intervals in S whether Ingrid is Swedish, it is enough to determine whether he knows at all the intervals except t 3 whether Ingrid is Swedish. Consequently, we can show that the use of the LF (49a) violates (P) by considering a domain variant of [[ (49a) ]]g whose domain consists of all the intervals except t3 . What this means is that, to the extent that we feel that the sentence is bizarre when we have no clue as to what the relevant time intervals are, we must be making some additional assumption along these lines: the distance between intervals is smaller than the amount it would take for John’s memory to fade. On the approach that I have been taking, this is the conclusion we are led to. To test whether the approach is on the right track, what we should do is consider two kinds of scenarios, one where it is clearly taken for granted that the distance between intervals is smaller than the amount of time it takes for John’s memory to fade, and one where it is clearly taken for granted that it is larger. A sentence like (48) should seem odd in the first scenario, but all right in the second. In fact, it is hard to think of any scenario where the kind of information that (48) conveys might be of interest, but the contrast between the following two dialogues might bear out the prediction. I think that the final sentence of the first dialogue sounds a lot worse than the final sentence of the second dialogue (where a natural paraphrase would use ‘remember’ instead of ‘know’). When
A Pragmatic Constraint on Adverbial Quantification
201
we hear the first dialogue, our first impression is that ‘always’ is just redundant. If we try to justify it to ourselves, then if anything we have the strange feeling that the speaker thinks that the relevant information might suddenly vanish from John’s mind, and when we know nothing else about John this comes as a surprise.12 (52) – I heard that Ingrid is Scandinavian, and I wanted to know whether she is Swedish. – For some reason this question comes up a lot. I always send people to John to find out. He isn’t very knowledgeable, it’s true, but he does always know whether Ingrid is Swedish. (53) – I heard that Ingrid is Scandinavian, and I wanted to know whether she is Swedish. – For some reason this question comes up a lot. I always send people to John to find out. He has a lousy memory for many things, but he does always know whether Ingrid is Swedish.
4
Consequences for ‘semantic partition’
Here is a very general pattern of reasoning. Premise: We do not use a sentence S to express the proposition that would be derived from a logical form L.13 Conclusion: L is not a possible logical form for S. Now that we have seen that principles of pragmatics can prevent us from using an LF, we can see that this kind of reasoning is questionable. If we do not use a sentence to express the proposition that would be derived from L, that could be because other principles prevent us from using L. It doesn’t have to be because our mechanism for generating LFs prevents us from generating L for the sentence in question. One place where this questionable pattern of reasoning has been used is in the treatment of sentences like Ingrid was always Swedish. We saw that there an independent principle of pragmatics, (P), could explain why we do not use an LF that contains an item coindexed with the quantifier. I want now to look at another case where this pattern of reasoning has been used. Perhaps again, instead of saying that the sentence lacks an LF that we might otherwise expect, we can say that in principle we can generate this LF but in practice (P) will typically prevent us from using it. The sentences that I am interested in are sentences with ‘individual-level predicates,’ and I am interested in the way they behave with respect to ‘semantic partition.’ Work by Kratzer (1995) and Diesing (1992) gave rise to a view which one can characterize as follows. Sentences with indefinite subjects like (54) have among their possible LFs one that gives rise to a proposition
202
Presupposition and Implicature in Compositional Semantics
like (55). The same kind of LF is not, however, available for sentences like (56).14 Behind this view is an argument that follows our familiar pattern of reasoning: We do not use (56) to express the proposition that we would derive from the relevant LF. Therefore, the sentence does not have the relevant LF. (54) (55)
(56)
A secretary is always on the phone. λw. For all intervals t in the set of intervals that the context makes salient, there is some individual who in w is a secretary for the duration of t and who in w is on the phone for the duration of t A secretary is always Swedish.
The kind of LFs that we are interested in are LFs like those in (57) and (58), or minor variants of these. I have sketched in (59)–(61) a few aspects of the way these LFs are interpreted. (Some things that are worth noting are that the indexed e items are silent elements that function as variables over times, and that the DP functions as a quantificational expression with existential force.) (57) (58) (59)
(60)
(61)
always1 [VP [DP a2 secretary e1 ] [t2 on the phone e1 ] ] always1 [VP [DP a2 secretary e1 ] [t2 Swedish e1 ]] ]]g [[ [ [DP ai secretary ej ] α ] = λw. there is some individual x such that, in w, x is a secretary for the duration of g(j) and such that [[ α ]]g i−>x (w) = 1. [[ VP(57) ]]g = λw. there is some individual x such that, in w, x is a secretary for the duration of g(1) and such that, in w, x is on the phone for the duration of g(1) [[ VP(58) ]]g = λw. there is some individual x such that, in w, x is a secretary for the duration of g(1) and such that x is Swedish-born in w
The propositions that we would get out of these LFs are given in (62) and (63). The LF in (57) would say that we can find a secretary on the phone at each of the contextually salient time intervals; the parallel LF in (58) would say that we can find a Swedish secretary at each of the contextually salient time intervals. Now, the position I am examining has it that, while we can generate the LF in (57) for the sentence A secretary is always on the phone, we cannot generate the parallel LF in (58) for the sentence A secretary is always Swedish. Is this position justified?
A Pragmatic Constraint on Adverbial Quantification
(62)
(63)
203
[[ (57) ]]g = λ w. For all intervals t in the set of intervals that the context makes salient, there is some individual who in w is a secretary for the duration of t and who in w is on the phone for the duration of t [[ (58) ]]g = λ w. For all intervals t in the set of intervals that the context makes salient, there is some individual who in w is a secretary for the duration of t and who is Swedish-born in w
The point that I want to make is that, even if A secretary is always Swedish does have the LF in (58), the use of this LF will be very restricted. As long as we take it for granted that at two of the contextually salient time intervals exactly the same people are secretaries, (P) will prevent us from using it. Suppose for instance that it is accepted that the pool of secretaries is the same from, say, t2 to the end of t3 ; then (P) will prevent a speaker from using (58) by virtue of the domain-variant whose domain simply excludes t3 . This is because, to determine whether there is a Swedish-born secretary at each of the relevant intervals, it is enough to determine whether there is a Swedish-born secretary at every interval but t 3 . (By contrast, to determine whether there is a Swedish-born secretary on the phone at each of the relevant intervals, it is not enough to determine whether there is a Swedish-born secretary on the phone at every interval but t3 . So (P) will not prevent a speaker from using (57) in the same situation.) In fact, if the contextually salient time intervals are close together, it is pretty reasonable to imagine that the pool of secretaries will stay the same from one time interval to the next. contextually salient set: { t1 , . . . , t5 }
(64)
t1
t2
t3
t4
t5
------|-----------|---|-----------|---|-----------|---|-----------|---|-----------|----------->
the pool of secretaries is the same from here to the end of t3
204
Presupposition and Implicature in Compositional Semantics
What does this mean? It means that, to the extent that we do not naturally take the sentence to express the proposition in (63) in the absence of information about the relevant time intervals, this might not be because the sentence lacks the LF in (58). It could just be because, in the absence of contrary information, we tend to imagine that the pool of secretaries will stay the same from one interval to the next – maybe because we tend to imagine that the intervals are close together. This kind of default assumption would be analogous to the kind of default assumption we considered at the end of the last section: that, in the absence of contrary information, we imagine that knowledge of whether Ingrid is Swedish will stay present from one interval to the next. If we really want to know whether the LF in (58) is a possible one, we should at least make sure to look at cases where it is clear that changes in the secretary pool occur between the contextually salient intervals, and cases like these will have to have intervals that are fairly far apart. A candidate case is (65). (To judge by the nods of native speakers at a number of workshops,) it seems that (65) can convey that every change of office staff results in the inclusion of a Swede in the secretary pool, and this suggests that the LF in (58) is possible.15,16 (65)
Every five years, we change office staff entirely. We never wind up with any Scandinavians as technical assistants, but by sheer chance a secretary is always Swedish.
If this line of reasoning is right, there are still some questions to ask. But they have nothing to do with the LFs that (56) admits. The main question is why, when we ask ourselves whether it gives rise to the reading in (63), we have a strong tendency to imagine that the times being quantified over are close together rather than far apart. We seem not to exhibit this tendency, for example, when we ask ourselves what a sentence like (66) expresses. Here, we seem to imagine that there is only one relevant time interval per reign.17 (66)
A blond is always King of Sweden, and a brunette is always Queen.
I want to close this section by mentioning a prediction that we can now make. I just said with regard to sentences like (56) that, when we have to guess what kind of time intervals might be under consideration, unless we know otherwise, we apparently do not imagine that changes in the secretary pool occur between the intervals. That is the conclusion we have to draw if (56) allows the LF in (58). As we saw earlier, we can
A Pragmatic Constraint on Adverbial Quantification
205
draw a parallel conclusion from the fact that a sentence like John always knows whether Ingrid is Swedish sounds bizarre in the absence of information about the time intervals under consideration. The conclusion there is that, unless we know otherwise, we apparently do not imagine that changes in a person’s knowledge as to whether Ingrid is Swedish can occur between the intervals. When we put these two conclusions together with what we have said so far, we make a prediction. Consider the sentences in (67a) and (68a), which in some sense are parallel to the sentences we started out with in this section. On assumptions that can be reconstructed easily enough from the discussion thus far, one possible LF for (67a) – abbreviated in (67b) – yields the proposition in (67c), and similarly one possible LF for (68a) – abbreviated in (68b) – yields the proposition in (68c). The LF for (67a) says that, at each of the relevant time intervals, there is some secretary who knows at the time whether Ingrid is Swedish. The LF for (68a) says that, at each of the relevant time intervals, there is some secretary who knows at the time whether Ingrid is on the phone (at the time). (67)
(68)
a. A secretary always knows whether Ingrid is Swedish. b. always1 [VP [DP a2 secretary e1 ] [t2 knows whether Ingrid is Swedish e1 ] ] c. λw. For all intervals t in the set of intervals that the context makes salient, there is some individual who, in w, is a secretary for the duration of t and knows for the duration of t whether Ingrid is Swedish a. A secretary always knows whether Ingrid is on the phone. b. always1 [VP [DP a2 secretary e1 ] [t2 knows whether Ingrid is on the phone e1 ] ] c. λw. For all intervals t in the set of intervals that the context makes salient, there is some individual who, in w, is a secretary for the duration of t and knows for the duration of t whether Ingrid is on the phone
The prediction is this: in the absence of information about what time intervals are under consideration, it will not be natural to take (67a) to express the proposition that derives from that first LF; by contrast, it should be natural to take (68a) to express the parallel proposition that derives from the second LF. In other words, the “semantic partition” difference that we find between A secretary is always Swedish
206
Presupposition and Implicature in Compositional Semantics
and A secretary is always on the phone should be preserved across this kind of attitude context. (Why do we make this prediction? Because, if at one time interval we can find a secretary who knows whether Ingrid is Swedish, then we should also be able to find such a secretary at another time interval – the very same secretary. So the relevant LF for (67a) will violate (P). But if at one time interval we can find a secretary who knows whether Ingrid is on the phone, there is no guarantee that we can find such a secretary at another time interval. So the relevant LF for (68a) will not violate (P).) This prediction seems to me to be correct. This is of interest because, contrary to what is often claimed, the ability to use LFs like those in (67b)–(68b) apparently does not depend on the identity of the matrix verb: the matrix verb is the same (know) in both LFs, but it looks as though the second LF is usable while the first is not.
5
Concluding remarks
In this chapter, I have argued for a conception of grammar under which principles of use may prevent us from availing ourselves of representations that we nonetheless have the resources to generate. I pointed out specifically that, since principles of use may sometimes block the use of LFs, we have to be careful when arguing that a sentence does not admit such and such an LF. I tried to motivate a particular principle of use, (P), and my concern was to explore its effects. The questions that arise are the usual ones. Are there alternative lines of explanation that would have accounted for the same facts that I used (P) to account for? If indeed a theory that incorporates a principle like (P) is on the right track, is the relevant principle really (P), or is there a better way of formulating it? Does the principle follow from anything? Does it relate in any way to other principles of grammar that we know about? To motivate (P), I used the fact that the sentence The student who finished first was always Swedish is unsuited to communicate the nationality of a certain student of whom it is known that she finished first each time. What other lines of explanation might one pursue to account for this fact? One possibility is that the semantics of the sentence is not what I claimed, and that its interpretation alone renders it incompatible with the situation in question. Alternatively, one might attribute the responsibility for this fact to other principles of use. One position to take, for example, is that it follows as a quantity implicature from the use of the sentence that the same student did not finish first each time. I can’t address all the different options, but I am skeptical of this last position.
A Pragmatic Constraint on Adverbial Quantification
207
To take this position is in part to say that we have a systematic way of generating alternatives to sentences like the one at issue, and that in the case of this sentence, the procedure will yield a sentence that is logically stronger and that entails that the same student did finish first every time. (One such alternative sentence might be: The student who always finished first was (always) Swedish.) I am skeptical because I do not see exactly how this procedure for generating alternatives would work. Assuming something like (P) is on the right track, is the formulation of the principle in need of refinement? Probably. For one thing, some provision has to be made somewhere for the communicative intentions of the speaker: while in the scenario considered it is odd to utter The student who finished first was always Swedish in order to communicate the student’s nationality, it is not so odd to utter the sentence in order to communicate a simple statistical generalization. Maybe a better formulation of the principle would reflect this. Apart from this, I think there is a further inadequacy with (P) and (Q) as they stand now. They predict that sentences of the kind in (69) should have the same status, when my intuition is that (69b) is less bizarre than (69a). (I haven’t come up with a context in which (69b) sounds perfect, but my impression is that it evokes the kind of scenario in which someone checks every so often to see whether Napoleon has died.) They predict that the two sentences should have the same status because, in both cases, one can determine the sentence’s truth by determining the truth of a domain-variant whose domain consists of a single time interval – in the case of (69a), the relevant time interval is the earliest one in the contextually salient set, and in the case of (69b) it is the last one. Accounting for the difference between (69a) and (69b) would certainly mean revising (P), perhaps in such a way as to take into account the chronological order (or some other natural ordering) of the time intervals in the domain of quantification.18 (69)
a. Napoleon was always dead. b. Napoleon was never dead.
There is another potential refinement worth mentioning: since my concern in this chapter has been exclusively with adverbial quantifiers, I have formulated (P) to account only for facts involving adverbial quantifiers, but probably (P) should be extended to cover parallel facts involving quantifier phrases in other positions. If I tell you (70) alluding to the Marx Brothers, that sounds as odd as telling you The student who finished first was always Swedish when we know that the same student finished first each time. Presumably it sounds odd for the same reason.
208
(70)
Presupposition and Implicature in Compositional Semantics
Each one’s mother was named Minnie.
Is there any connection between the principle of use that I have argued for, and anything else that we know about? As a pragmatic principle, (P) looks very different in character from certain others that have been posited, in that it compares propositions that are equally informative. Perhaps it could be viewed as a subcase of Grice’s Maxim of Manner (‘Be brief.’) (Its effect is radically different from the effect of the Maxim of Quantity, since, of the propositions it compares, it instructs the speaker to reject the logically stronger ones.) As for the facts that (P) is designed to account for, there do seem to be facts that bear a surface similarity: the bizarreness of sentences like those in (71) is reminiscent of the bizarreness of the familiar sentences in (72). (71)
(72)
a. ?? At that time, Ingrid was Swedish. b # On Tuesday, the student who finished first was Swedish. (in our second scenario) a. ?? Ingrid was always Swedish. b. # The student who finished first was always Swedish. (in our second scenario)
(P) does not account for the bizarreness of the sentences in (71). Since it is tempting to think that the problems with (72) can be reduced to the problems with (71), superficial similarities like these might tempt one to pursue an approach very different from the one I have taken here.19,20 Though I could quit here, I would like to end this chapter with a speculation. Once we recognize that a principle like (P) regulates the LFs that we use, we may be in a position to tackle old puzzles in new ways. I think in particular that a principle like (P) could play a role in explaining the phenomenon of ‘quantificational variability’ – in explaining why we often feel that a sentence like A secretary is always Swedish, which contains a quantifier that ranges over time intervals, is equivalent to a sentence like Every secretary is Swedish, which contains a quantifier that ranges over secretaries. So let me close by sketching the idea, very roughly.21 The sentence A secretary is always Swedish might well have several LFs, including the one that we have seen in Section 4 whose use generally violates (P). The idea starts from the view that, of the LFs that this sentence has, there is one whose use potentially does not violate (P). What is this LF like? Kratzer (1998) has proposed on independent grounds that what we think of as the indefinite determiner a is in actuality ambiguous between several different items, and that one of the lexical items that we
A Pragmatic Constraint on Adverbial Quantification
209
pronounce as ‘a’ is an item whose denotation is a ‘parametrized choice function.’ (A ‘parametrized choice function’ is Kratzer’s term for a function that yields choice functions as values.) The LF I am talking about is an LF that contains this item. Recall that a speaker who asserts a sentence has an LF L and assignment g in mind, and means to convey that [[L]]g characterizes the actual world. One aspect of Kratzer’s proposal is that the denotation of the item we are considering differs under different assignments – that is, what ‘parametrized choice function’ we obtain depends on what assignment we use. So a speaker who has in mind an LF that contains this item will be claiming that in the actual world some property holds of a certain ‘parametrized choice function’ Fg , where the identity of Fg is determined by the assignment g that the speaker has in mind. I don’t want to go into the details of the LF I am talking about – those familiar with Kratzer’s proposal will hopefully be able to figure it out.22 I just want to note one important aspect of it. When a speaker who uses this LF claims that in the actual world a certain property holds of a certain ‘parametrized choice function’ Fg , this will be just like saying that a related property holds of another kind of function that we can call SECg – a function that yields as its values people who are secretaries in the actual world. Specifically, a speaker who has in mind the LF we are considering will be claiming of w0 that, for all t in the contextually salient set S of time intervals, SECg (t) is Swedish-born in w0 – where SECg is a function from the set S to secretaries in w0. Now, suppose you are a hearer who hears A secretary is always Swedish. Suppose you think that the speaker had in mind an LF of this kind but you don’t know what assignment g the speaker had in mind. Since you don’t know what g is, you don‘t know the exact nature of the claim the speaker is making about w0, and in particular you don’t know the identity of SECg . But if you think that the speaker is obeying (P), you can conclude something about SECg . You can conclude that SECg yields a different secretary for each time interval in S. Why? Imagine that SECg yielded the same secretary for two time intervals t2 and t3 . Then the speaker could have communicated the same claim about w0 by quantifying over fewer time intervals – for example, by quantifying over all the time intervals in S except t3 . So the speaker would have violated (P). But now consider: if SECg yields a different secretary for each time interval in S, then SECg is a one-to-one mapping from S onto a certain set T that consists of secretaries in the actual world. And in that case, if the speaker is claiming that, for all t in S, SECg (t) is Swedish-born in w0, then he must be claiming that every secretary in T is Swedish-born
210
Presupposition and Implicature in Compositional Semantics
in w0. And this latter claim is the kind of claim that a speaker could make with a sentence of the form Every secretary is Swedish. This, I tentatively suggest, could be the basis for our feeling that the two sentences convey the same thing – though naturally some important questions remain to be addressed. End of speculation.
Notes 1. This is an old chapter, slightly trimmed for inclusion here. The original version appeared in the ZAS Working Papers, vol. 22 (2001). All errors should be blamed on the earlier self of mine who wrote the paper: I respect his opinions but do not completely share them. I am grateful for comments from Andrea Bonomi, Gennaro Chierchia, Rose-Marie Déchaine (at the Florence Cartography Workshop), Manfred Krifka (at the ZAS Workshop on Predicative Constructions), and reviewers (some of whom reviewed this chapter for a volume that never saw daylight). My approach to facts about ‘individual-level predicates’ has many antecedents – I would single out work by de Swart, and by Higginbotham and Ramchand. 2. See for example the brief discussion early on in Runner 1994 of The linguist I met last year at the LSA is always intelligent. The origins of this argument are can be found in Kratzer (1995), but Kratzer herself gives a more complicated paradigm and a more complicated argument. The examples of hers that are parallel to (1) and (2) are (i) and (ii) (which she maintains contain an implicit always). (i) When Mary speaks French, she speaks it well. (ii) When Mary speaks French, she knows it well. 3. See de Swart 1991 for arguments of a similar nature. 4. A reviewer gave the following succinct paraphrase: ‘Let Q be [the relevant kind of] generalized quantifier and A,B two sets. Then the statement ‘Q A B’ is pragmatically deviant in a common ground CG if there is a proper subset A’ of A such that CG |= Q(A)(B) ⇔ Q(A )(B).’ 5. Call C the set of worlds compatible with the information that the parties to conversation are taking for granted (cf. Stalnaker 1979). Then another way of putting (ii) is as follows: For every world w in C, [[QUANTi α]]g (w) = 1 iff Δ(w) = 1. 6. Given the more realistic assumption that adverbial quantifiers make reference to a property of time intervals, I would say something like: don’t use [ QUANTi α ] when the parties to conversation take it for granted that the contextually salient property holds of fewer than four time intervals. This is a variation of a proposal in de Hoop/de Swart (1989) and de Swart (1991) (their proposal would entail . . . fewer than two time intervals). Analogous restrictions seem to apply to determiner quantifiers like every, so this constraint probably follows from more general restrictions on the use of quantifiers. De Hoop and de Swart have explored this issue in the cited works. 7. This can be seen as a minor change from what I have assumed so far in talking about The student who finished first was always Swedish, but it is not a change that affects the preceding discussion in any serious way.
A Pragmatic Constraint on Adverbial Quantification
211
8. My use of a ‘lifetime presupposition’ here is inspired by Musan (1995). 9. One formulation that will do for the purposes here is: Do not use [ QUANTi α ] to express that [[ QUANTi α ]]g holds of w0 unless it is taken for granted about w0 that, for every x in the domain of [[ QUANTi α ]]g , [[ α ]]g i−>x (w0) is defined. 10. Or more precisely: as long as we have no reason to think that the speaker is assuming this kind of departure from normal assumptions. 11. I will assume, as I have until now, that to be Swedish really means to be Swedish-born. 12. A more realistic minimal pair might be as in (i)–(ii), which I think behave similarly. – I was advised to buy a box of Blastitol, but I need to know first whether it contains antihistamines. – For some reason this question comes up a lot. Why don’t you go ask the assistant pharmacist? . . . (i) He isn’t very knowledgeable, it’s true, but he does always know whether Blastitol contains antihistamines. (ii) He has a lousy memory, it’s true, but he does always know whether Blastitol contains antihistamines. 13. I am using proposition here to mean a function from worlds to truth values. I am using express proposition p to mean ‘convey that p characterizes the actual world.’ See Section 2. 14. Kratzer and Diesing actually concentrated on sentences with bare plurals. However, similar examples with singular indefinites (my concern here) show up elsewhere, and I think it fair to say that the same view applies. 15. I think that there is an additional aspect of (65) that contributes to bringing out the reading I am interested in. This is the contrast between secretary and technical assistant – the natural way of pronouncing (65) places pitch accent on both technical and secretary. A near minimal pair (for me) that suggests that the contrast makes a difference is (i). The continuation in (a) is bizarre, suggesting that none of its possible LFs is appropriate. By contrast, the continuation in (b) is good and conveys that most staff reviews turn up a Swedish secretary. (Replacing ‘a secretary’ in (i a) by ‘one of the secretaries’ redeems the sentence. This might be significant for the analysis of the contrast.) (i) We require them to hire Swedes for as many positions as possible. With this in mind, we review their staff every year. In fact, a. ?? a secretary is usually Swedish. b. none of the technical assistants is ever Swedish. Still, a Secretary usually is. 16. One might imagine that the LFs for sentences like Some secretary is always Swedish or At least one of the secretaries is always Swedish yield propositions just like the one (58) yields. Suppose they do. I have suggested here that assumptions that we make in the absence of contrary information prevent us from using LFs like (58). Does this imply that these other sentences should sound odd in the absence of special information? Not obviously. When we are exposed to a sentence and find that (P) taken together with our natural assumptions rules out one of its LFs, we have various options other than
212
17. 18.
19.
20.
21. 22.
Presupposition and Implicature in Compositional Semantics rejecting the sentence. If the sentence has another LF that yields a different proposition, and that does not violate (P), we could decide to take the sentence as expressing that different proposition. Or we could decide to revise our assumptions. (In the case of sentences like (56), it has been argued that another kind of LF is available, and so this first route is open. In the case of sentences like Some secretary . . . , perhaps the option is not available.) Parallel sentences with quantifiers like half the time might be of use in verifying this claim. There are many candidates to think about. Here is one. Starting from a salient (strong) linear ordering of the intervals in the domain of quantification of [[ QUANTi α ]]g , establish a corresponding ordering of those domainvariants whose domain consists exclusively of a single element of that set. (For instance, if the domain of (84a) is {March 1810, April 1810, May 1810, . . .}, order λw. in w Napoleon was dead for the duration of March 1810 before λw. in w Napoleon was dead for the duration of April 1810 before λw. in w Napoleon was dead for the duration of May 1810, etc.) New principle: Don’t use [[ QUANTi α ]]g if the truth of one of these domain-variants guarantees the truth of the next domain-variant in the order. For example, one might try to build on the idea that verifying sentences like those in (72) involves verifying many nonquantificational sentences of the kind in (71). It is natural to diagnose the oddness of (71) as resulting from a conflict with a quantity implicature. To evaluate this diagnosis, it might help to have a good description of the conditions of use of sentences with temporal adverbs. I think giving such a description is not so straightforward (Percus, 1997 contains an attempt). See Percus (1998) for the theory that inspired this rough suggestion and that perhaps could inspire refinements. Hint. In Kratzer’s LFs, the item whose denotation is a ‘parametrized choice function’ combines with a variable (i.e. an indexed item) – this results in a constituent whose denotation is a choice function. Then the result combines with a constituent whose denotation is the set the choice function selects from. In the particular LF that I have in mind here, always binds that initial variable, but does not bind any variable in the constituent that determines the selection set.
References de Hoop, H. and H. de Swart (1989) ‘Over Indefiniete Objecten en te Relatie tussen Syntaxis en Semantiek’. Glot 12, 19–35. de Swart, H. (1991) ‘Adverbs of Quantification: A Generalized Quantifier Approach’. PhD dissertation, Rijksuniversiteit Groningen. Diesing, M. (1992) Indefinites. Cambridge, MA: MIT Press. Kratzer, A. (1995) ‘Stage-Level and Individual-level Predicates’. In: G. Carlson and F.J. Pelletier (eds) The Generic Book. Chicago: University of Chicago Press. Kratzer, A. (1998) ‘Scope or Pseudoscope?: Are there Wide Scope Indefinites?’ In S. Rothstein (ed.), Events in Grammar. Dordrecht: Kluwer.
A Pragmatic Constraint on Adverbial Quantification
213
Musan, R. (1995) ‘On the Temporal Interpretation of Noun Phrases’. PhD dissertation, MIT. Percus, O. (1997) ‘Aspects of’ A. PhD dissertation, MIT. Percus, O. (1998) ‘A Somewhat More Definite Article’. In D. Strolovitch and A. Lawson (eds), Proceedings of SALT VIII. Ithaca NY: CLC Publications (Cornell). Runner, J. (1994) ‘A Specific Role for AGR’. In E. Benedicto and J. Runner (eds), UMass Occasional Papers 17: Functional Projections. University of Massachusetts, Amherst: GLSA. Stalnaker, R. (1979) ‘Assertion’. In: P. Cole (ed.), Syntax and Semantics 9: Pragmatics, 315–32. New York: Academic Press.
8 Transparency: An Incremental Theory of Presupposition Projection1 Philippe Schlenker UCLA and Institut Jean-Nicod
The analysis of presupposition projection led researchers to propose in the early 1980s that the meaning of a clause should be viewed as its Context Change Potential rather than as its truth conditions (Heim, 1983; Stalnaker, 1974). We argue that this ‘dynamic turn’ was misguided, and that it leads straight into a dilemma: either one follows Stalnaker in his pragmatic analysis, in which case one obtains a beautiful analysis of presupposition projection in conjunctions, but not of much else; or one follows Heim in her semantic analysis, which yields broader empirical coverage but little explanatory depth (no predictions are made about connectives whose Context Change Potential was not stipulated to begin with). We sketch an alternative account, entirely developed within classical logic. We argue that in some cases a complex meaning m is conceptualized as involving a precondition p, with m=pp (Division). In this case a pragmatic principle, Be Articulate!, requires that if possible m should be expressed as a conjunction p and pp rather than as pp (in order to make explicit the special status of the pre-condition p). If so, why can pp ever be pronounced on its own? Because a principle of Minimization sometimes rules out the full conjunction p and pp , leaving pp as the sole contender. Specifically, Minimization prohibits any sentence S from starting with α[p and if it can be ascertained that no matter how S will end the string p and could be deleted from S without modifying its contextual meaning. This derives a principle of Transparency, which suffices to obtain Heim’s projection results, but which has the advantage of making predictions for connectives that she does not consider. We end the chapter with some speculations on the Principle of Division, and suggest (but do not prove) that it is itself pragmatic in nature. 214
Transparency 215
1 1.1
The dynamic turn and the Transparency theory The dynamic turn and an alternative
A new conception of meaning arose in the early 1980s, due in large part to the works of R. Stalnaker and I. Heim. It used to be thought that the meaning of a clause could, as a first approximation, be analyzed as its truth conditions. Building on ideas by Stalnaker, Heim suggested instead that one should view the meaning of a clause in a dynamic way, as the effect it has on the Context Set, i.e. on the beliefs of the speech act participants (these beliefs are henceforth modeled as a the set of those possible worlds that are compatible with what the speaker and addressee take for granted). An important part of the motivation for this ‘dynamic turn’ lay in the analysis of the ‘Projection Problem’ for presuppositions.2 To illustrate the puzzle, let us consider the sentence John knows that it is raining, which presupposes that it is raining. It is natural to posit that this clause can only be uttered in a context of speech in which it is assumed that it is raining (more formally: each world of the initial Context Set should be one in which it is raining). But this simpleminded analysis makes incorrect predictions for a slightly more complex example, which presupposes nothing at all: It is raining and John knows it. Why is there such a difference between the two examples? Heim (1983) (building on analyses due to Stalnaker) suggested that a dynamic view of meaning could solve the problem. The main idea is that the presupposition of the second conjunct is not checked with respect to the initial Context Set C, but rather with respect to an ‘intermediate Context Set’ C , obtained by updating C with the dynamic meaning of the first conjunct. After this initial update, C only contains worlds in which it is raining. As a result, the presupposition of John knows it is satisfied with respect to C , though it wasn’t with respect to C. The key insight, then, is that the computation of the meaning of a sentence makes reference to a variety of local Context Sets (for short: local contexts), which are obtained as modifications of the initial one. As seductive as it may be, we will argue that this dynamic turn was misguided. We suggest that the notion of a ‘local context’ is deceptively appealing in the case of conjunction because the assertion of a conjunction can plausibly be equated with a succession of two assertions (though even this idea is dubious when the conjunction in question is embedded under other operators). But in the case of other connectives, the idea makes little sense. With respect to which context are F and G asserted in the disjunction F or Q? The point of a disjunction is precisely that it does not commit the speaker to the assertion of either disjunct. How is the
216
Presupposition and Implicature in Compositional Semantics
dynamic analysis to deal with this case? We believe that it is faced with a dilemma. The pragmatic side of the theory, championed by Stalnaker (e.g. 1974), must either develop an unintuitive theory of local contexts or forego the extension to other connectives. The semantic side of the theory, developed by Heim, has broader empirical coverage because it can simply stipulate the Context Change Potential of connectives and operators. But for this very reason, it fails to be explanatory: it does not make predictions for connectives whose lexical entry is not stipulated to begin with, and it fails to explain why there couldn’t be a deviant conjunction and*, truth-conditionally equivalent to and but with a different projective behavior (this point was repeatedly emphasized by Heim herself). We will propose an alternative account, entirely developed within (static and bivalent) classical logic. We argue that in some cases a complex meaning m is conceptualized as involving a precondition p, with m = pp (Principle of Division; the notation pp indicates that a single clause has as its meaning the conjunction of p and p , and that p is the pre-condition of m). When m undergoes Division, a pragmatic principle, Be Articulate!, requires that if possible m should be expressed as a conjunction p and pp rather than as pp , thus making explicit the separate status of the pre-condition p. If no principle were to rule out the full conjunction, Be Articulate! would prevent pp from ever being pronounced on its own. However, a principle of Minimization (Minimization of Conjunctions) prohibits any sentence S from starting with α[p and if it can be ascertained that no matter how S will end the string p and could be deleted without modifying its contextual meaning (the intuition is that in such cases p and is semantically idle, and is thus blocked by Gricean considerations: one should not utter words in vain). It is only in case Minimization is violated by the beginning of the full conjunction α[p and p ] that the string α pp is admissible on its own. This derives a principle of Transparency, which suffices to obtain most of Heim’s results, but which has the advantage of making predictions about connectives that she does not consider, such as or, unless and while; our principle also rules out a deviant conjunction and* on principled grounds. Towards the end of the chapter we speculate (but emphatically do not prove) that the Principle of Division can itself be given a pragmatic derivation. It is thus hoped that the present analysis is an intermediate step towards a fully pragmatic account of presuppositions, one whose inspiration goes back to Stalnaker (1974), Grice (1981), Abbott (2000), and Simons (2001).
Transparency 217
1.2
The Stalnaker/Heim dilemma
We take the Stalnaker/Heim analysis to have the following logic, which leads straight into a dilemma. 1. Assumption: When a clause pp with presupposition p is uttered, it requires that p be taken for granted in the context (more properly: Context Set) of utterance. 2. Observation: In some cases, the Assumption seems to be violated, e.g. in It is raining and John knows that it is, which does not presuppose anything. 3. Conclusion: The notion of ‘context’ must be ramified. In the course of the evaluation of a sentence, there are a variety of local contexts, which are obtained as modifications of the initial one. In It is raining and John knows that it is, the local context obtained after the first conjunct is evaluated is one in which the presupposition that it is raining is indeed satisfied. We will argue in what follows that the Observation refutes the Assumption. Instead of claiming that p must be taken for granted in some Context Set, though not necessarily the initial one (as the Conclusion asserts), we will suggest that p must be ’transparent’ relative to the one and only Context Set there is, which is the initial one. Those who adopt the Conclusion are led straight into a dilemma. Stalnaker represents its pragmatic horn. For him, a conjunction is a succession of two assertions; by asserting F and G, the speaker performs two successive speech acts: he utters F in a background of assumptions C, and then utters G in a background of assumptions C , obtained by updating C with F. If we further assume that a clause pp with presupposition p and assertion p is felicitous in a context C just in case C entails p (i.e. if each C-world is a p-world), we obtain the following rules, which are formalized by Heim (1983): (1)
a. Elementary Clauses C[pp ]=# unless each w in C satisfies p. If =#, C[pp ]={w∈C: p (w)=true} b. Conjunctions C[F and G]=(C[F])[G]
(C is a Context Set, i.e. a set of possible worlds; pp is an elementary clause with a presuppositional component p and an assertive component p ; and C[F] is the update of C with F. # indicates an ‘error signal’ which is
218
Presupposition and Implicature in Compositional Semantics
obtained when a presupposition failure occurs.) It is immediate that these rules predict that p and qq should presuppose nothing at all whenever p entails q – apparently the correct result. Despite this positive result, three negative remarks should be made at the outset. (i) It is by no means clear that it is a rational strategy, as Stalnaker implies, to update one’s beliefs right after one’s interlocutor has uttered a sentence or a conjunct. Suppose that my interlocutor is a complete idiot and that I am fully aware of it. A rational strategy might well be not to update my beliefs at all. But certainly this doesn’t prevent me from understanding what he has to say. Whatever notion of update is relevant for the analysis of presupposition projection, it is by no means clear that it corresponds to the rules by which a speech act participant modifies his beliefs in the course of a conversation. (ii) It is also unclear how a conjunction that appears in the scope of other operators could be identified with a succession of two assertions. What should the relevant assertions be in a negative environments such as None of my students is both incompetent and proud of it? (iii) Finally, it is difficult to see how Stalnaker’s strategy can extend to other connectives, such as disjunction. The point of asserting F or G is precisely that one is neither committed to F nor to G. But then how can an analysis based on assertions in some ‘local’ contexts be made to work in this case? In order to solve these problems, Heim (1983) adopts the approach pioneered by Stalnaker, but reinterprets it in a semantic fashion. The rule in (1b) is not seen as a rational strategy that a conversation participant should adopt in interpreting what his interlocutor is saying. Rather, the very meaning of the connective and specifies that its effect is to modify the Context Set in such a way that C[F and G]=(C[F])[G]. Formally, the meaning of a clause is now a function from Context Sets to Context Sets (=a Context Change Potential). To put it in a slogan: meanings should not be viewed as truth conditions, but as Context Change Potentials. This semantic move frees Heim (1983) from the three problems that plagued Stalnaker’s approach: she is not forced to equate update rules with rational strategies; she has no special problems with embedded environments since the notion of assertion plays no role in her analysis; and her theory can mechanically extend to other connectives and operators. But the expressive power of her analysis is also its
Transparency 219
explanatory downfall.3 Nothing rules out a deviant conjunction and* which has the same truth-conditional contribution as and but a different projective behavior. Assume for instance that by definition C[F and* G]=(C[G])[F] (this is the same rule as before, except that F and G have been permuted in the right-hand side of the identity). This would predict that John knows that it is raining and it is presupposes nothing, while It is raining and John knows that it is presupposes that it is raining. In Stalnaker’s theory, this rule makes no sense because the update process does not mirror the chronological order in which the conjuncts are asserted or heard. But in Heim’s theory, nothing blocks such an entry, as was observed by Soames (1989). For the same reason, the theory of Heim (1983) makes no predictions – or very weak predictions – about the projective behavior of connectives that she does not consider. The heart of the matter is that the descriptive success of Context Change Potentials stems from the fact that they can encode more information than mere truth conditions; but for this very reason the theory is intrinsically incapable of predicting the projective behavior of connectives and operators from their truth conditions alone. 1.3
The transparency theory
We will sketch a purely pragmatic alternative to Heim’s theory. The main selling point of our analysis is that it derives the projective behavior of connectives from information about their syntax and their (classical) truth-conditional contribution. The theory is predictive precisely because it eschews Context Change Potentials. It does incorporate a ‘dynamic’ element, though not in anything like the technical sense of ‘dynamic semantics’. Part of the theory depends on the hypothesis that certain pragmatic principles are checked incrementally, as a sentence is pronounced or heard. But unlike Stalnaker’s analysis, our theory does not need the assumption that there are ‘local’ contexts or assertions below the sentence level; nor do we need to posit anything like Context Change Potentials to obtain our results. In fact, we won’t even need the assumption that ‘Context Sets’ are updated at all. In our system, someone who takes his interlocutor to be a complete fool is at liberty to leave his initial Context Set untouched. Here is how the theory works. (i) We argue that in some cases a complex meaning m is conceptualized as involving a precondition p, with m=pp (Principle of Division). p corresponds to what is in standard theories the presupposition, while p
220
Presupposition and Implicature in Compositional Semantics
is the assertion. For us, however, p is simply a distinguished conjunct (we don’t have any other choice since the logic we adopt is entirely classical, and thus lacks the resources of a third truth value to encode ‘presupposition failure’). For most of the discussion Division can be taken to be stipulated in the lexicon, though we will speculate towards the end of the chapter that Division is a general pragmatic principle that operates both at and above the word level. (ii) When a meaning undergoes Division, a pragmatic principle, Be Articulate!, requires that whenever possible it should be expressed as a conjunction p and pp rather than as pp . The intuitive idea is that it is sometimes good conversational practice to articulate a pre-condition as a separate conjunct in order to emphasize that it plays a distinguished role. Be Articulate! might itself originate from more basic pragmatic principles, but in this chapter we will be content to take it as given, with the following form:4 (2)
Be Articulate! When (following Division) p is analyzed as a pre-condition of the complex meaning pp , if p and ( p)p can be uttered felicitously in a syntactic context α__β, α[p and ( p)p ]β is preferable to αpp β .
We have written (p)p to indicate that the principle comes in two versions. According to Version 1, pp is in competition with p and pp : John knows that it is raining competes with It is raining and John knows it. According to Version 2, pp is in competition with p and p : John knows that it is raining may thus compete with something like It is raining and John believes it (the precise nature of p doesn’t matter for our analysis, and it is only for simplicity that we take John knows p to mean p and John believes p).5 As far as we can tell, the choice between Version 1 and Version 2 does not matter, as long as it is assumed that in the full conjunction p must come first. This is certainly the only reasonable possibility in Version 1 because the opposite choice (pp and p) would make p entirely redundant once pp has been uttered. In fact, it can be checked independently that conjunctions are entirely infelicitous when the second conjunct is entailed by the first: #John has cancer and is sick is extremely odd; by contrast, John is sick and has cancer is acceptable; and the same facts hold when we consider sentences with presuppositions: John knows that it is raining and it is is deviant, whereas It is raining and John knows that it is is far more natural. Thus in Version 1 of the theory, the only possible competitor for pp is p and pp . In Version 2, by contrast, there is no obvious reason why p and p rather than p and p should compete with pp (in other words, it is unclear
Transparency 221
why John knows that it is raining should compete with It is raining and John believes it rather than with John believes that it is raining and it is (raining)). This leaves open two possibilities: either we will make slightly different predictions with Version 2 than with Version 1; or we will have to find some independent evidence that the competitor of pp is p and p , not p and p. Be that as it may, the important observation to make at this point is that if no other principle ruled out the full conjunction p and ( p)p , we would incorrectly predict that pp can never be uttered on its own. But a natural principle of Minimization does rule out some conjunctions. The basic idea is that the beginning of a sentence α[p and uttered in a background of assumptions C is infelicitous if no matter what follows this string, the expression p and could be eliminated without modifying the contextual meaning of the result: (3)
Minimization of Conjunctions If the beginning of a sentence α[p and is uttered in a background of assumptions C, the sentence is infelicitous if it can be ascertained that no matter what the second conjunct p will be and no matter what the end of the sentence β will be, the string p and could be left out without modifying the contextual meaning of the entire sentence. More precisely, infelicity results if for each possible continuation β, C |= ∀X (α[p and X]β ⇔ αXβ)
To illustrate, Minimization of Conjunctions rules out It is raining and __ in a context in which it is assumed that it is raining; and it rules out If it is raining, it is raining and __ in any context whatsoever, since as soon as we hear the word and we can be certain that the entire expression it is raining and was uttered in vain (it could be eliminated without affecting the meaning of the sentence, no matter what the end of the sentence will turn out to be). If Minimization is the only principle that rules out conjunctions, we obtain a characterization of those circumstances in which pp can be uttered. Given Be Articulate!, [p and pp ] will be preferred to pp in all cases except when p and is redundant and thus violates Minimization. This occurs just in case the background assumptions C guarantee that: ∀X(α(p and X)β ⇔ αXβ). This gives rise to the following principle: [p and pp ] will be preferred to pp in all cases except when p and is redundant, i.e. exactly in case the background assumptions C guarantee that: ∀X(α(p and X)β ⇔ αXβ). Since p and X has the same interpretation
222
Presupposition and Implicature in Compositional Semantics
as pX, we obtain a Principle of Transparency, which can be stated as follows: (4)
Transparency For any initial part αpp of a sentence uttered in a background of assumptions C, where pp is a clause that has undergone Division, it should be the case that for any sentence completion β, C|= ∀X(α(p and X)β ⇔ αXβ), or equivalently (since p and X and pX have the same meaning): C|= ∀X(α(pX)β ⇔ αXβ), where X is a propositional variable.
To illustrate, if it is not assumed that it is raining, It is raining and John knows it is preferred to John knows that it is raining (by Be Articulate!). But if it is assumed that it is raining, It is raining and John knows it violates Minimization, and thus John knows that it is raining become acceptable (for lack of a competitor). By the same reasoning, It is raining and John knows that it is and If it is raining, John knows that it is are acceptable in any context because the would-be competitors are in blatant violation of Minimization: it is clear that the second occurrence of it is raining is redundant in It is raining and it is raining and John knows that it is and in If it is raining, it is raining and John knows that it is. As the reader may have noticed, this derives simple cases of presupposition projection. Our goal in the following section will be to show that we can match Heim’s predictions in almost all the cases she considers, but that we can do so in a predictive way, thanks to an algorithm that extends to connectives that Heim does not consider (as pointed out by an Amsterdam Colloquium referee, our system encounters a difficulty with post-posed if -clauses, which we discuss below). One final remark is in order. As is well-known, all existing theories must make provisions for the contrast between, say, The King of Moldova isn’t bald (which normally presupposes that Moldova is a monarchy) and The King of Moldova isn’t bald because there is no King of Moldova! (which doesn’t presuppose anything). In the latter case, existing theories (in particular Heim, 1983) posit a costly mechanism of ‘local accommodation’, according to which The King of Moldova isn’t bald gets interpreted as It is not the case that there is a king of Moldova and that he is bald. Like its competitors, our theory needs a stipulation for this case. We will posit that under duress the process of Division may be foregone, so that pp is simply analyzed as pp , with no requirement that p satisfy Transparency. This point will matter when we discuss disjunction.
Transparency 223
2 The projection problem: basic results of the Transparency theory We start with an extremely simple syntax for the Object Language. (5)
Syntax of the Object Language F ::= p | (pp ) | not F | [F and F ] | [F or F ] | [if F]F
When we discuss the effects of Transparency, we enrich this language with propositional variables and quantifiers, as well as with material implication (⇒) and material equivalence (⇔); we also allow p to be an atom (the additional rules are in bold): (6)
Syntax of the Extended Language F ::= p | (pp ) | not F | [F and F ] | [F or F ] | [if F]F | p| X | pX |(F ⇒ F ) | (F ⇔ F ) |∀XF
Boolean connectives have their standard semantics, and pp is interpreted as a simple conjunction (the syntax is intended to indicate that the conjuncts correspond to a single lexical item). The expression if F, G is taken to be a strict indicative conditional: with background assumptions C, if F, G evaluated in any C-world is true if and only if every C-world that satisfies F also satisfies G. The distinguished role of C requires that the modal semantics be relativized both to a world and to a set of worlds. Importantly, it will be enough for our purposes to interpret the universal quantifier to be restricted to the C-worlds. We can thus define the following semantics: (7)
Semantics Let W be a domaine of possible worlds and let I be a function from atomic propositions (of the form p or p ) to subsets of W. If w∈W, C⊆W and w∈C: a. If F is atomic (be it of the form p or p ), w, C |=I F iff w∈I(F). b. If F= not G, w, C |=I F iff it is not the case that w, C |=I G. c. If F=(GG ) or F=[G and G ], w, C |=I F iff (i) w, C |=I G, and (ii) w, C |=I G . d. If F=[G or G ], w, C |=I F iff (i) w, C |=I G, or (ii) w, C |=I G . e. If F=[if G]G , w, C |=I F iff for each w∈C, if w , C |=I G, then w , C |=I G . f. If F=(G ⇒ G ), w, C |=I F iff w, C | =I G or w, C |=I G .
224
Presupposition and Implicature in Compositional Semantics
g. If F=(G ⇔ G ), w, C |=I F iff (w, C |=I G and w, C |=I G ) or (w, C | =I G and w, C | =I G ). h. If F= ∀XG, w, C |=I F iff for each extension I[π →p] of I which assigns to the new constant π the value p⊆C, w, C |=I[π→p] G[π/X] (where G[π/X] is the result of replacing each free occurrence of X with π). Remark: If the quantifier were not interpreted as restricted to C-worlds, the condition p⊆C would be replaced with p⊆W. We henceforth write C |=F if for each w∈C, it holds that: w, C |= F. It is worth noting that when C is fixed, our system reduces, quite simply, to a traditional modal logic with universe C (whatever happens outside C is irrelevant for the truth of any formula evaluated within C). In particular, the formula [if G]G can be seen as a simple notational variant of ❑(G⇒G ). This simplification is only possible because we do not consider subjunctive conditionals, which would require a more complex analysis. It follows from the reduction to modal logic that if two (variable-free) expressions are materially equivalent in each C-world, they are substitutable salva veritate in any formula evaluated with respect to C (this is because two expressions that have the same intension are substitutable salva veritate in a modal logic): (8)
Substitutivity If G and G are variable-free and if C |= G ⇔ G , then for any formula F which contains G, the result F of replacing one or several occurrences of G with G in F satisfies: C |= F ⇔ F
Substitutivity will be used repeatedly to derive the results we discuss below. 2.1
Connectives I: standard cases (not, and, if )
We start by showing step by step that our theory matches Heim’s predictions in the simple cases: expressions of the form (pp ) as well as not ( pp ) and if (pp ) all presuppose (at least) p; and sentences of the form p and (qq ) and if p, (qq ) (where p is not presuppositional) presuppose if p, q. More general results comparing our system to Heim’s in the general case are left for future research. We consider each case in turn, and we follow the convention that |= is read as |=I .
Transparency 225
Example 1. Sentences of the form (pp ) Transparency requires that for any sentence completion β, C |= ∀X ((pX)β ⇔Xβ) Claim: Transparency is satisfied iff C |= p. i. If C |=p, for any p⊆C, C |=I[π→p] ((pπ) ⇔ π), hence by Substitutivity C |=I[π→p] ((pπ)β ⇔ πβ), and by (restricted) universal quantification C |= ∀X ((pX)β ⇔Xβ). ii. Taking β to be the null string, Transparency requires that C |= ∀X ((pX)⇔X). Therefore in particular C |=I[π→C] ((pπ)⇔ π). But π is true in every C-world, and thus C |=I[π→C] p. Since π does not occur in p, we obtain: C |= p. Note: Since the result applies to all sentences of the form (pp ), it applies in particular to sentences of the form (pp ) and . . ., ( pp ) or . . ., etc. Example 2. Sentences of the form [not (pp )] Transparency requires that for any sentence completion β, C |= ∀X ([not(pX)]β ⇔ [not X]β). Claim: Transparency is satisfied iff C |= p. i. If C |=p, the result follows as in Example 1. ii. Taking β to be the null string, we have C |= ∀X (not(pX) ⇔ not X), hence C |= ∀X ((pX) ⇔ X), and thus from Example 1, ii: C |= p. Example 3. Sentences of the form [p and (qq )] Transparency requires that for any sentence completion β, C |= ∀X ([p and (qX)] β ⇔ [p and X]β) Claim: Transparency is satisfied iff C |= (p ⇒ q). i. If C |= (p ⇒ q), for any p⊆C, C |=I[π→p] ([p and (qπ)] ⇔ [p and π]), hence for any sentence completion β, C |=I[π→p] [p and (qπ)] β ⇔ [p and π]β (by Substitutivity). By (restricted) universal quantification we obtain C |= ∀X ([p and (qX)]β ⇔ [p and X]β). ii. By taking β to be the null string, Transparency entails that C |=I[π→C] ([p and π]⇒ [p and (qπ)]). But π is true in every C-world, and thus C |=I[π→C] (p ⇒ [p and q]), whence C |=I[π→C] (p ⇒ q). Since π does not occur in (p ⇒ q), we have C |= (p ⇒ q). Example 4. Sentences starting with [if (pp ) We start by observing that, given our semantics, if w is a C-world, w, C |= [if F] F if and only if C |= (F ⇒ F ). Transparency requires that for all sentence completions β, C |= ∀X ([if (pX)]β ⇔ [if X]β)
226
Presupposition and Implicature in Compositional Semantics
Claim: Transparency is satisfied iff C |= p. i. Clearly, if C |= p, for any p⊆C, C |=I[π→p] ((pπ) ⇔ π), hence for any sentence completion β, C |=I[π→p] ((pπ)β ⇔ πβ) (by Substitutivity). It follows by (restricted) universal quantification that C |= ∀X ([if (pX)]β ⇔ [if X]β). ii. Taking β to be p, Transparency entails in particular that C |=I[π→C] ([if (pπ)]p ⇔ [if π]p). The left-hand side is true in every C-world, and thus C |=I[π→C] [if π]p, whence (given the value of π), C |=I[π→C] p. But since π does not occur in p, this yields: C |= p. Example 5. Sentences of the form [if p](qq ) Transparency requires that for all sentence completions β, C |= ∀X ([if p](qX)β ⇔ [if p]Xβ) Claim: Transparency is satisfied iff C |= (p ⇒q). i. If C |= p⇒ q, for any d⊆C, C |=I[π→d] ([if p](qπ)⇔ [if p]π), hence for any sentence completion β, C |=I[π→d] ([if p](qπ)β ⇔ [if p]πβ) (by Substitutivity). Applying (restricted) universal quantification, we obtain: C |= ∀X ([if p](qX)β ⇔ [if p]Xβ).6 ii. Taking β to be empty, Transparency entails that C |=I[π→C] ([if p] (qπ) ⇔ [if p]π). The right-hand side is true in every C-world, and thus C |=I[π→C] [if p] (qπ). But π is true in every C-world, and thus C |=I[π→C] [if p] q, i.e. C |=I[π→C] (p ⇒q). Since π does not occur in (p ⇒q), C |=(p ⇒q). We will come back at the end of this chapter to the case of post-posed if -clauses, for which we make incorrect predictions. 2.2
Connectives II: Other Cases (or, unless, while)
We now consider more intricate examples, which are not discussed by Heim (though disjunction is discussed by other theories of presupposition projection). 2.2.1
Disjunction
Heim (1983) does not discuss the projective behavior of disjunction. There are a variety of positions in the literature (see Krahmer, 1998; Beaver; 2001). Following Beaver (2001), we take the correct result to be that pp or q presupposes that p, and p or qq presupposes that if not p, q. Whatever their stand is on this issue, existing theories must stipulate the projective behavior of or, which never follows from anything
Transparency 227
else. One could attempt (as Beaver does) to claim that F or G must be defined as if not F, G. But of course the fact that that this definition is correct when one considers truth conditions alone (as in propositional logic) does not entail that it must also hold within a semantics based on Context Change Potentials, since the latter is strictly more expressive than the former. Thus it could well be that F or G and if not F, G have the same truth-conditions, but different Context Change Potentials. In fact, this is precisely the situation we described at the beginning of this chapter when we compared F and G to F and* G. Without the assumption of inter-definability, Beaver (2001) must do as everyone else, and stipulate the behavior of disjunctions. Contrary to existing proposals, our algorithm makes precise predictions: as already shown in Example 1, pp or q presupposes p (since the sentence starts with pp ); and we will now show that p or qq presupposes if not p, q. Example 6. Sentences of the form [p or (qq )] Transparency requires that for any sentence completion β, C |= ∀X ([p or (qX)]β ⇔ [p or X]β). Claim: Transparency is satisfied iff C |= ((not p)⇒ q). i. If C |= ((not p)⇒ q), for any d⊆C, C |=I[π→d] ([p or (qπ)] ⇔ [p or π]) (because [p or π] is equivalent to [p or [(not p) and π]], which entails [p or (qπ)] if (not p) entails q). Hence for any sentence completion β, C |=I[π→d] ([p or (qπ)]β ⇔ [p or π]β) (by Substitutivity). By (restricted) universal quantification, we obtain that C |= ∀X ([p and (qX)]β ⇔ [p and X]β). ii. By taking β to be the null string, Transparency entails that C |=I[π→C] ([p or (qπ)] ⇔ [p or π]). The right-hand side is true in every C-world, as is π, and thus C|=I[π→C] [p or q]. But since π does not occur in [p or q], we get C|=[p or q], i.e. C |= ((not p) ⇒ q). Are these predictions correct? The following example is correctly predicted to presuppose nothing: (9)
This house has no bathroom or the bathroom is in a funny place (after Partee)
So far, so good. But it would seem that we make incorrect predictions when the disjuncts are reversed: (10)
The bathroom is in a funny place or this house has no bathroom
228
Presupposition and Implicature in Compositional Semantics
Arguably (10) (which is also a standard example) can be understood as presupposing nothing, which does not square well with the claim that ( pp or q) presupposes that p. Although we will not give a full solution to this problem, we make two preliminary remarks. 1. As pointed out by B. Spector (p.c.), there are other cases that suggest that there is a weak but systematic asymmetry in the projective behavior of disjunctions, as illustrated by the contrast between (11a-b): (11)
a. John doesn’t have a violin or his instrument is well hidden. b. John’s instrument is well hidden or he doesn’t have a violin. (B. Spector).
(11a) doesn’t presuppose anything, as is predicted by the analysis of Example 6. If or were projectively symmetric, (11b) should presuppose nothing either. But as B. Spector points out, this is empirically dubious (on the other hand the contrast appears to disappear when have is stressed and violin is destressed in the second clause, as was pointed out by D. Büring; I have no account of this fact). Why should Spector’s contrast hold? Remember that our theory allows for Division not to be applied, though this should be a costly option, to which an interpreter can resort only in case the normal interpretive process leads to an absurdity (as was mentioned, this is our version of what other theories call ‘local accommodation’). Let us now consider what will happen in (10) if Division is applied in the first conjunct. After Transparency is checked, it will have to be assumed that the house has a bathroom. But then the second disjunct will be trivially false, and thus eliminable. General principles of minimization should then require that it not be uttered, which leads to a pragmatic absurdity. At this point one has no choice but to forego Division, which yields the impression that the sentence does not presuppose anything. But let us now apply the same reasoning to (11b). Suppose that we do in fact generate the presupposition that John has an instrument. There is still a possibility that he does or doesn’t have a violin, and thus the second disjunct is not idle as it was in the first case. No pragmatic absurdity arises, and thus there is no reason to forego Division. After Transparency is checked in the first disjunct (following the logic of Example 1), we obtain a presupposition that John has a musical instrument. 2. It should be added that if -in spite of Spector’s observation- one were to argue from (10) that or is projectively symmetric, one would also be forced to revisit the projective behavior of other connectives. For
Transparency 229
instance it is normally thought that if pp , q presupposes p but that if p, qq only presupposes if p, q. However (10) can be modified to suggest that in some cases if pp , q only presupposes that if not q, p: (12)
a. If this house has a bathroom, the bathroom is in a weird place. b. If the bathroom is not in a weird place, this house has no bathroom.
Of course when the (costly) option of local accommodation is taken into account, these examples appear less surprising. If the usual mechanism of presupposition generation and projection were applied in the antecedent of (12b), we would obtain a global presupposition that there is a bathroom, and the indicative conditional would become pragmatically odd: it would require that there be a possibility that the house has no bathroom even though it is presupposed that it has one. Local accommodation (or in our terms failure of Division) would seem to be the only way to save the sentence. The crucial question is whether the situation is really different when one considers examples in which local accommodation is not forced in this way: (13)
a. If John has a violin, his instrument is hidden in a weird place. b. If John’s instrument is not hidden in a weird place, he has no violin.
The present line of analysis would predict that (13a) presupposes nothing, and that (13b) presupposes that John has an instrument (since no pragmatic absurdity arises in this case: all the indicative conditional requires is that there be a possibility that John has an instrument, that the instrument in question is not hidden, and that John has no violin). I must say, however, that I find the contrast between a. and b. less clear in (13) than in (11). I leave a more thorough discussion of this issue for future research. 2.2.2
Other connectives: unless and while
Heim (1983) makes no predictions about other connectives that she does not consider, such as unless or while. But the present analysis is more constrained. From the equivalence between Unless F, G and Unless F, [not F] and G, we predict that any presupposition of G entailed by not F should automatically be transparent. This prediction is borne out in (14): (14)
Unless John didn’t come, Mary will know that he is here
230
Presupposition and Implicature in Compositional Semantics
Intuitively (14) presupposes nothing additional if it is assumed that if John came, he is here. This follows because the example has the form unless F, qq , where not F entails q (specifically: not (John didn’t come) entails John is here). This accounts for the data. Turning now to while, the equivalence between While F, G and While F, F and G explains the facts in (15): (15)
While John was working for the KGB, Mary knew that he wasn’t entirely truthful about his professional situation.
This sentence has the form While F, qq , where F contextually entails q (because we take for granted that a spy isn’t truthful about his profession). For this reason, Transparency is automatically satisfied. Similar examples can be constructed ad libitum. In (16) we consider the contrast between Before F, G and After F, G [we are only concerned with the projective behavior of the main clause; the pre-posed clauses trigger presuppositions of their own, a fact we disregard here]: (16)
a. Before John became a politician, his mother was glad that he had a normal job. b. After John became a politician, his mother was worried that he didn’t have a normal job.
If in a context C the speech act participants assume that John is a politician is true if and only if John does not have a normal job, we predict that in C both sentences should be acceptable, thanks to the following intuitive equivalences: (17)
a. Before John became a politician, G ⇔ Before John became a politician, John wasn’t a politician and G b. After John became a politician, G ⇔ After John became a politician, John was a politician and G
In each case, the right-hand side of the equivalence guarantees that in C the presupposition of the main clause will be transparent, which accounts for the acceptability of both examples. 2.3
Extension: quantifiers (simple cases)
Presupposition projection in quantified structures is a notoriously hairy topic, which we only treat superficially by considering [ every P]( QQ ), [at least one P]( QQ ) and [no P]( QQ ). Without defining a full formal system,
Transparency 231
we extend out notation from propositional to predicate constants and variables: the underlined part is the presuppositional one, and concatenation is interpreted as generalized conjunction. In all three cases Heim (1983) predicts the same presupposition, namely that every P-individual satisfies Q. For better or worse, we match her results. We assume in the following that the domain of individuals, D, is fixed across possible worlds. Example 7. Sentences of the form [Every P] (QQ ) Transparency requires that for any sentence completion β, C |= ∀Y ([Every P](QY)β ⇔ [Every P]Yβ), where Y is a predicate variable (note that the syntax requires that [Every P](QY) and [Every P]Y be constituents in this configuration). Claim: Transparency is satisfied iff C |= [Every P] Q i. If C |= [Every P] Q, for any d⊆D, C |=I[∂→D] ([Every P](Q∂)⇔ [Every P]∂). Thus for any sentence completion β, C |=I[∂→D] ([Every P](Q∂)β ⇔ [Every P]∂β) (by Substitutivity). By universal quantification, we obtain C |= ∀Y ([Every P](QY)β ⇔ [Every P]Yβ). ii. If some P-individual, say i, is not a Q-individual in a C-world w, Transparency fails. Let us take β to be the null string. Then w, C |=I[∂→D] [Every P]∂, but w, C | =I[∂→D] [Every P](Q∂) (because i satisfies P but not Q in w). Example 8. Sentences of the form [At least one P] (QQ ) Transparency requires that for any sentence completion β, C |= ∀Y([At least one P] ( QY) β ⇔ [At least one P]Yβ) Claim: Transparency is satisfied iff C |= [Every P]Q i. If C |= [Every P] Q, for any d⊆D, C |=I[∂→D] ([At least one P](Q∂)⇔ [At least one P]∂). Thus for any sentence completion β, C |=I[∂→D] ([At least one P](Q∂)β ⇔ [At least one P]∂β) (by Substitutivity). By universal quantification, we obtain C |= ∀Y ([At least one P](QY)β ⇔ [At least one P]Yβ). ii. If some P-individual, say i, is not a Q-individual in a C-world w, Transparency fails. Let us take β to be the null string. Then: C |=I[∂→{i}] [At least one P]∂, but C | =I[∂→{i}] [At least one P](Q∂) (because i, the only individual that satisfies ∂ in w, satisfies P but not Q). Although the predictions might be correct for at least one, they are probably too strong for indefinites, e.g. A fat man was pushing his bicycle, which doesn’t presuppose that every fat man in the domain of discourse owns a bicycle. We come back to this problem at the end of this chapter.
232
Presupposition and Implicature in Compositional Semantics
Example 9. Sentences of the form [No P] (QQ ) Transparency requires that for any sentence completion β, C |= ∀Y([No P] (QY)β ⇔ [No P] Yβ) Claim: Transparency is satisfied iff C |= [Every P]Q i. If C |= [Every P] Q, Transparency is satisfied (the argument is the same as in Example 8 and Example 9). ii. If some P-individual, say i, is not a Q-individual in a C-world w, Transparency fails. Let us take β to be the null string. Then: w, C |=I[∂→{i}] [No P](Q∂) (because i does not satisfy Q in w, which entails that no individual can satisfy both ∂ and Q in w). But w, C | =I[∂→{i}] [No P]∂ since i satisfies both P and ∂ in w.
3
The triggering problem: against a lexical treatment
The theory developed in the preceding sections is logically compatible with a lexicalist account of the Triggering Problem. Although our framework departs from standard analyses of the Projection Problem, it wouldn’t be incoherent to state that Division is lexically encoded, and that a presuppositional expression e appears in the lexicon as pp , which specifies that e should compete with [ p and ( p)p ] (and thus that p must obey Transparency). This hypothesis raises an unsettling question, however: why should such lexical entries exist in the first place? We will circumvent this problem by suggesting (speculatively) that Division is not lexically encoded, but results from a pragmatic desire to utter as a separate conjunct a part p of a complex meaning m, with m=pp . This idea was already explored by Stalnaker (1974), Grice (1981) and Abbott (2000), but a proper definition of ‘complex’ and of ‘pre-condition’ (which are crucial to turn the intuition into a theory) is left for future research. In this short section, we concentrate on two arguments against a lexical treatment. The first one is that presupposition-like phenomena can be triggered at the compositional level, with adverbial modification (cf. Simons, 2001). It is plausible that a pragmatic account is needed for these cases; but if so, conceptual economy suggests that the same account might apply at the lexical level as well. The second argument stems from pragmatic constraints on the triggering process. We argue that presuppositions fail to be triggered when the putative assertion is ‘too dependent’ on the presupposition, so that the latter cannot be seen as an autonomous condition for the former (because the full meaning m and its putative pre-condition p are too entwined, so to speak). If this sketch could be made into a theory, the same pragmatic principle would
Transparency 233
account for the Triggering Problem and for the Projection Problem – a desirable (if promissory) result. 3.1
Presupposition-like effects with adverbials
Consider the following data, both in English and in French (the French version allows for an additional control, discussed below; we use => for: intuitively implies): (18)
a. Jean n’est pas arrivé en retard. John didn’t arrive late => Jean arrived. a . John didn’t come late. => John came b. Aucun de ces étudiants n’est arrivé en retard. None of these students arrived late => Some of these students arrived (strong implication) => Each of these students arrived (possibly, weaker implication) b . None of these students came late. => Some of these students came (strong implication) => Each of these students came (possibly, weaker implication)
To my ear, (18a) carries a strong implication that Jean came; (18b) carries a strong implication that some student came; but it also carries a somewhat weaker implication that each student came. The facts in (18a) are unsurprising: it is likely that <Jean came, Jean came late> forms a Gricean scale in this context (because one could certainly have decided not to utter late).7 John didn’t come is more informative than John didn’t come late; as a result, the less informative member of the pair implicates that the more informative member is false, and thus that John did come. Applied to (18b), the same reasoning predicts that None of these students came late should implicate that it is false that none of these students came, and hence that some of these students came. But this doesn’t suffice to derive the second (possibly weaker) implication that each of these students came. The latter behaves very much as a presupposition rather than as a scalar implicature: None of these students knows that he’ll be unemployed doesn’t just imply that some of these students will be unemployed, but that all of them will be.8 We conclude, following Simons 2001, that adverbial modification can trigger presupposition-like phenomena. Why should that be? One could argue that adverbials are quite generally presupposition triggers, though it is a bit unclear how this theory
234
Presupposition and Implicature in Compositional Semantics
could be implemented and motivated; it would also fail to account for the observation that the presuppositional effect we get is a bit weaker than with ‘standard’ presuppositions. An alternative is that a presuppositionlike phenomenon is generated by Be Articulate!: x came late can be viewed semantically as a conjunction of two conditions: (i) x came, and (ii) x did so late. If the former can be seen as a pre-condition of the latter, Be Articulate! would require that one should utter (i) and (ii) as separate conjuncts, for instance as x arrived, and was late. Only in case the full conjunction is disallowed by the minimization principle can x came late be uttered felicitously. But this means that x came should behave as a presupposition of x came late. There is a catch, however. Unlike the cases we discussed earlier, the expression x came late is somewhat articulated, since it involves two expressions (came + late) rather than just one, as in the case of know. This should lead us to expect that the presuppositional effect should be slightly weaker than in standard presuppositions, as appears to be the case. This line of analysis suggests that the presuppositional effect should be strengthened if the same meaning is conveyed with a single word, and that it should be weakened if it is expressed with a more explicit form of modification. I believe that these effects are indeed found (I provide both French and English examples for the first part; I have only tested the second part in French): (19)
Strengthening the presuppositional effect a. Aucun de ces étudiants n’a été en retard. None of these students was late => Each of these students came. b. None of these students was late. => Each of these students came.
(20)
3.2
Weakening the presuppositional effect (French) Aucun de ces étudiants n’est arrivé en étant en retard. None of these students arrived while being late => Some of these students arrived. => Each of these students arrived The role of independence
Why should it sometimes be better conversational practice to articulate a condition as a separate conjunct? We do not have anything like a full answer to offer. But there appear to be general conditions under which
Transparency 235
a presupposition is not generated. In particular, when a putative condition cannot be taken as separate from the intended meaning because it pragmatically entails it, no presupposition is triggered. By ‘pragmatically entail’, we mean that in typical situations, if the condition holds, then the entire proposition is true as well, though this need not be an entailment in the logical sense (not even a contextual entailment, i.e. an entailment relative to the assumptions of the conversation). Let us now briefly illustrate this observation. (21)
Talking about a complicated mathematical statement, I say: If my brother learns that this statement is true, he’ll act accordingly. a. Situation 1: my brother is a 10-year-old. b. Situation 2: my brother is a noted mathematician.
The sentence in (21) gives rise to a pragmatic ambiguity: it may be understood with the presupposition that the statement is true, or without it (in the latter case the sentence means something like: If this statement is true and my brother comes to think so, he will act accordingly). However it seems to us that the non-presuppositional reading is much easier to obtain in Situation 2 than in Situation 1. Why should that be? Intuitively the reason is that in Situation 2 one might well expect that if the statement is true my brother will come to think so precisely because he is an excellent mathematician. In other words, this statement is true pragmatically entails my brother will learn that this statement is true. As a result, the former cannot be considered as a pre-condition of the latter, and no presupposition is generated. In Situation 1, by contrast, there is no expectation that if the statement is true my 10-year-old brother will learn about it. A similar contrast can be seen with discover in the following example: (22)
George hasn’t announced that dinner is ready. a. Situation 1: George is the family butler, who is entirely reliable. The sentence is uttered by the mother to her ten-year old son, who is rushing to the dinner table. b. Situation 2: George is a newly-hired employee at the information desk of a hotel. The sentence is uttered by his boss, who is thinking of firing him.
Here too we find a pragmatic ambiguity: the sentence may or may not presuppose that dinner is ready. In the latter case, it has roughly the
236
Presupposition and Implicature in Compositional Semantics
meaning It is not the case that dinner is ready and that George has said so. Importantly, the non-presuppositional reading is much more readily available in Situation 1 than it is in Situation 2. Why? Again, we wish to suggest that this is because in Situation 1 one expects that as soon as dinner is ready George should say that this is so; while in Situation 2 there is no such presumption because George is a newly-hired employee, who has not built a reputation of reliability. As a result, George announced that dinner is ready can be taken to have as a pre-condition that dinner is ready in Situation 2 but not in Situation 1 .
4
Problems and prospects
We conclude with a list of open problems, with possible directions for future research. The first four problems are ones we inherit from Heim (1983), precisely because we follow her in almost all her predictions (see van der Sandt, 1993 and Geurts, 1999 for a critique of Heim’s theory). The last two problems are artifacts of the present implementation, and will have to be considered in detail in future research. 1. Some of the predictions made for quantified statements by Heim (1983) and the present appear to be too strong. Consider indefinites: One of my students knows that he will be unemployed doesn’t presuppose that each of my students will be unemployed. The universal presupposition might well be easier to obtain with modified numerals than with standard indefinites: At least one of my students knows that he will be unemployed is more readily understood with a very strong presupposition. It is thus conceivable that our predictions are incorrect for standard indefinites due to an independent difference between those and modified numerals. A natural suggestion is that the unexpected projective behavior of indefinites is related to their -equally unexpected- ability to ‘scope out’ of syntactic islands, as in If some relative of mine dies, I’ll inherit a house. One way to solve the latter problem (due in particular to Schwarzschild, 2002) is to assume that for some reason indefinites differ from modified numerals in being able to have an implicit domain restriction with, combined with the overt restrictor, holds true of one individual only. The wide scope effect can in this way be analyzed as a semantic illusion; the sentence is treated as If some relative of mine satisfying D dies, I’ll inherit a house, where D is a domain restriction whose extension is small enough that only one individual is a relative satisfying D (in more sophisticated versions of the theory it
Transparency 237
could be postulated that the domain variable D can be existentially quantified, which introduces further complexities in the analysis). Importantly, this line of analysis can also account for the surprisingly weak projective behavior of indefinites. Consider the sentence Some relative of mine knows that he’ll inherit a house, analyzed as Some relative of mine satisfying D knows that he’ll inherit a house. What our theory – or Heim’s – predicts is that it is presupposed that each of my relatives who satisfy D will inherit a house. But if D holds of a small number of individuals, we will get the impression that the presupposition is not universal – just as is desired. 2. We have been entirely silent about the projective behavior of restrictors. The facts are not particularly clear, but there seem to be minimal differences that are hard to capture for a purely pragmatic approach. To my ear, there is a fairly sharp contrast between No student who knows that he is incompetent applied vs. No student who applied knows that he is incompetent: the first sentence presupposes nothing or something very weak, while the second sentence presupposes that each of the students who applied is incompetent. We could try to account for the difference by positing that implicit domain restrictions on quantifiers are fixed incrementally so as to satisfy Transparency whenever possible. This would have the effect that the first sentence gets analyzed as No student who satisfies D and who knows that he is incompetent applied, where D is chosen to hold of a subset of the students who are incompetent, in such a way that Transparency is automatically satisfied. By contrast, in the second sentence the implicit domain restriction has already been fixed when the nuclear scope knows that he is incompetent is processed, with the result that a stronger presupposition is obtained. Obviously more work is needed to determine what the full range of facts is, and whether the analysis we sketched can be made to work. 3. Heim’s theory has been criticized by van der Sandt (1993) and Geurts (1999) on the ground that it predicts presuppositions that are often too weak: (23)
a. If the problem was easy/difficult, it isn’t John who solved it. (Geurts, 1999) b. Peter knows that if the problem was easy/difficult, someone solved it. (Geurts, 1999)
Heim’s theory, followed by ours, predicts that (23a) should presuppose that if the problem was easy/difficult, someone solved it, and that it should be presuppositionally indistinguishable from (23b). But the facts are
238
Presupposition and Implicature in Compositional Semantics
very different in the two cases: the predicted presupposition appears to be correct for (23b), but not for (23a), which out of the blue has the presupposition that someone solved the problem. However the data change when (23a) is preceded by a sentence that asserts the proposition which, according to Heim and to us, is in fact presupposed: (24)
If the problem was easy/difficult, someone solved it. But if so, it isn’t John who solved it.
Van der Sandt (1993) and Geurts (1999) take these facts to argue for their ’anaphoric’ theory of presuppositions. Within our framework an additional mechanism is needed to derive the stronger-than-expected presupposition of (23a). In a nutshell, we believe that the problem might be solved by allowing Transparency to be checked in a ‘lazy’ fashion, according to which one does not consider the entire syntactic environment in which the presuppositional element appears to check for Transparency; rather, a lazy interpreter will adapt the set of assumptions C so as to ensure that the contextual meaning of a constituent satisfies Transparency no matter which syntactic environment it is embedded in (as B. Spector noted (p.c.), a similar measure could be adopted just as well to save Heim’s theory from van der Sandt’s and Geurts’s objections). To be concrete, one may elect to satisfy Transparency lazily by ensuring that in the background C the contextual meaning of it isn’t John who solved the problem, analyzed as pp [with p=someone solved the problem and p = John solved the problem], is equivalent to p . This will turn out to require that C |= ∀X (pX⇔X), and hence that C |= p, as is desired. But of course the repercussions of Laziness have yet to be explored. 4. We also inherit systematic problems with respect to the projective behavior of attitude verbs.9 John believes that he is sick and he wants Mary to know that he is doesn’t presuppose that John wants to be sick, contrary to what a straightforward application of our theory might lead us to expect (if John wants F is analyzed as meaning something like: F holds in each of John’s bouletic alternatives; as in other quantified examples, our analysis predicts a presupposition that each of John’s bouletic alternatives satisfies the presupposition of F). This problem has been discussed in some detail by Heim (1992), but it is quite general. The line adopted in Heim 1992 is to analyze want in terms of belief. Roughly, John wants F is taken to mean something like John believes that if F, things are better for him than if not F. This analysis could be adapted to the present theory, but it would need to receive some independent motivation.
Transparency 239
5. It was been pointed out by an Amsterdam Colloquium reviewer that post-posed if-clauses might well have the same projective behavior as pre-posed ones: Ok If there is a reviewer, the reviewer is mad, Ok The reviewer is mad, if there is a reviewer. But as was noted in earlier examples, more complex sentences must be considered to determine what the precise nature of the problem is. The foregoing sentences might well be treated in terms of local accommodation, since it would make no pragmatic sense to say if there is a reviewer unless one assumes that there might be no reviewer. Let us thus go back to Spector’s example: If John has a violin, his instrument is well hidden. Do the facts change when the if -clause is post-posed, as in: John’s instrument is well hidden, if he has a violin? If the facts do not change, our theory has a serious problem, and we will have to change some of our assumptions; if the modified sentence does presuppose that John has an instrument, we can leave the theory as it is. We would tend to go for the first solution but must confess that we are unsure of the data.10 6. At this point we do not have a worked out account of the presuppositions triggered by definite descriptions. But here is a direction we would like to explore: a. A definite description the F is of the individual type (type e), and denotes the non-existent object # (which encodes referential failure) unless there is exactly one F-individual. b. # does not lie in the extension of any atomic predicate (not even identity). Since the logic is bivalent, [the F]P is trivially false unless there is exactly one F-individual. In this way we derive Russellian truth conditions (even though [the F] is not quantificational). c. For reasons to be investigated (maybe following the discussion in Grice, 1981), Division occurs in [the F]P, which gets analyzed as qq , with q=there is exactly one F-individual, and q =some (or all) F-individuals are P. When q is required to satisfy Transparency, we obtain the expected presuppositional behavior of definite descriptions. d. We assume that two pragmatic violations are obtained when the expected presupposition is not satisfied: one is simply the violation of Transparency; and an additional deviance might be obtained because the non-existent object # was allowed to play a role in the truth conditions. This might yield stronger violations than with other presupposition triggers. 7. Our analysis provides a natural account of global accommodation (a rational interpreter will adapt the Context Set to ensure that Transparency is satisfied when it should be). We can easily re-interpret
240
Presupposition and Implicature in Compositional Semantics
local accommodation as a failure to apply Transparency, i.e. as a failure to generate a presupposition. However we have no mechanism to account for ‘intermediate accommodation’. We leave this question for future research. 8. Last, but not least, we do not have an account of presupposition triggers such as too, which have been claimed to have an anaphoric component ( John had dinner in New York too is infelicitous out of the blue even if one knows that someone other than John had dinner in New York; somehow too wants to be coindexed with a proposition that entails that someone other than John had dinner in New York). We also leave this problem for future research.
Notes 1. An extended abstract (6 pages) appeared in the Proceedings of the Amsterdam Colloquium 2005, though part of the present theory is somewhat different. [In the extended abstract, Transparency was essentially stipulated; here we derive it from more basic principles, Be Articulate! and Minimization. In addition, the statement of Transparency was different in the earlier chapter, and made better predictions with respect to post-posed if -clauses]. I wish to thank the following for helpful criticisms and suggestions: D. Carter, A. Cormack, P. Egré, D. Fox, B. Geurts, F. Recanati, U. Sauerland, B. Spector, D. Sperber, D. Wilson, an anonymous Amsterdam Colloquium reviewer, an anonymous reviewer for Palgrave, as well as audiences at ZAS (Berlin, Semantiknetzwerk, October 2005), University College London (November 2005), Institut Jean-Nicod (December 2005), and the Amsterdam Colloquium (December 2005). The author gratefully acknowledges the financial support of the American Council of Learned Societies (Ryskamp Fellowship) and of UCLA. 2. Another part of the motivation had to do with the analysis of donkey anaphora, a problem that we do not consider in this chapter (but see Heim, 1982; Kamp, 1981; Groenendijk & Stokhof, 1991). Suffice it to say that the arguments in favor of a dynamic analysis have been hotly debated, and that almost all problems can be solved if pronouns are treated as concealed definite descriptions rather than as variables. Strikingly, this is the position that Heim herself adopted – or at least explored – in Heim (1990). If the E-type analysis is correct, any account that yields an adequate treatment of definite descriptions will also explain the anaphoric data that originally served to motivate dynamic semantics. We do not discuss anaphora per se in the present chapter, and our account of definite descriptions is entirely promissory. 3. This observation might well be compatible with the position of Heim’s later writings. Thus Heim (1992) mentions in a footnote ‘the rash claim in Heim (1983) that CCPs [=Context Change Potentials, PS] are fully predictable from truth-conditional properties, for which I was rightly taken to task by Soames
Transparency 241
4. 5.
6.
7.
8.
(1989) and Mats Rooth (personal communication in a letter dating from 1986)’. Thanks to D. Wilson for a critical remark that led to a revision of the homologous principle given in Schlenker 2005. Philosophers have repeatedly argued that John knows that p does not simply mean p and John believes that p. We agree, but as is mentioned in the text this does not matter: all standard proposals agree that John knows that p should be analyzed as p and ____. The reader may simply take ‘John believes that p to be short-hand for whatever his favorite analysis of ____ turns out to be. The result follows because our very simple syntax guarantees that any sentence starting with [if F]G has [if F]G as a syntactic unit (otherwise G should be preceded by:[ ). As B. Spector (2003) points out, it would made no sense to assume that the same scale is at work when one utters Jean came: from the latter utterance, one certainly does not infer that Jean didn’t come late! An alternative possibility, which should be explored in future research, is that the standard account of scalar implicatures is incorrect. This has indeed been argued on independent grounds by Chierchia (2004), Sauerland (2004), Spector (2003) and van Rooij & Schulz (forthcoming). It is thus particularly important to have theory-neutral tests to determine whether the inference triggered by adverbial modification can or cannot be handled as a scalar implicature. One could try to compare the effect obtained with the implicature triggered by the scale in the scope of the same quantifier none of these students: (i) None of these students read (both) Chomsky and Montague.
It is clear that (i) suggests that some of these students read Chomsky and Montague, as the standard neo-Gricean account suggests (because it is implicated that the more informative alternative None of these students read Chomsky or Montague is false, hence the result). The question is whether something stronger is also implied, namely that each of these students read Chomsky or Montague. We leave this question for future research. 9. Thanks to B. Spector for pointing this out. 10. In Schlenker (2005), it was assumed that the data do not change when the if clause is post-posed. But because the theory was built somewhat differently, the predictions it made depended on the point of attachment of the if -clause. It was argued that it is in fact attached low enough to allow the theory to make the correct predictions. This solution is not available in the present version of the analysis.
Bibliography Abbott, B. (2000) ‘Presuppositions as Nonassertions’, Journal of Pragmatics 32: 1419–37. Abusch, D. (2002) ‘Lexical Alternatives as a Source of Pragmatic Presuppositions’. In B. Jackson (ed.), Proceedings of SALT XII, CLC Publications, Ithaca NY. Beaver, D. (2001) Presupposition and Assertion in Dynamic Semantics. CSLI, Stanford.
242
Presupposition and Implicature in Compositional Semantics
Bhatt, R. & Pancheva, R. (2001) ‘The Syntax of Conditionals’. Ms, USC and University of Massachusetts, Amherst. Chierchia, G. (2004) ‘Scalar Implicatures, Polarity Phenomena, and the Syntax/ Pragmatics Interface’. In A. Belletti (ed.), Structures and Beyond. Oxford University Press. Geurts, B. (1999) Presupposition and Pronouns. Elsevier. Grice, H. P. (1981) ‘Presupposition and Conversational Implicature’. In P. Cole (ed.), Radical Pragmatics, 183–198. New York: Academic Press. Groenendijk, J. & Stokhof, M. (1991) Dynamic Predicate Logic: Linguistics and Philosophy 14, 1: 39–100. Heim, I. (1982) The Semantics of Definite and Indefinite Noun Phrases. Doctoral dissertation, University of Massachusetts, Amherst. Heim, I. (1983) ‘On the Projection Problem for Presuppositions’. Reprinted in S. Davis (ed.), Pragmatics: A Reader, Oxford University Press, 1991. Heim, I. (1990) ‘E-Type pronouns and donkey anaphora’. Linguistics & Philosophy 13: 137–77. Heim, I. (1992) ‘Presupposition Projection and the Semantics of Attitude Verbs’, Journal of Semantics 9(3): 183–221. Kamp, H. (1981) ‘A Theory of Truth and Semantic Representation’. In J. Groenendijk, T. Janssen, and M. Stokhof (eds), Formal Methods in the Study of Language, Mathematical Center Amsterdam: 277–322, 1981. Krahmer, E. (1998) Presupposition and Anaphora, CSLI. Sauerland, U. (2004) ‘Scalar Implicatures in Complex Sentences’, Linguistics and Philosophy, 27, 367–91. Schlenker, P. (2005) ‘Transparency: an Incremental Theory of Presupposition Projection’, Extended Abstract, Proceedings of the Amsterdam Colloquium 2005. Schwarzschild, R. (2002) ‘Singleton Indefinites’. Journal of Semantics 19, 3: 289–314. Simons, M. (2001) ‘On the Conversational Basis of some Presuppositions’. In Hasting, R., Jackson, B. and Zvolenzky, S. (eds), Proceedings of SALT 11, CLC publications, Cornell University. Soames, S. (1989) ‘Presupposition’. In D. Gabbay and F. Guenthner (eds), Handbook of Philosophical Logic IV, 553–616. Spector, B. (2003) ‘Scalar implicatures: exhaustivity & gricean reasoning’, Proceedings of the 2003 ESSLLI Student Session. Sperber, D. and Wilson, D. (1989) La Pertinence (Communication et cognition), Minuit. Stalnaker, R. (1974), ‘Pragmatic Presuppositions’. In Munitz, M. and Unger, P. (eds), Semantics and Philosophy. New York: New York University Press. van der Sandt, R. (1993), ‘Presupposition Projection as Anaphora Resolution’. Journal of Semantics 9(4): 333–77. van Rooij, R. and Schulz, K. (forthcoming), ‘Pragmatic Meaning and Nonmonotonic Reasoning: The Case of Exhaustive Interpretation’. Linguistics and Philosophy.
9 Aspects of the Pragmatics of Plural Morphology: On Higher-Order Implicatures∗ Benjamin Spector Ecole Normale Supérieure Harvard University
This chapter argues that the semantic behavior of bare plurals (and some other plural indefinites) provides evidence for the existence of higher-order implicatures, i.e. pragmatic inferences based on the comparison of a given sentence with the pragmatic meanings of its alternatives (and not only with their literal meanings). In a nutshell, I claim that a bare plural noun like horses denotes the set of individuals, be they atomic or complex, whose atomic parts are individual horses, and that the at-least-two-reading that one gets in sentences such as John has seen horses in the garden is an implicature that derives from the comparison of this sentence with John has seen a horse in the garden. In terms of its literal meaning, the latter sentence is true if and only if John has seen at least one horse in the garden, but gets pragmatically strengthened into John has seen exactly one horse in the garden (scalar implicature). As a result, John has seen horses in the garden will implicate the negation of John has seen exactly one horse in the garden, so that it will end up meaning John has seen several horses. I give an explicit formalization of the processes that lead to higher-order implicatures, based on the possibility of iterating an exhaustivity operator. The proposal accounts for the readings of plural indefinites in various linguistic contexts (monotone-decreasing contexts, non-monotonic contexts, and universally quantified contexts).
1
A puzzle about plural indefinites
Consider the following sentences: (1) (2)
a. b. a. b.
The homework contains difficult problems The homework doesn’t contain difficult problems One of my students has solved difficult problems None of my students has solved difficult problems 243
244
Presupposition and Implicature in Compositional Semantics
These two pairs raise the following problem: in both cases, the (b) sentence is not equivalent to the logical negation of the (a) sentence. Indeed, (1a) conveys that the homework contains more than one difficult problem, and (2a) that one of my students has solved two, or more than two, difficult problems. But (1b) does not mean that the problem doesn’t contain several difficult problems; it has the same truth-conditional content as ‘The homework doesn’t contain any difficult problem’. Likewise, (2b) doesn’t mean that none of my students has solved two or more problems (which could be true in a situation where a student has solved exactly one difficult problem), but rather that none of my students has solved any problem at all. In a nutshell, difficult problems in the (a) sentences receives an at-least-two-interpretation, while it receives an at-leastone-interpretation in the (b) sentences. One can wonder whether this kind of behavior is restricted to Bare Plurals (or, in French, to DPs headed by des), or is found with full DPs as well. In fact, plural DPs headed by some display a similar behavior, though this is harder to show due to the fact that such DPs, being positive polarity items, cannot take scope immediately below negation. Consider however: (3)
There was a set of problems. Some of them were easy, others were difficult. a. Jack solved some difficult problems b. It is unlikely that Jack solved some difficult problems
Again, while (3a) conveys that Jack solved at least two difficult problems, (3b), on the reading where some is interpreted within the scope of it is unlikely,1 does not merely assert that it is unlikely that Jack solved at least two difficult problems, but rather that it is unlikely that he solved even one difficult problem.2 One more fact: (4)
a. Exactly one of my students has solved difficult problems b. Exactly one of my students has solved some difficult problems
On its most natural interpretation, (4a) is equivalent to none of the following sentences: (5)
a. Exactly one of my students has solved at least one difficult problem b. Exactly one of my students has solved two, or more than two, difficult problems
Aspects of the Pragmatics of Plural Morphology 245
Indeed, (5a) (on the reading where at least one difficult problem takes scope below exactly one of my students) is true in a situation where one student, say Jack, has solved exactly one difficult problem, and no other student has solved any problem, while, intuitively, (4a) suggests that the only student who has solved a problem has solved at least two problems. On the other hand, (5b) is true in a situation where one student, say Jack, has solved exactly two problems, and all other students have solved exactly one problem; but, intuitively, (4a) entails that no student, apart from the unique one who has solved difficult problems, has solved any problem at all. In fact, the most natural interpretation we get for (4a) is the following:3 (6)
One of my student has solved several difficult problems, and all other students have solved no difficult problem at all.
The same point can be made with respect to (4b), though one must be careful, in this case, not to ignore the fact that, on one reading, some difficult problems takes scope over exactly one of my students. On the other reading, though, where some difficult problems takes narrow scope, we get the interpretation given in (6).4 This last fact shows that the various possible readings of bare plurals and some NPs (hereafter, plural indefinites) cannot be adequately described in terms of an ambiguity of plural indefinites between an at-least-two and an at-least-one reading. Though the existence of the at-least-one-reading for bare plurals has been recognized long ago (partly in relation with the phenomenon of dependent plurals), there is, to my knowledge, no theory available that accounts for the exact distribution of the various readings. Sauerland (2004b) and Sauerland et al. (2005) note that the at-least-one reading is found in monotone decreasing contexts, and provide a tentative account of this fact (see Section 6), but, as far as I can see, don’t provide an account of the kind of reading exemplified in (6), nor of the readings that arise within the scope of a universal quantifier (see Section 4.2). I’ll return to Sauerland’s proposal in Section 6.
2
Sketch of the analysis
The hypothesis I will defend in this chapter is the following: the meaning of plural morphology (on nouns) is in itself insufficient to account for atleast-two-readings. In particular, what compositional semantics delivers in the case of (1) is the proposition that the homework contains at least
246
Presupposition and Implicature in Compositional Semantics
one difficult problem.5 The at-least-two-reading, in this case, is derived as an implicature, based on the comparison of (1) with a sentence whose propositional content is equivalent to6 (7): (7)
The homework contains exactly one difficult problem
This hypothesis, as such, obviously lacks motivation, since the phrase exactly one difficult problem is not a natural alternative for difficult problems, because it is syntactically much more complex. I want to derive the fact that (1a) competes with a linguistic form whose meaning is that of (7) from the much more natural hypothesis that difficult problems competes with a difficult problem, as already argued by Krifka (1989).7 Though The homework contains a difficult problem is not semantically equivalent to (7) (and is in fact, given my assumptions, equivalent to (1)), it actually generally implicates (7). My main theoretical claim can be summed up as follows: when computing implicatures, hearers take into account not only the literal meaning of a given sentence’s alternatives, but also their pragmatically enriched meaning, hence the term higher-order implicature. The underlying idea is that a speaker might prefer a given sentence A over another one B that is equivalent to A, because she is aware that B leads to pragmatic inferences that she is not ready to endorse. And the hearer can actually reflect on the speaker’s choice of A over B, and conclude that the speaker does not believe the implicatures of B to be true. The formal implementation of this idea is presented in Section III. At the moment, I simply assume that exactly one difficult problem is an alternative of difficult problems. Since (7) is logically strictly stronger than (1a), one can conclude, from the maxim of quantity, that the author of (1) does not hold the belief that the homework contains exactly one difficult problem. An additional step leads from this inference to the stronger one that the author of (1) believes that the homework doesn’t contain exactly one difficult problem, which together with the fact that he is supposed to believe that (1a) is true, leads to the conclusion that the speaker believes that the homework contains at least two difficult problems. This additional step, whose motivation is not entirely obvious, is present in most present-day theories of scalar implicatures, in order to account, for instance, for the not all inference triggered by some in increasing contexts.8 This simple hypothesis makes a direct prediction: the at-least-two reading should disappear in monotone decreasing contexts. The ‘strengthened’ meaning that scalar terms get in monotone increasing context (some but not all for some, exclusive reading for the disjunction) is indeed
Aspects of the Pragmatics of Plural Morphology 247
absent in decreasing context, due to the reversal of entailment patterns, which reverses the informativity ordering within the set of alternatives. In the case at hand, what happens is that that the sentence one gets by replacing an expression like difficult problems with exactly one difficult problem is not stronger, but weaker, than the original sentence if difficult problems occurs in a monotone decreasing context, so that the reasoning sketched above for (1) will not apply.9 This prediction, which already accounts for the readings of (1b) and (2b), also accounts for the interpretations of the following examples: (8)
a. Every student who has solved some difficult problems will be rewarded b. Every person who has seen horses will come up c. It is impossible that Jack solved difficult problems d. Every time Jack sees horses, he gets afraid
(8a) entails that every student who has solved one, and only one, difficult problem, will be rewarded, (8b) that every person who has seen one, and only one horse will come up, and (8c) that it is impossible that Jack solved even one difficult problem. Finally, in (8d), the phrase ‘every time that Jack sees horses’ quantifies over all situations in which Jack sees one, or more than one, horse.10 What about cases where the plural indefinite occurs in a nonmonotonic context? According to some standard accounts of scalar implicatures, non-monotonic contexts should cancel scalar implicatures, because they give rise to alternatives that are not ordered with respect to each other in terms of logical strength. Illustration: (9)
a. Exactly one student read all of Balzac’s novels b. Exactly one student read most of Balzac’s novels
Suppose (9a) is the only scalar alternative of (9b); since (9a) does not entail (9b), a speaker who utters (9b) who also believes that (9a) is true will not have violated Grice’s maxim of quantity, since it is not the case that the other alternative he could have truthfully used (i.e. (9a)) is strictly more informative.11 Still, let us see what happens if we add to the literal meaning of (9b) the negation of (9a): (10)
Exactly one student read most of Balzac’s novels, and it is not the case that exactly one student read all of Balzac’s novels.
248
Presupposition and Implicature in Compositional Semantics
This is provably equivalent to: (11)
Exactly one student read most of Balzac’s novels, he didn’t read them all, and no other student read most of Balzac’s novels.
It turns out that (11) is, intuitively, the pragmatic meaning of (9b). This suggests that scalar implicatures are not simply derived by negating stronger alternatives. Many recent works (Chierchia, 2002; Spector, 2003, 2005; Sauerland, 2004a; van Rooij & Schulz, 2006) have argued that such a procedure yields too strong results in some cases (cf. ‘the problem of multiple disjunctions’), and too weak results in other cases. Following van Rooij & Schulz (2006), and Spector (2006), I assume that the strengthened meaning of any sentence S whose scalar alternatives is the set ALT(S) is given by the application of an exhaustivity operator, as defined below: exh(S) = {w: S is true in w and ¬∃w (S is true in w and w <S w},12 with w <S w if the members of ALT(S) true in w’ are a proper subset of the members of ALT(S) true in w. The reader will check that in case ALT(S) contains only one member besides S itself, call it S , then: exh(S) = S ∧ ¬S if S does not entail S exh(S) = S if S entails S . More generally, for most of the cases discussed in this chapter, the exhaustivity operator can be described as follows:13 Let S be a sentence; let ALT*(S) be defined as: ALT*(S) = {S : S ∈ALT(S)∧S does not entail S }. Then exh(S) = ∧ (S∧ ¬S1 ∧ ¬S2 ∧ . . . . ∧¬Sn ), where {S1 ,…, Sn } = ALT*(S).14 In the case of (8b), the predicted strengthened meaning is therefore exactly the one paraphrased in (11). Let us now return to (4a), repeated below as (12): (12)
Exactly one of my students has solved difficult problems
Assume that (12) has only one other alternative, which is equivalent to (13a) and (13b): (13)
a. Exactly one of my students has solved exactly one difficult problem b. There is a unique student who has solved one difficult problem and hasn’t solved more than one
Aspects of the Pragmatics of Plural Morphology 249
Since (12) does not entail (13), the application of the exhaustivity operator to (12) yields the proposition that is expressed by the conjunction of (12) and the negation of (13). Recall that the literal meaning of (12), by hypothesis, is Exactly one of my students has solved at least one difficult problem: (14)
Exactly one of my students has solved has solved difficult problems, and it is not the case that there is a unique student who has solved one difficult problem and no more.
This is equivalent to: (15)
Exactly of my students has solved several problems, and all other students have solved no problem at all.
As argued above, (15) is the most natural reading for (12). At this point, it is necessary to make my basic hypothesis more precise. Four questions arise: (a) what is the exact semantic content of plural morphology on nouns: I will argue, following Krifka (1989) and Schwarzschild (1996) (among others), that a plural NP like horses denotes the set of groups of horses, where even a single, individual horse counts as a group of horses. I will not make any specific assumption about the way existential readings are derived for bare plurals. There are many reasons to think that bare plurals should not be treated as being intrinsically existentially quantified.15 However, since the pragmatic account I am assuming is based only on assumptions about the global meaning of the sentences that are alternatives to each other, I need not take a clear stand on the issue how existential readings are derived for bare plurals. In this respect, the fact that we find similar patterns with genuine quantified plural indefinites (such as some NPs) shows that the problem of the distribution of at-least-one and at-least-two readings for plural indefinites is independent of the question how existential readings for bare plurals are derived. However, due to the various additional constraints that regulate the interpretation of DPs headed by some (i.e. the fact the such DPs are positive polarity items, and can receive a specific interpretation), I will mostly be concerned with sentences in which a bare plural receives an existential interpretation. (b) How come that a sentence like Jack saw horses competes with a sentence whose meaning is Jack saw exactly one horse? It would be natural
250
Presupposition and Implicature in Compositional Semantics
to assume that the competitor for Jack saw horses is simply Jack saw a horse (as already assumed by Krifka, 1989, Sauerland, 2003, 2004b, Sauerland et al. 2005 – see Section VI), whose literal meaning is equivalent to Jack saw at least one horse, as such a competition is arguably reducible to a competition between the singular morpheme and the plural morpheme. As explained above, the hypothesis I will put forward is that Jack saw horses competes with the pragmatically enriched meaning of Jack saw a horse, i.e. with the proposition that Jack saw exactly one horse. In this sense, at-least-two-readings are higher-order implicatures. It is necessary to offer a formal implementation of this idea. (c) How are scalar implicatures derived exactly? Depending on the exact way in which we characterize alternatives, we can end up with different predictions, typically when the plural indefinite takes scope below some other operator. For this reason, I delay the discussion of sentences like Every student has solved (some) difficult problems. (d) Last but not least, bare plurals seem to get an at-least-one-reading when they are under the scope of some other plural DP, as in My friends have beautiful wives.16 The present proposal does not aim at accounting for this phenomenon, also known as that of dependent plurals. For this reason, I will not use any example in which a plural indefinite is c-commanded by some other plural element.17
The rest of the chapter is organized as follows: Section 3 explains the proposal, based on the notion of higher-order implicature, and applies it to simple cases such as The homework contains difficult problems. Section 4 discusses the consequences of the proposal for more complex cases, such as those in which a plural indefinite is within the scope of a nonmonotonic quantifier or a universal quantifier. In the latter case, I will show that we make different predictions depending on how exactly alternatives are defined. Section 5 shows that even in contexts where a plural indefinite gets an at-least-one-reading, something of the at-leasttwo-reading is preserved, under the form of a modal presupposition; I argue that a similar pattern is found with disjunction (something of the exclusive reading is preserved even in contexts where a disjunction is truth-conditionally clearly interpreted as inclusive), which motivates further the treatment of at-least-two-readings in terms of implicatures. Section VI presents and discusses a proposal by Sauerland (2003, 2004b) and Sauerland et al. (2005), which is in part similar to mine. Section 7 is the conclusion.
Aspects of the Pragmatics of Plural Morphology 251
3
The proposal
3.1 Background assumptions about the semantics of plural NPs Semantic accounts of plural morphology can be roughly divided into two types. Based on the existence of at least two interpretations, one can postulate that the denotation of a bare plural of the form NPs is the set of all groups18 of cardinality more than one whose atomic individuals are in the extension of NP. In other words, if, in a given model, the extension of the singular noun horse is {a, b, c}, then that of horses is {a+b, a+c, b+c, a+b+c}. This is, for instance, Chierchia’s (1998) view. As a result, a sentence such as Jack saw horses ends up equivalent to There is a group of at least two horses such that Jack saw this group. The at-least-two reading is therefore a direct consequence of compositional semantics. The main shortcoming of this approach is that, obviously, the negation of the previous sentence, i.e. Jack didn’t see horses, is predicted to be true if Jack saw exactly one horse, contrary to fact. I will assume, on the contrary, that the plural morpheme denotes a function which applies to a set of atomic individuals and returns the set of all individuals that are made up of these atomic individuals. That is, if, in a given model, the denotation of horse is {a, b, c}, then that of horses is {a, b, c, a+b, a+c, b+c, a+b+c}. This is in line with Krifka (1989) and Schwarzschild (1996),19 among others. In other words, I am assuming the following semantics for the plural morpheme: Ontology: Each model is based on a domain of individuals which is structured by a part–whole relation (a partial order), such that, for any two individuals x and y, there exists a unique individual z, noted x + y (the mereological sum of x and y), such that x is a part of z, y is a part of z, and z is a part of any z that also contains x and y as parts. Furthermore, there exist atomic individuals, i.e. individuals whose only parts are themselves. The structure that is commonly used to embody these axioms is that of a join semi-lattice. I take the denotation of a singular NP to be a set of atomic individuals. x X reads as ‘x is a part of X’ (⊆ is used only to represent standard set-theoretic inclusion). The semantic contribution of the plural morpheme is then captured as follows: [[pl]] = λP.λX.(∀x(atomic(x) ∧ x X) → P(x) = 1)20 As to the singular morpheme, I assume it has no specific semantic import.21 A singular NP denotes the set of atomic individuals that are in the extension of the NP. For any NP, we thus have [[NP-sg]] ⊆ [[NP-pl]].
252
Presupposition and Implicature in Compositional Semantics
Furthermore, note that if a given NP, in a model M, is true of more than one atomic individual, then [[NP-sg]]M is properly included in [[NP-pl]]M , i.e. [[NP-sg]]M ⊂ [[NP-pl]]M .22 Such a semantics directly accounts for the reading we get for Jack hasn’t seen horses, but not for the at-least-two reading triggered by plural indefinites in Jack has seen horses and Jack has seen some horses. The goal of this chapter is to derive such readings as implicatures. Assume that, for any NP, forms a scale. Suppose a certain plural NP occurs in a sentence S and receives an existential interpretation, and let us write S(NPpl ) to denote this sentence, and S(a NPsg ) to denote the sentence that one obtains by replacing the plural NP by its singular counterpart and making the necessary adjustments in S with respect to agreement rules. If NPpl occurs in a monotone increasing context, then, given that, in general, [[NPsg]]⊂ [[NPpl]], we have [[S (a NPsg )]] ⊆ [[S(NPpl )]]. This as such does not exclude that S(a NPsg ) and S(NPpl ) be in fact equivalent. But if not, i.e. if S(a NPsg ) asymmetrically entails S(NPpl ), then, assuming that S does not itself contain any other scalar term, the pragmatic (i.e. exhaustive) meaning of S(NPpl ) will be S(NPpl ) ∧ S(a NPsg ). Unfortunately, in the most simple cases, S(a NPsg ) and S(NPpl ) are predicted to be equivalent. Let me illustrate with a simple example: (16) (17)
Jack saw a horse Jack saw horses
(17) has the following literal meaning: ∃ X (X ∈ [[horses]] ∧ Jack saw X), i.e. ∃ X (∀x((atomic(x) ∧ x X) → x ∈ [[horse]]) ∧ Jack saw X) On the other hand, (16)’s predicted meaning is: ∃ x (x ∈ [[horse]] ∧ Jack saw x) Because λX. Jack saw X is a distributive predicate, the following equivalence generally holds, for any X: Jack saw X ⇔ ∀x((atomic(x) ∧ x X) → Jack saw x) It is clear that (16) entails (17). But it turns out that (17) entails (16). Suppose indeed that Jack saw a certain group X of horses. Because he must then have seen every atomic member of X, he must have seen at least one atomic horse. It is therefore impossible to reason as if (16) asymmetrically entailed (17).
Aspects of the Pragmatics of Plural Morphology 253
3.2
The mechanism: higher-order implicatures
My main proposal is that when computing the pragmatic ‘strengthened’ meaning of (17), we take into account not only the literal meaning of its only competitor, but also its strengthened meaning. Note indeed that (16) itself implicates the following: (18)
Jack saw exactly one horse
The assumption that the computation of the pragmatic meaning of a given sentence should take into account not only the literal meanings of its competitors, but also their own pragmatic meaning is entirely natural in a Gricean framework: scalar implicatures, indeed, are not only inferences produced by the hearer regarding the speaker’s mental state; as they are based on general conversational principles, the speaker himself knows that the hearer will derive them, which is why he can be said to mean them, i.e. to take responsibility not only for the literal reading of his utterance, but for its pragmatic implications as well. For this reason, the fact that a given sentence triggers a certain implicature is truly common knowledge; it follows that there is no reason whatsoever why pragmatic processes that are based on the assumption that a given sentence competes with its alternatives should not take into account the implicatures of the competitors as well as their literal meanings. As I have already pointed out, it is perfectly natural to assume that a speaker might choose not to use a certain sentence because he does not want to convey the implicatures of this sentence. In this sense, the hypothesis that higher-order implicatures exist is the null hypothesis.23 In the case at hand, one can implement the mechanism as follows: I assume that bare plurals contain a null plural determiner, whose singular syntactic counterpart is the indefinite article a/an,24 noted as Øpl . I do not want to assume that this null determiner is responsible for the existential reading of bare plurals; it may be semantically null as well.25 I furthermore assume that the singular morpheme and the plural morpheme constitute a scale. The scalar alternatives or a given sentence are obtained by (possibly multiple) substitutions of a scalar item with a member of the same scale, together with all the adjustments that are required by agreement rules.26 In the case at hand, Øpl NP−pl gets replaced with a NP−sg. On the other hand, I assume that the singular indefinite articles a and some compete with several, to the effect that a sentence like There is a horse in the garden triggers the implicature that There aren’t several horses in the garden. I therefore postulate the two following scales: and
254
Presupposition and Implicature in Compositional Semantics
.27 The mechanism giving rise to higher-order implicatures can be formalized as follows:28 For any S, I call i0 (S) the literal interpretation of S (a proposition, i.e. a set of worlds), which is the result of compositional semantic rules, and ALT(S) the set of its scalar alternatives. I call ALT0 (S) the set of the propositions that are expressed by the members of ALT(S). In other terms: ALT0 (S) = {φ: ∃ S (S ∈ ALT(S) ∧φ=i0 (S )} The first-order pragmatic meaning of S is then noted i1 (S), and defined as follows: i1 (S) = exh(S), i.e: i1 (S) = {w: w ∈ i0 (S) ∧¬∃w (w ∈ i0 (S) ∧ w <S,0 w)}, with: w <S,0 w iff the members of ALT0 (S) that are true in w are a proper subset of the set of members of ALT0 (S) that are true in w. Then we define inductively, for any n, and any S, ALTn (S) and in (S), as follows: – ALTn (S) = {φ: ∃ S (S ∈ ALT(S) ∧φ = in (S )} – in+1 (S) = {w: w ∈ in (S) ∧¬∃w (w ∈ in (S) ∧ w <S,n w)}, with: w <S,n w iff the members of ALTn (S) that are true in w are a proper subset of the set of members of ALTn (S) that are true in w . For any n, in+1 (S) is the exhaustive interpretation of S relatively to the semantic values of S s alternatives at step n. In other words, in (S) is the nth -order pragmatic meaning of S. Whenever each alternative set contains only two members, one can rephrase this inductive definition as follows: if S is the unique alternative of S, then in+1 (S) = in (S) ∧ ¬in (S ) if in (S) does not entail in (S ). Otherwise, in+1 (S) = in (S). Illustration: (19) (20) (21)
Jack saw horses Jack saw a horse Jack saw several horses
By hypothesis, the scalar alternatives of these three sentences are as follows (assuming that several itself is not part of a scale, which is certainly a simplifying assumption, but one that does not in fact affect the general
Aspects of the Pragmatics of Plural Morphology 255
result; if we assumed the scale <several, many>, for instance, this would not affect the pragmatic meaning of (19)): ALT((19)) = {Jack saw horses, Jack saw a horse} ALT((20)) = {Jack saw a horse, Jack saw horses, Jack saw several horses} ALT((21)) = {Jack saw several horses} We first need to turn these sets of sentences into the sets of propositions that are the literal meanings of these sentences. For any S, I note ∧ S the proposition that S expresses. By hypothesis, we have: i0 ((19)) = ∧ Jack saw at least one horse i0 ((20)) = ∧ Jack saw at least one horse i0 ((21)) = ∧ Jack saw at least two horses As ALT((19)) includes two equivalent sentences (given the semantics I am assuming for the plural morpheme), ALT0 ((19)) will contain only one proposition: ALT0 ((19)) = {∧ Jack saw at least one horse} We also get: ALT0 ((20)) = {∧ Jack saw at least one horse, ∧ Jack saw at least two horses} ALT0 ((21)) = {∧ Jack saw at least two horses} For any n, and any sentence S, I say that a given proposition S is a level-n alternative of S if S belongs to ALTn (S). At the next stage, we compute the exhaustive reading of each of these three sentences with respect to its level-0 alternatives. As (19) and (21) have no level-0 alternative distinct from themselves, they are obviously equivalent to their exhaustive readings, and we thus have: i1 ((19)) = ∧ Jack saw at least one horse i1 ((21)) = ∧ Jack saw at least two horses On the other hand, (20) has two distinct level-0 alternatives, and, since (20) does not entail its only level-0 alternative, its interpretation at stage 1 is obtained by adding the negation of this alternative: i1 ((20)) = i0 ((20)) ∧¬i0 (21) = ∧ Jack saw exactly one horse29
256
Presupposition and Implicature in Compositional Semantics
We now turn to the sets of level-1 alternatives. By definition, we have: ALT1 ((19)) = {i1 ((19), i1 ((20))} ALT1 ((20)) = {i1 ((20)), i1 ((21))} ALT1 ((21)) = {i1 ((21))} i.e.: ALT1 ((19)) = {∧ Jack saw at least one horse, ∧ Jack saw exactly one horse} ALT1 ((20)) = {∧ Jack saw exactly one horse, ∧ Jack saw at least two horses} ALT1 ((21)) = {∧ Jack saw at least two horses} The level-2 reading of (19) is then given by: i2 ((19)) = i1 ((19)) ∧¬ iI ((20)) = ∧ (Jack saw at least one horse and Jack didn’t see exactly one horse) = ∧ Jack saw at least two horses. This is the desired result. On the other hand, we have i2 ((20)) = i1 ((20)) ∧¬ ii ((21)). But because ii ((20)) anyway entails the negation of ii ((21)), this is simply equivalent to: i2 ((20)) = i1 ((20)). Likewise, i2 ((21)) = i1 (21), which is itself equal to i0 ((21)). The reader will convince himself that for any n > 1, in ((20)) = i1 ((20)) and in ((21)) = i0 ((21)) At this point, let me consider what happens at the next stage for (19). We have: ALT2 ((19)) = {i2 ((19)), i2 ((20))}, i.e. ALT2 ((19)) = {∧ Jack saw at least two horses, ∧ Jack saw exactly one horse} As the two members of this set are mutually incompatible, i2 ((19)) ∧¬ i2 ((20)) is equivalent to i2 ((19)), from which it follows that i3 ((19)) = i2 ((19)). In fact, for any n > 2, in ((19)) = i2 ((19)). It is fortunate that for each example, we reach a state after which its interpretation cannot be altered by any number of additional applications of the exhaustivity mechanism. Otherwise, we might expect an indefinite process that never yields a final pragmatic interpretation. In the appendix, I prove that, under certain hypotheses, this is always the case: for any sentence S, there exists an integer n such that for any m > n, im (S) = in (S).30 In fact, in stereotypical cases of scalar implicatures, the final interpretation is reached at level 1, which makes the concept of higher-order implicature empirically irrelevant. In order for there to be higher-order implicatures, we in fact need highly unusual scales. Let me now briefly turn to monotone decreasing contexts. If a bare plural occurs in a decreasing context, its competitor contains a singular
Aspects of the Pragmatics of Plural Morphology 257
indefinite NP in a decreasing context: as a result, the strengthening from a NP to exactly one NP (or from [….. a NP….] to [….a NP…] and not […. several NPs….]) will not take place, thus blocking higher-order implicatures for the original sentence.
4
More complex cases
4.1
Non-monotonic contexts
Let me now return to sentence (4a) (repeated below as (22)), i.e. to a case where the plural indefinite is interpreted within the scope of a nonmonotonic quantifier: (22)
Exactly one of my students has solved difficult problems
In Section II, I assumed that this sentence’s only alternative was: (23)
Exactly one of my students has solved exactly one difficult problem
However, this assumption is no longer valid; rather, the only alternative for (22) is the following:31 (24)
Exactly one of my students has solved a difficult problem
By hypothesis, (22) will be compared with the strengthened meaning of (24), which is not equivalent to (23), as we’ll see below. Let us apply the mechanism described above. At the first stage, nothing happens for (22), because i0 ((22)) = i0 ((24)). In other words, we have i1 ((22)) = i0 ((22)). In order to compute i2 ((22)), we have to determine the strenghtened meaning of its only alternative, i.e. i1 ((24)). By hypothesis, ALT((24)) = {Exactly one of my students has solved a difficult problem, Exactly one of my students has solved difficult problems, Exactly one of my students has solved several problems}, and ALT0 ((24)) = {∧ Exactly one of my students has solved at least one difficult problem, ∧ Exactly one of my students has solved at least two different problems} Since the first of these two propositions is compatible with the negation of the second proposition, we have: i1 ((24)) =∧ (Exactly one of my students has solved at least one difficult problem & ¬(Exactly one of my students has solved at least two difficult problems)).
258
Presupposition and Implicature in Compositional Semantics
This turns out to be equivalent to: i1 ((24)) =∧ (Exactly one of my students has solved exactly one difficult problem, and no other student has solved any problem at all) Note that this last proposition is distinct from ∧ (Exactly one of my students has solved exactly one difficult problem). The prediction that i1 ((24)) is the pragmatic reading of (24) seems to be correct, and is a strong argument in favour of an exhaustivity–based account of scalar implicatures.32 The level-1 alternatives of (22) are therefore given by the following set: ALT1 ((22)) = {i1 ((22)), i1 ((24))} = {∧ Exactly one of my students has solved at least one problem, ∧ (Exactly one of my students has solved exactly one difficult problem and no other student has solved any problem at all)} We now have to compute i2 ((22)), i.e. i1 ((22)) ∧¬ i1 ((24)), i.e.: (25)
Exactly one of my students has solved exactly at least one difficult problem, and it not the case that exactly one of my students has solved exactly one difficult problem and no other student has solved any difficult problem at all
I will informally prove that this last sentence is equivalent to: (26)
Exactly one of my students has solved at least two difficult problems, and all other students have solved no difficult problem at all.
Let me prove that (25) entails (26): suppose (25) is true. Then there is a student, call him Jack, who solved a difficult problem, and is such that all other students have solved no difficult problem. Suppose Jack had solved exactly one difficult problem. Since it is not the case that exactly one student has solved exactly one difficult problem and no other student has solved any difficult problem, it follows that either a student different from Jack has solved exactly one difficult problem, or that Jack is the only one that has solved exactly one difficult problem but there are other students that have solved a difficult problem; in both cases, there must be a student different from Jack who has solved a difficult problem, which is contradictory (since Jack is the only one who solved a difficult problem). Therefore Jack has solved several difficult problems; furthermore, no other student solved any difficult problem.
Aspects of the Pragmatics of Plural Morphology 259
(26) entails (25): suppose (26) is true. There is then a unique student, call him Jack, who solved several difficult problems, and all other students have solved no difficult problem at all. Jack is also the only student who solved at least one difficult problem, and it is furthermore the case that there is no student who solved exactly one difficult problem, from which the truth of (25) follows. We therefore end up with the same interpretation as in Section II for (12), a welcome result. 4.2
Plural indefinites under the scope of a universal quantifier
Consider the following sentence: (27)
Yesterday, every student solved (some) difficult problems33
Most informants report the following: on the one hand, one tends to understand, from sentence (27), that every student solved several difficult problems. On the other hand, if uttered in a situation where every student solved at least one difficult problem, and only some of them solved several difficult problems, the sentence is not judged false. Finally, in a situation where every student has solved exactly one difficult problem, the sentence is felt as ‘misleading’, as one informant puts it. We are therefore dealing with very delicate judgments. It seems that though (27) suggests that every student solved several difficult problems, it is not incompatible with a situation in which some students solved only one while all the others solved more than one problem. For this reason, I assume that (27) can have one of the two following readings:34 (28)
a. Yesterday, every student solved several difficult problems b. Yesterday, every student solved at least one difficult problems, and at least one student solved several difficult problems
I call the (a) reading the ‘strong reading’ and the (b) reading the ‘weak reading’. Before turning to the predictions that are made by a theory in terms of higher-order implicatures, let me first see what the predictions are if we assume that difficult problems competes directly with exactly one difficult problem. Under such a hypothesis, (27)’s alternative would be: (29)
Yesterday, every student solved exactly one difficult problem
260
Presupposition and Implicature in Compositional Semantics
As (29) is not entailed by (27), the predicted strengthened meaning for (27) is the conjunction of (27) and the negation of (29), i.e.: (30)
Yesterday, every student solved at least one problem, and not every student solved exactly one difficult problem
This is in turn equivalent to: (31)
Yesterday, every student solved at least one problem, and at least one student solved several problems
Such a hypothesis therefore predicts the weak reading. What about the predictions made by the analysis in terms of higher-order implicatures? As we’ll see in a moment, the predictions that are made in this case depends on the exact way we characterize alternatives. Because the sentence contains two scalar items (every student, difficult problems), one has to wonder whether both of them contribute to the definition of alternatives. I will show that if only the alternatives induces by the plural indefinite are ‘activated’, then we end up with the strong reading. On the other hand, if both every student and difficult problems are ‘activated’, i.e. contribute both to the definition of the alternatives, then we end up with the weak reading (see below). Ignoring, for the moment, the fact that every student may well itself induce additional alternatives (given the fact that every and some belongs to the same scale – since this is necessary to derive the fact that not every student came implicates some students came), the only alternative of (27) (repeated below as (32)) is (33): (32) (33)
Yesterday, every student solved difficult problems Yesterday, every student solved a difficult problem
These two alternatives are, as far as their literal meaning is concerned, equivalent. It follows that the level-1 meaning of (32) is the same as its level-0 meaning, i.e. its literal meaning. But (33), by hypothesis, competes with (34) (34)
Yesterday, every student solved several difficult problems
As a result, the level-1 interpretation of (33) is expressed by (35): (35)
Yesterday, every student solved a difficult problem and not every student solved several difficult problems
Aspects of the Pragmatics of Plural Morphology 261
It follows that the level-2 reading of (32), which is equivalent to the conjunction of its level-1 meaning and the negation of the level 1-meaning of its competitor, is given by (36): (36)
Yesterday, every student solved at least one difficult problem and it is not the case that every student solved a difficult problem and not every student solved several difficult problems
The reader will check that this is in fact equivalent to: (37)
Yesterday, every student solved several difficult problems
The theory in terms of higher-order implicatures is therefore able to derive the strong reading. More generally, as long as we don’t take into account the alternatives that may be induced by the presence of other scalar items, we predict that a plural indefinite that occurs in a monotone increasing context will always get an at least two interpretation,35 even when it is under the scope of some other operator. Let me show this informally. Let S(NPpl) be a sentence containing a plural indefinite NPpl that occurs in an increasing context. For any X, I write S(X) to refer to the sentence that is obtained from S by replacing NPpl by X in S. By hypothesis, we have ALT(S(NPpl)) = {S(NPpl ), S(a NPsg)}, and ALT(S(a NPsg)) = {S(a NPsg), S(several NPpl)}. Assume that S(NPpl ) and S(a NPsg ) have the same literal meaning, which is always the case if NPpl is an argument of a distributive predicate. In this case, i1 (S(NPpl)) = i0 (S(a NPsg)). On the other hand, since a NPsg occurs in a monotone increasing context in S(a NPsg), S(a NPsg) implicates S(a NPsg) ∧¬ S(several NPpl). In other words, i1 (S(a NPsg)) = i0 (S(a NPsg) ∧¬ i0 (S(several NPpl)). As a result, we have i2 (S(NPpl)) = i1 (S(NPpl)) ∧¬i1 (S a NPsg), i.e. i2 (S(NPpl)) = i0 (S(NPpl) ∧ ¬(i0 (S(a NPsg)) ∧ ¬i0 (S(several NPpl))), i.e. i2 (S(NPpl)) = i0 (S(NPpl) ∧ (¬i0 (S(a NPsg)) ∨ i0 (S(several NPpl))). Since i0 (S(NPpl)) is equivalent to i0 (S(a NPsg)), this is equivalent to: i2 (S(NPpl)) = i0 (S(NPpl) ∧ i0 (S(several NPpl)). Since S(NPpl) defines a monotone increasing context for the element that occurs in the position of NPpl, i0 (S(several NPpl)) entails i0 (S(NPpl)), so that this last formula is equivalent to: i2 (S(NPpl)) = i0 (S(several NPpl)) In other terms, the level 2 interpretation of S(NPpl) is equivalent to the level-0 interpretation of S(several NPpl). We therefore get an at-leasttwo-reading for the plural indefinite at stage 2. Furthermore, this reading is not altered when we get to stages 3, 4, etc.
262
Presupposition and Implicature in Compositional Semantics
What if we also take into account the scale <some, every>? To make things sufficiently simple,36 let me now assume the following scales: , , <some NPsg, every NP> The set of alternatives for (32), repeated as (38), is given below in (39): (38)
Yesterday, every student solved difficult problems
ALT((38)): (39)
a. b. c. d.
Yesterday, every student solved difficult problems Yesterday, every student solved a difficult problem Yesterday, some student solved difficult problems Yesterday, some student solved a difficult problem
We therefore have (ignoring the adjunct yesterday): ALT0 ((38)) = {∧ every student solved at least one difficult problem, ∧ some student solved at least one difficult problem} Since the first member of this set entails the second one (as soon as there exist students), the meaning of (38) at stage 1 is the same as its literal meaning: i1 ((38)) = i0 ((38)). In order to compute i2 ((38)), we must first determine the content of ALT1 ((38)), i.e. the strengthened meanings of all of the alternatives of (38): Alternatives of (39)b: (40)
a. b. b. c. d. e.
Yesterday, every student solved a difficult problem Yesterday, every student solved difficult problems Yesterday, every student solved several difficult problems Yesterday, some student solved a difficult problem Yesterday, some student solved difficult problems Yesterday, some student solved several difficult problems
ALT0 ((39)b) includes all the propositions expressed by one of the sentence in (40), i.e.: ALT0 ((39)b) = {∧ every student solved at least one difficult problem, ∧ every student solved at least two difficult problems, ∧ some student solved at least one difficult problem, ∧ some student solved at least two difficult problems} In order to compute i1 ((39)b), we apply the exhaustivity operator to (39b) relatively to ALT0 ((39)b), which results in the conjunction of
Aspects of the Pragmatics of Plural Morphology 263
((39)b) and the negations of all the alternatives that are not entailed by (39b) (that is, every alternative except some student solved at least one difficult problem). We end up with: i1 ((39)b) = i0 (((39)b) ∧ ¬ (every student solved at least two difficult problems) ∧¬ (some student solved at least two difficult problems), which is equivalent to: i1 (((39)b) = every student solved exactly one difficult problem Similar considerations yields the following 1-level readings for the other alternatives of (38) (listed in (39)): i1 (((39)c) = ∧ some student, but not all the students, solved at least one difficult problem i1 ((39)d) = ∧ some student solved exactly one difficult problems, and not every student solved a problem Therefore, ALT1 ((38)) ={∧ every student solved at least one difficult problem, ∧ every student solved exactly one difficult problem, ∧ some student, but not all, solved at least one difficult problem, ∧ some student solved exactly one difficult problem, and not every student solved a problem} In order to compute i2 ((38)), we need, for each sentence S of ALT1 ((38)) that is not entailed by i1 ((38)), to add the negation of S. In the case at hand, the members of ALT 1 ((38)) that are not entailed by i1 ((38)) are the following ones: (41)
a. every student solved exactly one difficult problem b. some student, but not all, solved at least one difficult problem c. some student solved exactly one difficult problem, and not every student solved a problem
In fact, (41b) and (41c) contradicts (38), so that i1 ((38)) already entails their negation. For this reason, we end up with: i2 ((38)) = i1 ((38)) ∧ ¬ (∧ every student solved exactly one difficult problem) = every student solved at least one difficult problem, and at least one student solved several difficult problems We therefore now derive the weak reading, a welcome result as well. The theory in terms of higher-order implicatures is thus able to predict both the weak and the strong reading, provided we allow different options regarding the way alternatives are characterized. We have to assume that some scalar items can be ignored when we compute alternatives. The exact constraints that regulate the computation of alternatives when several scalar terms are present are in themselves a topic for research.
264
Presupposition and Implicature in Compositional Semantics
5 An enigma that confirms the hypothesis: the modal presupposition induced by plural indefinites I have not yet mentioned another crucial property of the interpretation of plural indefinites: even in contexts where the plural indefinite receives an at-least-one-reading (i.e. those in which no pragmatic strengthening takes place), something of the intuitive notion of plurality is retained, as evidenced by the following contrasts: (42)
a. b. c. d.
Jack doesn’t have a father # Jack doesn’t have fathers Jack doesn’t have a brother Jack doesn’t have brothers
The reason for the oddness of (42b) is felt as obvious: such a sentence suggests that it is at least possible to have several fathers. If we were talking about adoptive fathers in a society where one can be adopted by several men, then (42b) would be perfectly natural. It could seem that this phenomenon is readily explained by the fact that it makes no sense to deny something that is obviously false: because Jack has fathers conveys that Jack has several fathers, something which we know is false, there is no point to negate it. This reasoning, however, is in this case obviously flawed; indeed, as we explained at length, the truthconditional content of (42b) when it is appropriate (as in the case where we are talking about adoptive fathers) is equivalent to Jack has no father, and not to Jack doesn’t have several fathers. As noticed before, (42b) and (42c) are hardly distinguishable, and, at the very least, have the same truth-conditional content.37 More generally, it seems that even in cases where a bare plural NPpl receives an at least-one interpretation, it triggers the presupposition that the predicate of which the bare plural is an argument could be true of more than one NP. This general property is further illustrated by the following examples. (43)
a. In France, every student who gets a high-school diploma can enter a university b. # In France, every student who gets high-school diplomas can enter a university
(44)
a. I forgot to book a room
Aspects of the Pragmatics of Plural Morphology 265
b. #I forgot to book rooms In contrast with this, note that the two following sentences, within the stipulated context, are equally appropriate: (45)
a. I forgot to read a paper b. I forgot to read papers
Where could such a condition come from? We could of course include a presuppositional condition in the semantic characterization of plural indefinites. As I will now show, such a move probably misses a generalization that pertains to other scalar items as well. Let me first sum up our observations in the following table:
pragmatic interpretation of a plural indefinite increasing contexts at-least-two reading decreasing contexts at-least-one reading
presupposition
Poss(…at least two…)
Poss(. . . at least two . . .) is a schematic way of representing the presupposition according to which it was compatible with common knowledge, before the sentence was utterred, that the predicate of which the bare plural NP is an argument be true of at least two NPs. By analogy, let us see what a similar table would look like for a scalar item such as or:
pragmatic interpretation of a disjunction presupposition increasing contexts exclusive disjunction decreasing contexts inclusive disjunction
Poss(. . . orexcl . . .)
In other words, if disjunction and bare plurals behave similarly, we expect that a sentence like Mary didn’t meet Peter or Jack (on the reading where or
266
Presupposition and Implicature in Compositional Semantics
takes scope below negation) asserts, on the one hand, that Mary didn’t meet Peter and didn’t meet Jack, and presupposes, on the other hand, that it could have been true that Mary met Peter or Jack but not both. Is this prediction borne out? Consider the following context: Peter and Jack are Siamese twins. As a result, it is impossible to meet Peter without meeting Jack, and the other way around. That is, common knowledge excludes that Mary met Peter or Jack but not both; she either met both or none of them. The prediction is that the above sentence should be inappropriate in this case, and this prediction seems to be correct. To see why this fact is surprising, one must pay attention to the fact that, in such a context, the following four sentences are contextually equivalent:38 (46)
a. b. c. d.
Mary didn’t meet Peter and Jack Mary didn’t meet Peter or Jack Mary didn’t meet Peter Mary didn’t meet Jack
Since, in this context, it is common knowledge that Mary met neither Jack nor Peter, or met both Jack or Peter, all these four sentences are true only in situations in which Mary met both. Yet only (46a) is felicitous.39 Of course, we observe the same constraint if we turn to the positive counterparts of the previous sentences: (47)
a. b. c. d.
Mary met Peter and Jack Mary met Peter or Jack Mary met Peter Mary met Jack
In this case, it could be argued that the reason why (47a) is preferred to all others is due to the maxim of quantity: among these four alternatives, the speaker ought to choose the one that is logically the strongest. Note, however, that such an explanation works only if we force this kind of pragmatic mechanism to have no access to contextual knowledge: indeed, these sentences are contextually equivalent, so that if the maxim of quantity takes into account a contextual notion of logical strength, the above reasoning is invalid, as none of these sentences is contextually strictly stronger than another one.40 Furthermore, such an account would make wrong predictions in the case of (46); indeed, in this case, the sentence that is actually chosen is the one that is logically the weakest (the variant with and).
Aspects of the Pragmatics of Plural Morphology 267
At the moment, I can offer no convincing account of these contrasts, a topic that I leave for further research.41 However, the main point of this discussion is to stress the similarity between the behavior of plural indefinites and at least some other scalar items: it seems to be a general fact that some scalar items, even in contexts where they lose their “strengthened” meaning, induce a modal presupposition in which the strengthened meaning resurfaces.
6 Sauerland’s alternative (Sauerland 2003, 2004b, Sauerland et al. 2005) Sauerland (2003, 2004b) and Sauerland et al. (2005) suggest a pragmatic and semantic analysis of the singular/plural distinction which shows clear similarities with the present proposal. According to these works, the singular morpheme triggers the presupposition that the denotation of its sister (the noun or NP it combines with) includes only atomic individuals. The plural morpheme, on the other hand, has no semantic contribution (it denotes the identity function). Sauerland (2003, 2004b) and Sauerland et al. (2005) furthermore assume that there exists a phonologically null operator, noted ∗ , which turns a predicate into its distributive variant, so that, if a given predicate is true of {a, b, c}, then the operator yields a predicate that is true of {a, b, c, a+b, a+c, b+c, a+b+c}. This operator applies freely, but, given the presupposition induced by the singular morpheme, a structure like [sg * NP] is always ill-formed. The result is that when the distributivity operator is present, then the plural morpheme must be used. This, however, does not prevent as such the plural morpheme from applying to a predicate that is true only of atomic individuals. Consider, in this respect, the following two sentences:
(48) (49)
Jack met the student Jack met the students
In the case of (48), the uniqueness presupposition triggered by the student is explained in the following manner: since student is singular, it must denote a set of atoms, which entails that the operator ∗ is absent. Assuming a standard maximality based-semantics (Link 1983) for the definite article, we derive that the student must denote the maximal element in the denotation of student, and presupposes that such a maximal element
268
Presupposition and Implicature in Compositional Semantics
exists. But this condition can only be met if the denotation of student-sg contains only one atom. Let me turn to (49): in this case, the plural morpheme imposes no specific condition on the denotation of students. Whatever this denotation is, the students denote the maximal individual that is in the denotation of students; if ∗ is present, we end up with the individual that contains all students as parts. In a situation where there exists only one student, the students will actually denote the only student that exists. This is not a welcome result, as in such a situation, (49) is inappropriate, i.e. presupposes that there exist at least two students. To account for this presupposition, Sauerland (2003, 2004b) and Sauerland et al. (2005) use the notion of implicated presupposition, based on a maxim called Maximize presupposition!. In a nutshell, given the semantics that he assumes, (48) and (49) are Strawson-equivalent, i.e. have the same truth conditions in all the contexts in which their presuppositions are satisfied. In such a case, Maximize presupposition! requires that a speaker who wants to convey the propositional content of these sentences should use the one whose presuppositions are the strongest. It follows that, if it is common knowledge that there is only one student, then (48) only can be used. As a result, (49) is appropriate only in contexts where it is not common knowledge that there exists only one student; this as such delivers the presupposition it is possible, according to what is common knowledge, that there exists more than one student. Sauerland (2003, 2004b) and Sauerland et al. (2005) assume that this presupposition gets in fact strengthened, so that we finally get: there exists more than one student. This proposal and mine shares the thesis that the at-least two interpretation associated with the plural morpheme is not part of its intrinsic semantics, but results from a competition between the plural and the singular morpheme. However, Sauerland (2003, 2004b) and Sauerland et al. (2005) face conceptual and empirical difficulties when dealing with indefinites and existentially interpreted bare plurals. I am going to present the account offered in Sauerland et al. (2005), which is actually different from the one in Sauerland (2003), (which Sauerland et al. (2005) explicitly rejects – as far as I can see, Sauerland (2004b) gives more or less the same account as Sauerland et al. (2005)). Sauerland et al.’s (2005) proposal is based on the following two principles :
(A) presuppositions that are triggered by a presuppositional item under the scope of an existential quantifier do not project at all, i.e. are incorporated into the assertive component.
Aspects of the Pragmatics of Plural Morphology 269
Example (mine): (50)
Some country killed its king >> Presupposes nothing, and asserts that some country that had a king killed its king
(B) Maximize presupposition applies to the scope of an existential if this strengthens the entire utterance Let me then return to the following examples: (51)
a. The homework contains difficult problems b. The homework contains a difficult problem
(52)
a. The homework doesn’t contain difficult problems b. The homework doesn’t contain a difficult problem
As a result of principle A, (51b) presupposes nothing, and asserts that the homework contains an atomic entity that is a problem, while (51b) asserts that the homework contains a group of at least one problem. Given that the predicate ‘being contained in the homework’ is distributive, these two sentences have in fact the same literal content, just as in my own account. Now Maximize presupposition! applies to the scope of the existential operator, which, by assumption, is present in the logical form of (51a) and (51b) (accounting for the existential interpretation of the indefinites); that is, the presupposition that is triggered by difficult problem-sg, i.e. that the extension of difficult problem contains only atomic entities is compared to the presupposition triggered by difficult problems, i.e. to a null presupposition. Though Sauerland et al. (2005) does not say explicitly how Maximize presupposition! is defined in this case, one has to assume that difficult problems, as a result of Maximize presupposition, now triggers the presupposition that the extension of difficult problems does not contain atomic entities.42 Let me stress that this cannot be obtained simply by negating the presupposition triggered by the competitor, i.e. that the extension of difficult problem contains only atomic entities, since negating this last statement would give us something weaker, i.e. that the extension of difficult problems does not contain only atomic entities. When the result of this process is incorporated into the scope of the existential operator, one gets the homework contains several difficult problems, which is strictly stronger than the literal reading of (51a), and therefore meets the condition imposed by Principle (B). It is clear that in the case of (52a), one would not get a stronger meaning by means of a similar mechanism, but a weaker one, so that
270
Presupposition and Implicature in Compositional Semantics
Principle (B) prevents the above-described process to apply, to the effect that we get an at-least-one-reading for the plural indefinite (and, in fact, in all cases where a plural indefinite occurs in a monotone-decreasing context). Sauerland must be credited for the generalization that at-least-tworeadings vanish in monotone decreasing contexts, and for making a general case that the interpretative effects of the plural morpheme are to be understood as resulting from a competition between the singular and the plural. In this respect, my own proposal is a direct heir to his. Yet the account sketched by Sauerland et al. (2005) seems to me quite stipulative, mostly because of the role played by Principle (B). In addition to the lack of motiviation for this principle, one has to provide an explicit description of how Maximize presupposition! works at the level of a sub-constituent, and it seems to me that Sauerland et al. (2005) does not provide a complete formal description of this mechanism. As observed above, it is not clear at all how one reaches, from the negation of the presupposition attached to a singular NP (i.e. the negation of the presupposition that the extension of the NP contains only atomic individuals), the conclusion that the extension of the corresponding plural NP contains only non-atomic individuals. Note also that principle (B) cannot account for the exact behavior of plural indefinites within the scope of a non-monotonic existential operator; one could think of the following weakening of Principle (B): Principle B : Maximize presupposition applies to the scope of an existential if this does not weaken the entire utterance Such a modification predicts that sentence (4)a, repeated below as (53), gets interpreted as Exactly one of my students has solved several difficult problems, contrary to fact, as argued above.43 (53)
Exactly one of my students has solved difficult problems
Likewise, Sauerland et al. (2005) cannot match the predictions of the present proposal for cases where a plural indefinite is under the scope of a universal quantifier. For these reasons, my own account seems to me to be preferable. It is worth mentioning that Sauerland’s primary goal in these various works is not to account for the readings of plural indefinites, but to give a general theory of the presuppositions induced by number-features (and, more generally, by all phi-features, cf. Sauerland (2004b)); he derives many non-trivial facts from the principle Maximize Presupposition! and the semantics he assumes for number morphemes. It turns out that my
Aspects of the Pragmatics of Plural Morphology 271
own account is compatible with Sauerland et al.’s (2005) general theory, and could be used to replace their own account of the semantics of plural indefinites, without entailing any modification of their general framework. On the other hand, on the basis of my own account and Maximize presupposition!, one can in fact analyze plural definite descriptions in exactly the same way as Sauerland et al. (2005) does: if students denotes the set of all groups of students, while student denotes the set of all individual students, then the students denotes the maximal element in the denotation of students, i.e. the maximal group of students, and the student denotes the unique student, if there is only one student; Maximize pressupposition! then predicts that in case it is known that there is only one student, the students should be blocked by the student.
7
Conclusion
This chapter’s central claim is that at-least-two-readings of plural indefinites are a special kind of scalar implicature: they arise from the fact that the plural morpheme competes with the singular morpheme, and from the existence of a mechanism of higher-order implicatures, to the effect that the scalar implicatures triggered by the competitors of a given sentence are taken into account for the computation of the pragmatic meaning of the sentence in question. I have offered an explicit formalization of this mechanism. I assumed that the strengthened meaning of a given sentence is equivalent to the result of an application of an exhaustivity operator; this as such does not mean that the actual process whereby scalar implicatures are derived involves an exhaustivity operator. Rather, as shown by van Rooij & Schulz (2006) and Spector (2003, 2005, 2006), the exhaustivity operator can be viewed as a shortcut, i.e. as a way of computing the result of a purely pragmatic reasoning. I am agnostic as to whether or not one should incorporate an exhausitivity mechanism into the grammar proper; let me note that if we choose to do so, it is possible to implement the present account by allowing the presence of an indefinite number of exhaustivity operators in the logical form of a given sentence.44
272
Presupposition and Implicature in Compositional Semantics
Appendix The goal of this appendix is to prove that, if a certain condition is met by the alternative set of certain sentence S (the finiteness condition, see below), then there exists an integer n such that the nth -level reading of S (as we have defined it) does not change when you iterate the exhaustivity mechanism. In other words, there exists an integer n such that for any m greater than n, im (S) = in (S). I first recall the relevant definitions: If S is a sentence, ALT(S) is the set of the scalar alternatives of S. Then we define inductively, for any n, and any S, ALTn (S) and in (S), as follows: – i0 (S) is the proposition corresponding to the literal meaning of S. Propositions are sets of worlds. – ALT0 (S) is the set of propositions expressed by the members of ALT(S), i.e. ALT0 (S) = {φ: ∃ S (S ∈ ALT(S) ∧ φ = i0 (S )} For any n, – ALTn (S) = {φ: ∃ S (S ∈ ALT(S) ∧ φ = in (S )} – in+1 (S) = {w : w ∈ (S) ∧ ¬∃w (w ∈ in (S) ∧ w <S,n w)}, with: w <S,n w if the members of ALTn (S) that are true in w are a proper subset of the set of members of ALTn (S) that are true in w . I need one additional definition: For any sentence S, *ALT(S) is the transitive closure of ALT(S), i.e. the set one gets by adding to ALT(S) the alternatives of each member of ALT(S), and then the alternatives of the alternatives of each member of ALT(S), etc. (recall that the relation ‘being an alternative of’ is not, in general, transitive) More formally: *ALT(S) is the smallest set X having the following properties: – ALT(S) ⊆ X – ∀S1 ∀S2 (S1 ∈ X ∧ S2 ∈ ALT(S1 ) → S2 ∈ X) Finiteness condition: S meets the finiteness condition if *ALT(S) is finite.
Aspects of the Pragmatics of Plural Morphology 273
FACT TO BE PROVED: If *ATL(S) is finite, then there exists an integer n such that for any m > n, im (S) = in (S) Proof First note that, for any n, and any m greater that n, im (S) ⊆ in (S). Suppose the fact that has to be proved were false. In this case, for any n, there exists an integer m greater than n such that im (S) = in (S), i.e., given that im (S) ⊆ in (S), im (S) ⊂ in (S). I will show that this leads to a contradiction. Assume now that *ALT(S) is finite ∼S w2 ’ means: Notation: given two worlds w1 and w2 , ‘w1 = ∀S ∈ *ALT(S), w1 ∈ i0 (S ) if and only if w2 ∈ i0 (S ) (in words: the members of *ATL(S) whose literal meanings are true in w1 are the same as those whose literal meanings are true in w2 ). Obviously, ∼ =S is an equivalence relation. For any world w, I note w/S its equivalence class relatively to ∼ =S . Since *ALT(S) is finite, there is a finite number of equivalence classes (since each equivalence class is uniquely determined by the truth value of each member of *ALT(S), there are at most 2m equivalence classes, where m is the cardinality of *ALT(S)). Let us assume the following lemma, which will be proved below: Lemma: For any S in *ATL(S), and any n, in (S ) is the union of a subset of all the equivalence classes. In other words, if a given world w is in in (S ), then its whole equivalence class is included in in (S ), i.e. w/S ⊆ in (S ) Assume that the FACT is false, i.e. that, for any n, there exists an integer m greater than n such that im (S) ⊂ in (S). It follows that there exists an infinite sequence < j1, j2, j3, . . . .> such that for any n, ij(n+1) (S) ⊂ ijn (S). This entails that for any n, there is a world w in ijn (S) which does not belong to ij(n+1) (S). Applying the lemma, we have w/S ∩ ij(n+1) (S) = Ø. It follows that the number of equivalence classes that are included in ij(n+1) (S), which is finite, is smaller than the number of equivalence classes that are included in ijn (S). Therefore, as we move on in the sequence < ij1 (S), . . .,ijn (S), . . .>, we get propositions that contain less and less equivalence classes; since the number of equivalence classes included in ij1 (S) is finite, there must be an integer m such that ijm (S) contains no equivalence class, i.e. (given the lemma) is the empty set.
274
Presupposition and Implicature in Compositional Semantics
But then it is impossible that ij(m+1) ⊂ ijm , which contradicts the hypothesis. Proof of the lemma I will prove that for any S in *ALT(S), and for any n, w ∈ in (S ) if and only if w/S is included in in (S ). In fact, one direction of this biconditional (from right to left) is trivial, so we just have to prove that for any S in *ALT(S), and any n, if w ∈ in (S ), then w/S ⊆ in (S ) Let us note *ALT0 (S) the set of the literal meanings of the member of *ALT(S) (a) suppose n = 0. Let S be a member of *ALT(S). Suppose w ∈ i0 (S ). Let w be such that w ∼ =S w. By hypothesis, w makes true exactly the same members of ∗ ALT0 (S) as w. Since w makes true i0 (S ), w does too, i.e. w ∈ i0 (S ). Therefore all the worlds equivalent to w belongs to i0 (S ), i.e. w/S ⊆ i0 (S ) (b) Inductive reasoning Induction hypothesis: For any S in *ALT(S) and any w, w ∈ in (S ) if and only if w/S ⊆ in (S ). The induction hypothesis entails that, for any two worlds w and w , if w∼ =S w , then they make true exactly the same element of ALT n (S ), for any S . Let S be a member of *ALT(S). Let w1 be a member of in+1 (S ), and w2 be such that w2 ∼ =S w1 . We have to prove that w2 ∈ in+1 (S ). Since in+1 (S ) ⊆ in (S ) (by definition), w1 ∈ in (S ), and, by the induction hypothesis, w2 ∈ in (S). Recall that in+1 (S ) = {w: w ∈ in (S ) ∧ ¬∃w (w ∈ in (S ) ∧ w <S ,n w}, with: w <S ,n w if the set of members of ALTn (S ) true in w is properly included in the set of members of ALTn (S ) that are true in w. It follows from this definition that if two worlds w and w makes true exactly the same members of ALT n (S ), then either both belong to in+1 (S ), or none of them does, for any S . By the induction hypothesis, w2 and w1 make true exactly the same members of ALTn (S ). Therefore, since w1 ∈ in+1 (S ), w2 ∈ in+1 (S ). QED
Notes ∗ Many thanks to Manfred Krifka and Uli Sauerland for useful comments on a first
draft of this chapter. I also benefited from useful discussions with Francis Corblin, Danny Fox, Nathan Klinedinst, Alain Rouveret and Philippe Schlenker. All these
Aspects of the Pragmatics of Plural Morphology 275 people do not necessarily agree with the content of this chapter. All mistakes are mine. 1. It may be hard for some speakers to get a narrow-scope interpretation for some problems here, because some is a positive polarity item (though recent works on positive polarity items do not in fact predict that the narrow-scope reading is blocked in this context – see Szabolcsi, 2004). In any case, the only thing that matters here is that the narrow-scope reading, though maybe unnatural, forces an at least one- interpretation – but see note 2. It is in fact possible to force the narrow-scope reading by adding the focus-sensitive adverb even and focusing some, as in: (a) It is unlikely that Jack solved even SOME difficult problems (b) It is impossible that Jack solved even SOME difficult problems
2.
3. 4. 5. 6. 7.
8.
9.
10.
Unambiguously, some difficult problems, in these two examples, has an at least one-interpretation. Judgments are a little bit more complex. For instance, stressing some may allow a genuine at-least-two-reading. Furthermore, an informant says that a pedantic speaker could pretend that he was right in saying ‘It is impossible that Jack solved some difficult problems’ in case Jack actually solved exactly one problem. Similarly, Only Jack has solved difficult problems is most naturally understood as Nobody except Jack has solved any problem, and Jack has solved several problems. Or, to be more precise, this interpretation is the most straightforward one. I thank Tim Stowell for sharing his judgments with me. As argued independently by Krifka (1989), among others. In Section 3, this hypothesis will be derived from a notion that I term higherorder implicature. Krifka (1989) has already claimed that plural indefinites can be true of atomic individuals and get pragmatically strengthened as a result of a competition with singular indefinites. As Krifka himself points out (p.c.), the difficulty raised by the fact that John has a child and John has children are then literally equivalent is not explicitly addressed in this chapter. A similar view is present in Hoeksema (1983) and van Eijck (1983). The additional step was implicitly present in neo-gricean analyses such as Horn (1972), Levinson (1983) and Gazdar (1979), according to whom scalar implicatures are derived by negating stronger alternatives. Sauerland (2004a) assumes that the strenghtening from ‘the speaker doesn’t have the belief that P’ to ‘the speaker believes that not-P’ is based on an hypothesis according to which the speaker is opinionated, i.e. knows the truth-value of any scalar alternative of a given sentence, unless a prior pragmatic reasoning established that he must be uncertain whether this scalar alternative is true. van Rooij & Schulz (2006) and Spector (2003, 2005) assume that the speaker is supposed to be as informed as possible, given that he has obeyed conversational maxims. In the next section, though, what will count is simply that the exactly oneinterpretation of the singular indefinite article, which is itself an implicature, is absent to begin with in monotone-decreasing contexts. Note, however, that one tends to infer from (8d) that there must be occasions in which Jack sees several horses. Likewise (8a) would not be natural if we
276
11.
12. 13.
14. 15.
16.
17.
Presupposition and Implicature in Compositional Semantics knew for sure that every student who has solved any problem has in fact solved exactly one problem. See Section 5. Things would not be different if we also took into account the alternatives induced by the scale of numerals, i.e. the sentences we get by replacing exactly one by exactly two. In this chapter, propositions are viewed as sets of worlds. The definition that follows, which is close to Krifka’s (1993) semantics for only, is not equivalent to the previous one in the general case, but will suffice for most examples. Van Rooij & Schulz (forthcoming) shows that the Krifkainspired exhaustivity operator fails to make accurate predictions for some sentences in which a disjunction is embedded under another scalar item. I will not consider such cases in this chapter. The Krifka-inspired exhaustivity operator yields the same reading as the ‘official’ one whenever it does not yield the contradictory proposition. I am using here an informal notation: ‘∧ . . .’ is to be read as ‘the proposition that . . . is true’. The existence of generic readings for bare plurals, the phenomenon of quantificational variability, as well as the fact that existential bare plurals must take narrow-scope with respect to other operators, have motivated various proposals in which bare plurals are not treated as quantifiers, but either is introducing variables (Diesing, 1992) or as referring to kinds or properties (see Carlson, 1977; Chierchia, 1998; Dayal, 1999, among many others). Note, however, that such a sentence would be false if all my friends were married to the same wife, though, in this situation, if would be true that my friends each have a beautiful wife (even with a narrow-scope reading for a beautiful wife). Zweig (2005) claims that this overall plurality requirement is itself an implicature, a view that is fully compatible with my general account, as is also the view that dependent plurals are a specific instance of cumulative interpretations, which is argued for by Beck & Sauerland (2000). Francis Corblin (p.c.) suggests that French plural indefinites headed by des are interpreted as number-neutral when they occur in a subordinated discourse condition structure as defined by Discourse Representation Theory (cf. Kamp and Reyle 1993). Such an account predicts number-neutral readings as soon as the plural indefinite is under the scope of an operator that is not a singular indefinite, and therefore captures both dependent plurals and number-neutral readings in decreasing contexts. Such a move cannot easily account for the kind of readings exemplified by (12), and does not account either for the reading of plural indefinites within the scope of a distributive universal quantifier (see Section 4.2). The need for a specific account of dependent plurals is shown by the following contrast (which also holds for the corresponding French sentences): (a) (All) the professors are wearing ties (b) #Every professor is wearing ties
See also Corblin (1987). 18. I use the word group in an informal manner. See below for the formal implementation.
Aspects of the Pragmatics of Plural Morphology 277 19. Though Schwarzschild (1996) does not assume an ontology based on plural individuals, but on sets of individuals. 20. This is the same as Landman’s (2000) *-operator. 21. Sauerland (2003, 2004b) and Sauerland et al. (2005) assume, on the contrary, that the plural morpheme has no semantic import, while the singular morpheme presupposes that the NP to which it applies has a denotation that includes only atomic individual. The end-result is actually the same as here, i.e. the denotation of horses is, in general, the set of all (atomic or nonatomic) individuals whose parts are atomic horses, while that of horse is the set of all atomic horses. See Section 6. The claim that the plural is semantically unmarked is also present in Hoeksema (1983), van Eijk (1983) and Krifka (1989). Even though, due to the particular mode of exposition I have adopted, I assume that the singular morpheme has no semantic import, there is a sense in which the singular morpheme represents the marked item in the singular/plural opposition: namely, singular nouns are semantically more specific than plural nouns. 22. ⊆ represents set-theoretic inclusion; ⊂ represents proper inclusion. 23. Manfred Krifka (p.c.) notes that even though Jack solved difficult problems and Jack solved a difficult problem have the same truth-conditions (according to my account), they are not necessarily equivalent if a richer notion of meaning is adopted, as in dynamic semantics. In DRT (Kamp & Reyle 1993), for instance, Jack solved a difficult problem would be true in all worlds in which Jack solved at least one problem, but a difficult problem would introduce a discourse referent whose denotation is necessarily an atom, while difficult problems would introduce a discourse referent whose denotation can be an atom as well as a plural individual. As a result, Jack solved a difficult problem is semantically more specific, and one could try to derive the at-least-tworeading of plural indefinites from a competition mechanism that would be sensitive to this difference (and thus dispense with higher-order implicatures). Van Roiij & Schulz (forthcoming) exploit dynamic semantics in order to account for the interpretative differences between bare numerals like three and modified numerals like more than two, by means of an exhaustivity operator that is formulated in terms of dynamic semantics. Their specific account, though, would actually not derive the desired result for bare plurals, even on the assumption that bare plurals competes with singular indefinites. Note furthermore that if an indefinite is embedded within the scope of a universal quantifier, it does not introduce a discourse referent at the level of the main DRS, so that Every student solved a difficult problem and Every student solved difficult problems are equivalent even dynamically. Krifka’s suggestion is therefore unable to account for the facts that are the object of Section 4.2 below. More generally, Krifka’s alternative account predicts that at-leasttwo-readings for bare plurals vanish as soon as the bare plural is interpreted within the scope of an operator that is not an indefinite (i.e. within a subordinated DRS), which is I think incorrect. As far as I can see, this suggestion is in the end equivalent to Corblin’s proposal for French des, as described in note 17. 24. Though I do not want to assume that Øpl and a have the same semantic contribution. The existence of a null determiner in bare plurals is of course debatable. Nothing essential in this chapter hinges on this hypothesis, which
278
25. 26. 27.
28.
29.
30. 31.
32.
33. 34.
35. 36.
Presupposition and Implicature in Compositional Semantics is made only for the sake of explicitness. Krifka (1995) argues that the indefinite article and the plural morpheme fulfill the same syntactic function: both fill the number argument of count nouns. As he himself points out (p.c.), this assumption could serve as a motivation for the view that bare plurals and singular indefinites belong to the same scale. I don’t want to commit myself to any specific analysis of the mechanism whereby existential readings are derived for bare plurals. To illustrate: ‘Two students have read this book’ is an alternative of ‘One student-sg has read this book’ In many languages, the singular indefinite determiner and the numeral for one are undistinguishable. For such languages, it might be natural to assume the scale <sg. indef. det, two, three, . . . .>, which would in fact make no difference here. I owe the ideas underlying this formalization to Danny Fox’s suggestion that the recursive application of an exhaustivity operator may account for the freechoice interpretation of disjunction under the scope of possibility modals (Fox, this volume). Contrary to Fox, however, I am not assuming that the exhaustivity operator has any syntactic expression, though I do not exclude it either. In this, I make an informal use of propositional logic notation : ¬S, where S is a proposition (and not a sentence), represents the negation of S, i.e, in set theoretic terms, its complementary set. Likewise, ∧ and ∨ are used to represent intersection and union in case we are dealing with propositions, i.e. sets of words, and not sentences. Furthermore, I often use a sentence to name the proposition that it expresses. I thank Philippe Schlenker for pointing out this issue. Again, ignoring the alternatives that are induced by the presence of a numeral (exactly one). In fact, replacing one by a greater numeral always leads to sentences that contradict the literal meaning of the original sentence, so that the negation of such alternatives is already entailed by the literal meaning of the original sentence. Chierchia (2002), contrary to standard neo-gricean accounts that predict no implicature at all in non-monotonic contexts, offers an analysis according to which a possible strengthened meaning for (24) is Exactly one of my students has solved exactly one difficult problem. The variant with some is to be considered only under the reading where some takes scope below every student. As observed by Uli Sauerland (p.c.), such a claim may be problematic: since the a. reading entails the b. reading, the existence of a genuine ambiguity is hard to show, because there can be no situation in which the a. reading is true and the b. reading is false. However, as explained above, speakers tend to infer (28a) from (27), while they acknowledge at the same time that the truth of (28b) is enough to deem (27) true. Putting aside, of course, cases where a bare plural does not receive an existential interpretation, but a generic interpretation. That is, I ignore the fact that a/an as well belongs to the scale of every, and I also ignore intermediate members of the scale <some, every>, e.g. many. As far as I can see, these simplifications do not in fact affect the analysis, but make the exposition easier.
Aspects of the Pragmatics of Plural Morphology 279 37. Of course, we get exactly the same effects with #Jack has no fathers, vs. Jack has no father and Jack has no brothers. 38. Two sentences are contextually equivalent in a given context if they are true in exactly the same worlds among those that are compatible with common knowledge before the sentence is uttered. 39. A caveat: maybe (46c) is better if it is known that, for some reason, Mary wanted to meet Peter (and, as a result, Jack as well) because she has a specific question to ask him, and not Jack. Likewise for (46d), substituting Jack for Peter and Peter for Jack. 40. Danny Fox (p.c.) argues, on the basis of similar facts, that the computation of scalar implicatures is not a purely pragmatic process, but is governed by a set of rules that are context-independent. 41. Danny Fox (p.c.) made a suggestion that can be reformulated in the following way: when all the members of a given alternative set are contextually equivalent, none of them can be used felicitously. This is apparently falsified by the appropriateness of the variant with and in the Siamese twins context. D. Fox, however, notes that in such a context, Mary met both Peter and Jack is infelicitous, as it suggests that Mary could have met one without the other; he suggests that Peter and Jack can be seen as ambiguous between the quantifier ‘both Peter and Jack’ and a definite description (‘the group made up of Peter and Jack’), in which case the definite description variant would not be part of a scale. I don’t see, however, how to extend such an account to the modal presupposition triggered by plural indefinites. Krifka (p.c.) points out that and in (47) does not represent boolean conjunction, but sum-formation, and that there could be a rule that says that one should ‘maximize’ the sum one talks about. 42. In fact, this presupposition is contradictory as long as there is a group of difficult problems, since in this case there must also exist atomic difficult problems. However, if I understand Sauerland et al.’s proposal correctly, once this presupposition is incorporated into the assertive component below the scope of the existential operator, there is no more contradiction. 43. I argued that such a sentence is interpreted as ‘One student has solved at least two difficult problems, and all other students solved no difficult problem’, while principle B’ predicts that this sentence would be true in a situation in which one student has solved three difficult problems, and all others have solved exactly one difficult problem. 44. Fox (this volume) uses the possibility of applying recursively an exhaustivity operator in the syntax to account for the free-choice interpretation of disjunctions under possibility modals (crucially, he needs another exhaustivity operator than the one used here – I could have used his own operator as well for all the cases mentioned in this chapter).
References Beck, S. & U. Sauerland (2000) ‘Cumulation is needed: reply to Winter (2000)’, Natural Language Semantics 8: 349–71. Carlson, G. (1977) Reference to Kinds in English, PhD thesis, University of Massachusetts.
280
Presupposition and Implicature in Compositional Semantics
Chierchia, G. (1998) ‘Reference to kinds across languages’, Natural Language Semantics 6: 339–405. Chierchia, G. (2002) ‘Scalar implicatures, polarity phenomena, and the syntax/pragmatics interface’, in A. Belletti (ed.), Structures and Beyond. Oxford: Oxford University Press. Corblin, F. (1987) Indéfini, défini et démonstratif: Constructions linguistiques de la référence, Genève, Droz Dayal, V. (1999) ‘Bare NP’s, Reference to Kinds, and Incorporation’, Proceedings of SALT IX Diesing, M. (1992) Indefinites. MIT Press. Eijck, J. van (1983), ‘Discourse representation theory and plurality’, in Alice ter Meulen (ed.), Studies in Model-theoretic Semantics, pp. 85–106. Foris, Dordrecht. Fox, D. (2006) ‘Free-choice and the theory of scalar implicatures’, this volume. Gazdar, G. (1979) Pragmatics. Implicature, Presupposition, and Logical Form. New York: Academic Press. Hoeksema, J. (1983) ‘Plurality and conjunction’, in Alice ter Meulen (ed.), Studies in Model-theoretic Semantics, pp. 63–83. Foris, Dordrecht. Horn, L. (1972) ‘On the semantic properties of logical operators in English’, PhD thesis, UCLA, Los Angeles, CA. Kamp, H. and U. Reyle (1993) From Discourse to Logic. Dordrecht: Kluwer. Krifka, M. (1989) ‘Nominal reference, temporal constitution and quantification in event semantics’, in R. Bartsch et al. (ed.), Semantics and Contextual Expressions, pp. 75–116. Foris. Krifka, M. (1993) ‘Focus, presupposition, and dynamic interpretation’, in K. Bimbó & A. Máté (eds), Proceedings of the Fourth Symposium on Logic and Language. Budapest: Aron Publishers, 31–60. Krifka, M. (1995) ‘Common Nouns: A Contrastive Analysis of English and Chinese’, in G.N. Carlson & F. J. Pelletier (eds), The Generic Book, pp. 398–411. Chicago University Press. Landman (2000) Events and Plurality, Dordrecht: Kluwer. Levinson, S. (1983) Pragmatics, Cambridge: Cambridge University Press. Link, G. (1983) ‘The logical analysis of plural and mass terms: A lattice-theoretical approach’, in R. Bauerle, C. Swharze and A. von Stechow (eds), Meaning, Use and Interpretation of Language, pp. 302–23. Berlin: de Gruyter. Rooij, R. van & K. Schulz (2006) ‘Pragmatic Meaning and Non-monotonic Reasoning: the case of Exhaustive Interpretation’, in Linguistics and Philosophy 29: 205–50. Rooij, R. van & K. Schulz (forthcoming) ‘Only: Meaning and Implicatures’, in M. Aloni et al. (eds), Questions in Dynamic Semantics: Current Research in the Semantics/Pragmatics Interface. Elsevier. Sauerland, U. (2003) ‘A new semantics for number’, in R. Young and Y. Zhou (eds), Proceedings of SALT 13, CLC Publications, Cornell University, Ithaca, N.Y. Sauerland, U. (2004a) ‘Scalar implicatures in complex sentence’, Linguistics and Philosophy 27: 367–91. Sauerland, U. (2004b) ‘A comprehensive semantics for agreement’, Paper presented at the ‘Phi-Workshop, McGill University, Montreal, Canada. August 2004. Sauerland, U., J. Anderssen & K. Yatsushiro, 2005. ‘The plural is semantically unmarked,’ in Stephan Kepser and Marga Reis (eds) Linguistic Evidence: Empirical, Theoretical, and Computational Perspectives, 413–34. Berlin: Mouton de Gruyter.
Aspects of the Pragmatics of Plural Morphology 281 Schwarzschild, R. (1996) Pluralities, Dordrech: Kluwer. Szabolcsi, A. (2004) ‘Positive polarity-negative polarity’, Natural Language and Linguistic Theory 22: 2. Spector, B. (2003) ‘Scalar implicatures: exhaustivity and gricean reasoning’, in B. ten Caten (ed.), Proceedings of the ESSLLI 2003 student session, Vienna. Spector, B. (forthcoming) ‘Scalar implicatures: exhaustivity and Gricean reasoning’ (rev. version of Spector, 2003), to appear in M. Aloni et al. (ed), Questions in Dynamic Semantics, Current Research in the Semantics/Pragmatics Interface, Elsevier. Spector, B. (2006) Aspects de la pragmatique des opérateurs logiques, PhD thesis University of Paris 7. Zweig, E. (2005) ‘The implications of dependent plural readings’, Ms, New York University.
Index accommodation global, 132, 153, 239 intermediate, 131, 134–6, 240 local, 129, 134, 135, 153, 155, 222, 228, 229, 239, 240 presupposition (see also presupposition) adverb of quantification, 125 adverbial, 3, 155, 233, 241 ‘again’, 7, 8, 13, 14–31 presuppositions of, 8 alternative bouletic, 238 epistemic, 35, 42, 58–61, 63 deontic, 37, 53–7, 68 excludable, 98–100, 107, 111 scalar, 3, 247–8, 253–4, 272, 275 set, 92, 107, 114, 116, 254, 272, 279 set of, 56, 72, 75, 79, 83–4, 91, 96–104, 107, 115, 117–18, 247, 262 ‘always’, 133, 178, 180, 184, 186–7, 192, 195, 210, 212 ‘also’, 8, 28–31 ambiguity (see also readings), 16–18, 155, 235, 245, 278 repetitive/restitutive, 16–19 anaphoricity, 18 ‘and’, 75, 90, 216, 218–19, 224, 241 appropriateness, 14–15, 279 Authority Principle, 42–3
Be Articulate, 9, 214, 216, 220–2, 234, 240 Bidirectional Optimality Theory, 166, 170–1, 177 binding, 28, 58, 127, 129, 131–3, 135–6, 140, 142, 151–3 borderline case, 169 bound pronoun problem, 129, 139, 145, 150
choice function approach, 8, 122, 127–9, 136, 138, 150–1, 154, 160 parametrized, 209, 212 competition/competitor, 166, 220–2, 250, 253, 255, 256, 261, 268–70, 277 complex variables, 31 conditionals (see also ‘if-clauses’) anaphoric, 52 conjunction, 4, 36–7, 43, 50, 54, 56, 60, 65–6, 68, 71, 74, 76, 80, 86–8, 90, 97, 105, 111, 115–16, 118, 151, 214–21, 223, 231, 234, 249, 260–2, 279–80 constraint pragmatic, 178 , 191, 232 trapping, 130, 133 context change potential, 214, 216, 218–19, 227, 240 downward entailing (downward monotone)/downward decreasing, 34, 36, 39, 45, 50, 63, 65, 82, 115 non-monotonic, 243, 247, 257, 278 set, 14, 215, 217–19, 239 upward monotone, 36 Cooper storage, 125 counterfactual, 116, 119–20 cumulative interpretation, 276
decomposition, 217 definite description, 151, 153, 160, 239–40, 271, 279 deontic, 35, 42, 56, 65 discourse, 4, 18–19, 21, 30, 32, 46, 52, 56, 66, 125, 147, 231, 276–7 discourse representation theory (DRT), 8, 121–2, 125, 126, 133, 136, 138, 151, 154, 156, 276, 277 282
Index disjunction alternatives of, 90, 93, 96, 104, 112 free-choice (see also free-choice effect) projective behaviour of, 226, 228 scope of, 111 division principle of, 214, 216, 219 domain variant, 187–91, 193–4, 196–8, 200, 203, 207, 212 Donald Duck problem, 126, 129, 139, 141, 146 double negatives, 163–6, 175–7 dynamic semantics, 219, 240, 277 empty set problem, 127–8, 138–9 epistemic, 7, 36–7, 40–3, 49, 53, 59, 60, 63, 65–6, 67, 69 exhaustivity/exhaustification recursive, 7, 80, 102, 109, 118 anti-, 84–5, 116 existential closure, 121, 125–6, 138–9, 152–4, 156, 171 modal, 35, 46, 53–6, 59, 63 partitive use of, 47–9 quasi-specific, 38 specific, 59–60, 67 specific use of, 48–9 free-choice (see also disjunction) effect, 4–6, 11, 34–5, 42, 44–56, 62, 64–9, 71, 104, 111–12 puzzle, 35, 42, 44, 46 French, 210, 233–4, 244, 276, 277 functional wh-questions, 23 generic -ity, 133–4 reading, 276 use of existentials, 48 Gricean scale, 233 Grice’s Maxim conversational, 79 of Manner, 49, 55, 62, 114, 166, 208 of Quality, 49, 55, 73, 77 of Quantity, 72–7, 83, 95, 166, 208, 246–7, 266
283
I-principle, 38, 169, 175 ‘if’-clauses (see also conditionals), 51, 52, 239 indefinite intermediate scope of, 123–4, 129–30 non-specific, 155 plural, 5, 243, 245, 249, 264–5, 267, 270–1, 275–7, 279 scope of, 121, 127, 139 specific, 8–9, 121, 125, 129–32, 153–5, 160 individual-level predicate, 178, 201, 210 inference, 7, 8, 15, 20, 22, 25–6, 30, 40, 60, 72–3, 79, 86, 88, 241, 246 anti-conjunctive, 109–10 free choice, 41, 85, 86, 108 ignorance, 72–5, 77–8, 80, 82, 84–5, 90–5, 97, 102–4, 112, 114, 117 implicature, 170, 172, 206, 212, 243, 246, 253, 275, 276 higher order, 243, 246, 250, 253–4, 256–7, 259, 260–1, 263, 271, 277 Neo-Gricean approach to, 5, 82 scalar, 3, 5, 10, 71–3, 233, 241, 243, 246–8, 250, 253, 256, 258, 271, 275, 279 syntactic approach to, 77
LF, 81, 138, 154, 185–96, 198–206, 208–9, 211–12
M-principle, 170, 175 minimization, 222, 228, 234, 240 of conjunctions, 214, 216, 221 modal, 64, 86 deontic, 42, 53, 55 existential, 5, 35, 71, 80–1, 86, 94, 105, 117 universal, 40, 85, 87, 117 monotone decreasing context, 243, 245–7, 256, 270, 275
284
Index
(natural language) variable, 13 anaphoric, 24–5 bound, 23, 130, 141, 143, 146, 154 covert, 23 definedness condition of, 8, 122, 137–9, 139, 141, 146–7, 150, 154 domain, 12–13, 22, 135, 237 free, 8, 16, 21, 23, 31, 121, 125–6, 141–2, 152 partial, 8–9, 121–2, 136–40, 141, 143, 146–7, 149–51, 153–6, 158 restriction, 9, 12, 121, 138, 141, 142–3, 146–7, 151–3, 155–6, 158 negation, 243–4, 247, 249, 251, 255–7, 260–1, 263, 266, 270, 278 negative concord, 40 neither nor, 40, 64 not un-A, 164–5, 175–6 Novelty Condition, 9, 132, 136, 151, 153 numeral, 78, 236, 276, 277, 278 scale of, 276
‘only’, 71, 78–9, 95, 97, 99, 101, 276 ‘or’, 34–69, 86, 92, 216 contemplative, 61 temporally dynamic, 58 operator (see also quantifier) downward monotonic (downward entailing), 81, 124, 128, 160 exhaustivity, 71, 79, 243, 248–9, 262, 271, 276, 277, 278, 279
P-principle of pragmatics, 184, 187–95, 198–9, 200–1, 203, 206–8, 211 partition, semantic, 179, 201, 205 partitive (see also existential), 47–9, 53 permission, 34, 54, 59, 64, 69, 80–1, 86 plural bare, 243, 249, 251, 256, 264–5, 277–8 dependent, 245, 250, 276, 281
polarity item, 244, 249, 275 possibility epistemic (see also modal, epistemic), 36, 42–3, 53, 59–60, 65–6, 70, 120 presupposition anaphoric, 21, 30 accommodation (see also accommodation), 129, 131, 133, 136, 152, 154 binding, 129, 132, 135–6, 152, 153 maximize, 268–71 modal, 250, 264, 267, 279 projection, 9–10, 21, 24, 151–2, 195, 214, 218, 222, 226, 230 trigger, 129, 151 Projection Problem, 215, 223, 232–3
quality, 46, 68, 76, 90–2, 94, 98, 116, 165 quantifier/quantification (see also operator) adverbial, 178, 188 dependence, 13, 22, 24, 26, 30–1 downward monotonic (downward entailing), 81, 124, 128, 160 existential, 53, 71, 87, 109, 111, 137, 141, 268, 269–70, 279 raising, 125 scope of, 21, 24, 40, 63, 65, 71, 87, 122, 134,136, 137, 146, 218, 245, 250, 257, 259, 261, 269, 270, 276, 277, 279 universal, 37, 39–40, 63, 71, 80, 87–8, 93, 101, 107, 111, 115, 140, 144–5, 156, 160, 223, 225–7, 231, 245, 250, 259, 270, 276, 277 quantificational variability, 208, 276 question, 80, 96
R-principle, 169 Reinhart (1997), 122, 127, 147 reinterpretation, 34, 62–4
Index scales, 114, 253, 256, 262 scope intermediate, 123–4, 129–30 island, 8, 122, 130 of disjunction, 111 of indefinite, 121, 127, 139, 160 scop[e/ing facts] of negation, 6, 36, 69, 71, 87 scop[e/ing of] ‘or’/disjunction (see also disjunction), 66, 69 semantic composition, 50, 52, 181 singular, 6, 106, 122, 251–3, 256, 267, 270, 277 Skolem function, 124, 128–9 speaker opinionated, 76–7, 83, 89–92, 94, 275 specific (see existential), 249, 276, 277 specificity, 8–9, 46, 49, 69, 122, 124–5, 129, 131–2, 134, 136, 154, 156 Stalnaker’s bridge, 14 stereotypical interpretation, 169–70, ‘stop’, 9, 28, 30
Strawson-equivalent, 268 stress ed quantifier, 48 supercover, 44 supervaluation, 174 symmetry condition, 44–5, 64
‘too’, 8, 28–30, 240 transparency, 232, 237–40 triggering problem, 232–3
‘unless’, 216, 226, 229–30
vague ness, epistemic theory of, 168–9, 171 predicates, 174
weak positives, 163 ‘while’, 216, 226, 229–30 ‘wieder’, 18, 27, 32 Williamson (1994), 7, 168
285