Uploaded by Phú Vũ Hữu

understanding-semantics-second-edition - Giáo trình

advertisement
Understanding Semantics
Understanding Semantics, Second Edition, provides an engaging and accessible introduction
to linguistic semantics. The first part takes the reader through a step-by-step guide to the
main phenomena and notions of semantics, covering levels and dimensions of meaning,
ambiguity, meaning and context, logical relations and meaning relations, the basics of noun
semantics, verb semantics and sentence semantics. The second part provides a critical
introduction to the basic notions of the three major theoretical approaches to meaning:
structuralism, cognitive semantics and formal semantics.
Key features include:
∑ A consistent mentalist perspective on meaning
∑ Broad coverage of lexical and sentence semantics, including three new chapters
discussing deixis, NP semantics, presuppositions, verb semantics and frames
∑ Examples from a wider range of languages that include German, Japanese, Spanish and
Russian
∑ Practical exercises on linguistic data
∑ Companion website including all figures and tables from the book, an online dictionary,
answers to the exercises and useful links at routledge.com/cw/loebner.
This book is an essential resource for all undergraduate students studying linguistics.
Sebastian Löbner is Professor for general linguistics at the University of Düsseldorf,
Germany. His main research interests are linguistic semantics and cognitive linguistics.
Understanding Language series
Series Editors:
Bernard Comrie, Max Planck Institute for Evolutionary Anthropology, Leipzig, Germany
Greville Corbett, Surrey Morphology Group, University of Surrey, UK
The Understanding Language series provides approachable, yet authoritative, introductions
to major topics in linguistics. Ideal for students with little or no prior knowledge of linguistics,
each book carefully explains the basics, emphasising understanding of the essential notions
rather than arguing for a particular theoretical position.
Other titles in the series:
Understanding Child Language Acquisition (forthcoming)
Caroline Rowland
Understanding Language Testing
Dan Douglas
Understanding Morphology, Second Edition
Martin Haspelmath and Andrea D. Sims
Understanding Phonetics
Patricia Ashby
Understanding Phonology, Third Edition
Carlos Gussenhoven and Haike Jacobs
Understanding Pragmatics
Jef Verschueren
Understanding Second Language Acquisition
Lourdes Ortega
Understanding Sociolinguistics (forthcoming)
Enam Al-Wer
Understanding Syntax, Third Edition
Maggie Tallerman
For more information on any of these titles, or to order, go to www.routledge.com/linguistics
Understanding
Semantics
Second edition
Sebastian Löbner
First published in Great Britain 2002 by Hodder Arnold, an imprint of Hodder Education
Second edition published 2013
by Routledge
2 Park Square, Milton Park, Abingdon, Oxon OX14 4RN
Simultaneously published in the USA and Canada
by Routledge
711 Third Avenue, New York, NY 10017
Routledge is an imprint of the Taylor & Francis Group, an informa business
© 2002, 2013 Sebastian Löbner
The right of Sebastian Löbner to be identified as author of this work has been asserted by him in accordance
with sections 77 and 78 of the Copyright, Designs and Patents Act 1988.
All rights reserved. No part of this book may be reprinted or reproduced or utilised in any form or by any
electronic, mechanical, or other means, now known or hereafter invented, including photocopying and
recording, or in any information storage or retrieval system, without permission in writing from the publishers.
Trademark notice: Product or corporate names may be trademarks or registered trademarks, and are used only
for identification and explanation without intent to infringe.
British Library Cataloguing in Publication Data
A catalogue record for this book is available from the British Library
Library of Congress Cataloging in Publication Data
Löbner, Sebastian.
Understanding semantics / Sebastian Löbner. -- Second Edition.
pages cm. -- (Understanding language)
Previous edition published under Hodder Education, the second edition is now published under Routledge
after Hodder Education Linguistic titles were acquired in 2012.
Includes bibliographical references and index.
1. Semantics. I. Title.
P325.L567 2013
401’.43--dc23
2012048797
ISBN: 978-0-415-82673-0 (hbk)
ISBN: 978-1-4441-2243-5 (pbk)
ISBN: 978-0-203-52833-4 (ebk)
Typeset in 11 on 12pt Minion
by Phoenix Photosetting, Chatham, Kent
Contents
Preface
Acknowledgements
ix
xi
1 Meaning and semantics
1
1.1 Levels of meaning
1.2 Sentence meaning and compositionality
1.3 Semantics: its scope and limits
Exercises
Further reading
2 Dimensions of meaning
2.1 Meanings are concepts
2.2 Descriptive meaning
2.3 Meaning and social interaction: the dimension of social meaning
2.4 Meaning and subjectivity: the dimension of expressive meaning
2.5 Connotations
2.6 Dimensions of meaning
Exercises
Further reading
3 Ambiguity
3.1 Lexemes
3.2 Lexical ambiguity
3.3 Compositional ambiguity
3.4 Contextual ambiguity
3.5 Meaning shifts and polysemy
Exercises
Further reading
4 Meaning and context
Part 1: Deixis
4.1 Person deixis
4.2 Demonstratives and place deixis
4.3 Time deixis
Part 2: Determination
4.4 Definiteness and indefiniteness
4.5 Quantification
1
7
14
16
17
18
18
21
28
33
36
37
39
40
41
41
44
48
49
57
59
60
62
62
63
70
72
74
74
83
Contents
vi
4.6 Generic NPs
Part 3: Presuppositions
4.7 Presuppositions
4.8 Summary
Exercises
Further reading
5 Predication
5.1 Predications contained in a sentence
5.2 Predicate terms and argument terms, predicates and arguments
5.3 Verbs
5.4 Nouns and adjectives
5.5 Predicate logic notation
5.6 Thematic roles
5.7 Selectional restrictions
5.8 Summary
Exercises
Further reading
6 Verbs
6.1 Argument structure, diatheses and alternations
6.2 Situation structure
6.3 Aspect
6.4 Tense
6.5 Selected tense and aspect systems
6.6 Concluding remark
Exercises
Further reading
7 Meaning and logic
7.1 Logical basics
7.2 Logical properties of sentences
7.3 Logical relations between sentences
7.4 Sentential logic
7.5 Logical relations between words
7.6 Logic and meaning
7.7 Classical logic and presuppositions
Exercises
Further reading
8 Meaning relations
8.1
8.2
8.3
8.4
Synonymy
Hyponymy
Oppositions
Lexical fields
90
94
94
101
102
104
106
106
108
111
115
120
122
125
130
131
133
134
135
140
150
157
163
164
165
166
167
167
172
175
184
187
191
197
201
202
203
203
205
208
215
Contents
Exercises
Further reading
9 Meaning components
9.1 The structuralist approach
9.2 Applying the structuralist approach to meaning
9.3 Semantic features
9.4 Semantic formulae
9.5 Semantic primes: Wierzbicka’s Natural Semantic Metalanguage
9.6 Summary and evaluation of the approaches to decomposition
Exercises
Further reading
10 Meaning and language comparison
10.1 Translation problems
10.2 Headache, international
10.3 Relativism and universalism
10.4 Berlin and Kay’s investigation of colour terms
10.5 Consequences
Exercises
Further reading
11 Meaning and cognition
11.1 Categories and concepts
11.2 Prototype theory
11.3 The hierarchical organization of categories
11.4 Challenges to prototype theory
11.5 Semantics and prototype theory
11.6 Semantic knowledge
11.7 Summary
Exercises
Further reading
12 Frames
12.1 Barsalou frames
12.2 Verbs and frames
12.3 Nouns and frames
12.4 Frames and composition
12.5 Frames and cognition
12.6 Conclusion
Exercises
Further reading
13 Formal semantics
13.1 Japanese numerals: a simple example of a compositional analysis
vii
219
220
221
223
226
228
237
242
245
246
246
248
248
251
256
258
262
263
264
265
266
267
276
279
284
292
298
300
300
301
301
311
313
319
321
322
323
324
325
325
Contents
viii
13.2 A small fragment of English
13.3 Model-theoretic semantics
13.4 Possible-world semantics
13.5 The scope and limits of possible-world semantics
Exercises
Further reading
330
342
352
358
362
363
References
364
Index
369
Preface
As for many others, my entrance into semantics was formal semantics, about
forty years ago. At present, this is still the standard approach in many linguistics
departments around the world. Working in the field of semantics my entire academic
life, my conviction has grown that formal semantics is not the ideal framework for
working one’s way into the rich and fascinating reality of natural language meaning.
The perspectives allowed by the formal apparatus developed in formal semantics are
far too restrictive. And the aspects of meaning and the semantic phenomena that are
neglected, or are simply problematic to deal with, are far too numerous. Above all,
formal semantics has little to say about lexical meaning – which, after all, provides
the ultimate basis of all linguistic meaning – and, not by chance, it fails to connect
semantic theory to cognition.
In Understanding Semantics, I have taken a different approach. It is driven by
the idea that students of semantics should first grasp the level of meaning which
linguistic semantics aims to describe and how this level is related to higher levels
of interpretation; they should learn that there are different dimensions of meaning,
in addition to descriptive meaning; they should know about ambiguity and about
the existence of meaning shifts that interfere with lexical meaning; they should get
a notion of the rich inventory of indexical means of expression including deixis,
determination and presupposition carriers; they should learn the basics of lexical
semantics of nouns and verbs; they should know that there are different theoretical
approaches to meaning; and they should get a notion of the fact that linguistic
meaning is ultimately a matter of conceptualizing the things we talk about: when we
put things into words, we are not just giving a one-to-one mapping of what the world
is like – we make a choice by putting things in the particular way we do. Meaning is
not just a matter of logical relations and truth conditions. As to sentential meaning,
the students need to know about the basic semantic functions of NP determination
and the verbal categories of aspect and tense, and they should know the basics of
predication. All this should be discussed from a perspective that also takes a look at
other languages. On this complex background, the more advanced students may start
to work their arduous way into the theory and technicalities of formal semantics.
In order to give an idea of the basic notions of this approach, the book offers a
substantial basic introduction in the last chapter, and a critique.
The second edition of Understanding Semantics is not only a more up-to-date
version of the first edition, but is supplied with new sections that considerably
broaden the coverage of the field. These include:
x
Preface
∑ basic notions of the semantics of word formation (chapters 1 and 12)
∑ deixis and demonstratives (chapter 4)
∑ presuppositions (chapters 4 and 7)
∑ NP semantics (chapter 4)
∑ verb semantics including voice, aspect and tense (chapter 6)
∑ Barsalou frames (chapter 12).
The book is accompanied by a website that provides additional support for
students and instructors (http://www.routledge.com/cw/loebner). Along with a
number of additional minor features, the webpage provides
∑ a checklist of key notions for each chapter, interlinked with
∑ a glossary of all technical terms
∑ pdf versions of all figures and tables for your use in teaching, presentations, term
papers, etc.
∑ solutions to the exercises (instructors only).
Acknowledgements
Among the many people that directly or indirectly were of help in writing this book
and preparing the second edition, I want to express my gratitude to Daniel Hole,
Berlin, who took the trouble of commenting in detail on substantial new parts of
the second edition. I would also like to thank Rachel Daw and Nadia Seemungal at
Routledge for their kind and competent support and guidance for the second edition.
This page intentionally left blank
1
Meaning and semantics
Semantics is the part of linguistics that is concerned with meaning. This is, of course,
a very superficial definition. The crucial term ‘meaning’ on which the definition
rests has several different readings – a first semantic observation which you will
find at the beginning of almost every textbook on semantics. Among the many
uses of the notion ‘meaning’, only some belong to the field of linguistic semantics.
Meaning is always the meaning of something. Words have meanings, as do phrases
and sentences. But deeds may have meaning too. If a government pursues a certain
policy, we may ask what the meaning is of doing so. The ‘meaning’ of an action or a
policy is what sense it makes or what purpose it serves or what it is good for. More
generally, we apply the notion of meaning to all sorts of phenomena that we try to
make sense of.
The first thing to be stated is that linguistic semantics is exclusively concerned with
the meanings of linguistic expressions such as words, phrases, grammatical forms
and sentences, but not with the meanings of actions or phenomena. We will approach
the problem of linguistic meaning step by step, to arrive at a more precise definition
of semantics at the end of this chapter. A more concrete idea of what semantics is
about will result when you learn about the many facets of this fascinating discipline
in the course of this book.
1.1 LEVELS OF MEANING
Even if we restrict the study of meaning to words and sentences, the notion of
meaning has to be further broken down into different levels at which we interpret
words and sentences.
1.1.1 Expression meaning
Let us get started by looking at a simple example that will illustrate what semantics
is about.
(1)
I don’t need your bicycle.
This is an ordinary English sentence. Without even noticing, you have already
recognized it as such, you have interpreted it and you are probably imagining
a situation where you would say it or someone would say it to you. Since you
understand the sentence, you know what it means. But knowing what the sentence
2
Understanding semantics
means is one thing, describing its meaning is another. The situation is similar with
almost all our knowledge. We may exactly know how to get from one place to another,
yet be unable to tell the way to someone else. We may be able to sing a song by heart,
but unable to describe its melody. We are able to recognize tens of thousands of words
when we hear them. But the knowledge that enables us to do so is unconscious.
Uncovering the knowledge of the meanings of words and sentences and revealing its
nature is the central objective of semantics.
Let us now try to determine the meaning of the sentence in (1). We start from the
meanings of the words it contains. The main verb in a sentence occupies a key role. So,
1
what is the meaning of the verb need ? Actually, there are two verbs need: an auxiliary
verb (as in I need not go) and a full verb. In (1) we have the full verb. It is used with
2
a direct object (your bicycle) and roughly means ›require‹. We ‘need’ something if
it is necessary or very important for us. In (1), what is needed is described by an
expression composed of the possessive pronoun your and the noun bicycle. The noun
means some sort of vehicle, usually with two wheels and without a motor.
The words need and bicycle are the main carriers of information in the sentence,
so-called content words. The meanings of most content words are very differentiated
because there are thousands of the same kind. All the other elements in our sentence
are different in that they represent items from a very limited choice of expressions of
the same kind. Such words are called function words and include articles, pronouns,
prepositions, conjunctions and other ‘small’ words. We will examine these elements
one by one.
The subject expression I is one of seven personal pronouns in English (I, you, he, she,
it, we and they). What is the meaning of I ? If Mary says the sentence in (1), it is Mary
who is said not to need the bicycle. If John says (1), it is John. In other words, I is used
for the one who says it; more technically: for the one who produces an occurrence of
this pronoun. The technical term for using an expression for something is reference.
When people use I, they refer to themselves. The entity referred to by an expression
is called its referent. The meaning of the pronoun can thus be described as follows:
I indicates reference to the speaker. Similarly, the pronoun you indicates reference to
the addressee or the addressees.
For each personal pronoun there is a corresponding possessive pronoun: I–my,
you–your, etc. Your in (1) indicates that the bicycle referred to is linked to the
addressee(s). For such a link, there is a broad variety of relations possible. Possession
in the sense of ownership is only one option: the expression your bicycle may also
refer to the bicycle the addressee is just riding or cleaning or repairing, or even the
3
bicycle they have been talking about for the last ten minutes. The meaning of your
can roughly be described as ›linked to the addressee(s)‹.
1 It is common practice in linguistic texts to mark words which are referred to in a sentence, rather
than just used, by using italics. In addition, I use italics for emphasis. Whether a word is referred
to or used emphatically is always clear from context.
2 › … ‹ quotes are used for meanings and concepts.
3 I use they as a gender-neutral 3rd person singular pronoun.
Meaning and semantics
3
The form don’t is a contraction of the auxiliary verb do and the negation particle
not. Don’t contributes two things to the meaning of the sentence. It negates the verb
need and thereby turns its meaning into the contrary. In addition, the form don’t
contributes present tense. Tense is the indication that the sentence refers to a certain
time, e.g. present, past or future time. The actual time referred to depends on when
the sentence is uttered. Due to the present tense in (1), we will by default relate the
situation described to the ‘present’ time, i.e. the time when the sentence is being
uttered. Combining these two components of don’t, we may say: the meaning of don’t
is an indication of reference to the time when the sentence is uttered and it turns the
situation expressed by the main verb into the contrary.
So far this has been an attempt to determine the meaning of each word in the
sentence I don’t need your bicycle. This is typical of the work of a semanticist.
As you will have noticed, it is not trivial. For a content word, the description of
its meaning must be specific enough to distinguish it from all words with other
meanings. It would not suffice to describe the meaning of bicycle merely as ›vehicle
with two wheels‹ because there are other kinds of vehicles with two wheels, such as
motorcycles, kick scooters or kids’ balance bicycles. At the same time, the description
must be general enough to cover all cases in which this word could be used. Since
one usually imagines a particular context when one tries to think of a word and its
meaning, one tends to take the meaning too specifically, disregarding other cases in
which the word can also be used.
As for function words like pronouns and auxiliaries and for grammatical forms
such as present tense, their meanings may at first view seem elusive. But it is possible
to account for them too, as our little discussion may have illustrated.
If we put all the pieces together, we can describe the meaning of the sentence as a
whole. It can be roughly formulated as: ›for the speaker, the two-wheeled vehicle of
the addressee(s) is not required at the time when this is being uttered‹.
It is very important to realize that the sentence leaves open who the speaker and
the addressee(s) are, what particular time is referred to and which bicycle. This is not
part of its meaning. Such questions can only be settled if the sentence is actually used
on a concrete occasion. What is, however, determined by the meaning of the sentence
is how the answers to these questions depend on the occasion when the sentence is
used. First, if it is actually used, it is necessarily used by someone who produces the
sentence. With I in subject position, the sentence ‘tells’ us that it is the speaker who
does not need the bicycle. The use of I functions like an instruction: find out who
produced this sentence, this is the referent of I. Second, the use of your presupposes
that there are one or more addressees. The sentence describes the bicycle as related
to them. Third, if a sentence is uttered, it is necessarily used at a certain time. The
time of utterance serves as the reference time for determining what is present, past or
future. The present tense part of the meaning of the sentence conveys the instruction:
attribute the situation described to the time when the sentence is said. Thus the
meaning of the sentence specifies the way in which its reference is determined if and
when it is used at some occasion.
The meanings of words, phrases and sentences, taken out of any particular context
constitute the level of meaning which will henceforth be called expression meaning.
4
Understanding semantics
Expression is just a cover term for words, phrases and sentences. The term expression
meaning covers in particular word meaning and sentence meaning. As you have
noticed, the determination of expression meaning requires an abstraction from the
use of the expressions in concrete contexts. In this sense, the notion of expression
meaning itself is an abstraction and a theoretical construct. But it is justified in the
way language is conceptualized not only in linguistics but also in common thinking:
we do talk about the meanings of words and complex expressions as such, i.e. we do
address this level of meaning.
1.1.2 Utterance meaning
Let us now examine what happens when the sentence in (1) is actually used. We will
consider two scenarios.
SCENARIO 1
1 August 2012, morning. Mary has been planning a trip to town that
afternoon. Two days before, she talked with her neighbour John about the trip
and asked him to lend her his bicycle. She had lent her car to her daughter and
did not know if she would get it back in time. Meanwhile her daughter is back
and has returned Mary’s car. Mary is talking with John on her mobile, telling
him:
I don’t need your bicycle.
Used in this context, the sentence receives a concrete interpretation. References are
fixed: the personal pronoun I refers to Mary, the possessive pronoun your links the
bicycle to her neighbour John and the time reference is fixed, too: in the given context,
the present tense verb will be taken to refer to the afternoon of 1 August 2012. This is
clear from the fact that Mary could have said: I don’t need your bicycle this afternoon,
without changing the meaning of her utterance. Furthermore, the reference of the
grammatical object your bicycle is fixed: it is the bicycle Mary asked John to lend her
two days before.
This is a different level of meaning, called utterance meaning. It comes about
when a sentence with its expression meaning is actually used in a concrete context
and all references get fixed. When this happens, another central notion comes into
play, the notion of truth. If Mary says (1) in scenario 1, the sentence is true. But
in a slightly different scenario it might be false. As long as the sentence (1) is not
actually used with concrete reference, it fails to be true or false. The question of
truth primarily concerns ‘declarative’ sentences such as the one under review. Only
such sentences, when uttered, are true or false. But it matters also for interrogative
and other types of sentences. For example, if John asked Mary Do you need my
bicycle?, the use of the question form would convey that he wants to know from his
addressee whether it is true or false.
Let us now imagine a different scenario:
Meaning and semantics
5
SCENARIO 2
Same time and place. John’s five-year-old daughter Maggie is playing at her
place with her friend Titus. They are playing with a game of cards that display
all kinds of vehicles. Titus holds a card that shows a snowmobile. Maggie
is eager to exchange this card for one of hers and offers Titus a card with a
bicycle. Titus rejects the exchange:
I don’t need your bicycle.
In this scenario, references of I, your and the present tense are fixed accordingly. What
is interesting is that in such a context the word bicycle can be naturally interpreted
as referring not to a real bicycle but to a card carrying the picture of a bicycle. Are
we to conclude that the lexical meaning of bicycle must be taken as covering not only
real bicycles but also pictures of this kind of vehicle and things that display such a
picture? The answer is ‘No’. The word bicycle literally means real bicycles, but when
used in special contexts it can also mean ›picture of a bicycle‹, ›card with a picture
of a bicycle‹, ›toy bicycle‹, ›replica of a bicycle‹, etc. or also ›someone riding on a
bicycle‹ in utterances like ‘Stop, there’s a bicycle coming!’ This, however, is a matter of
utterance meaning. What happens in such cases is that the lexical meaning is shifted
for obtaining an utterance meaning that fits into the given context. Such shifts are
quite common; there are many shifting-patterns at our disposal.
For a general definition of utterance meaning, we need a notion for what was
called ‘occasion’, ‘context’ or ‘scenario’ above. The technical term for this is context of
utterance. The context of utterance, CoU for short, is the sum of circumstances that
bear on reference and truth.
DEFINITION 1 Context of utterance
The context of utterance (CoU) comprises the following aspects of the
situation in which an utterance is produced:
∑ the speaker (or producer) of the utterance
∑ the addressee(s) (or recipient(s)) of the utterance
∑ the time at which the utterance is produced and/or received
∑ the place where the utterance is produced and/or received
∑ the facts given when the utterance is produced and/or received
We have seen in connection with (1), how utterance meaning may depend on who
the speaker and addressees of an utterance are and at which time it is produced. The
place where an utterance is made matters for the reference of expressions such as
here, there, upstairs, downtown, etc. as well as for the truth of sentences like It’s raining.
Facts matter principally for truth as well as for reference. For example, Mary can only
refer to John’s bicycle in such CoUs where a certain bicycle is related to John. CoUs
may be real or fictitious. If we read a work of fiction or watch a movie, the relevant
facts and figures are those of the story.
6
Understanding semantics
Against this background, utterance meaning can be defined as the meaning that
results from an expression being used and interpreted in a given CoU. Utterance
meaning derives from expression meaning on the basis of the particulars provided
by the CoU.
The notion of utterance meaning does not include all that an addressee may
make of an utterance in a particular CoU. Addressees usually make all kinds of
inferences. For example, in scenario 1, John may infer that Mary is still planning to
make the trip since otherwise she would have told him; that she would have asked
him to lend her his bicycle if she could not have used her car; that, however, her
daughter is back with the car and that Mary is not going to lend her the car again
on that afternoon; that Mary will take the car for her trip; that she considers herself
able to drive, etc. All this is not explicitly said with that sentence, and it need not be
true under different circumstances. In the given scenario, these inferences can be
considered communicated because Mary can rely upon John’s understanding them.
Although these inferences are triggered in the addressee’s mind by the utterance of
the sentence, it is important to separate what is actually being said from what is only
inferred. The investigation of such inferences, their role in communication and how
they are related to the utterance meaning of what is actually said, is an important
part of pragmatics, the scientific study of the rules that govern the use of language.
Within pragmatics, Grice’s theory of ‘conversational implicatures’ and Relevance
Theory by Sperber and Wilson deal with inferences of this kind.
1.1.3 Communicative meaning
Neither the level of expression meaning nor that of utterance meaning is the final and
crucial level of interpretation. In an actual exchange, our main concern inevitably is
this: what does the speaker intend – in particular, what does the speaker want from
me? Conversely, when we say something, we choose our words in pursuit of a certain
communicational intention. Verbal exchanges are a very important form of social
interaction. They will always be interpreted as part of the whole social exchange and
relationship entertained with the speaker.
One and the same sentence can be uttered with quite different communicative
results. The utterance of (1) in scenario 1 will be taken as a withdrawal of a former
request. In scenario 2, the utterance of the same sentence constitutes the rejection of
an offer. In other CoUs, uttering the sentence could serve still other communicative
ends. A theory that addresses this level of interpretation is speech act theory,
introduced in the 1950s by the philosopher John L. Austin (1911–60) and developed
further by others, in particular John R. Searle. The central idea of speech act theory
is that whenever we make an utterance in a verbal exchange we act on several levels.
One level is what Austin calls the ‘locutionary act’, defined as the act of saying an
expression with a certain utterance meaning in the given CoU. In doing so, we
also perform an ‘illocutionary act’, i.e. a certain type of ‘speech act’: a statement,
a question, a request, a promise, an offer, a refusal, a confirmation, a warning, etc.
When Titus in scenario 2 says I don’t need your bicycle, he performs the locutionary
Meaning and semantics
7
act of saying that he doesn’t need Maggie’s card with the bicycle and the illocutionary
act of rejecting her offer. The speech act level of interpretation will be referred to as
communicative meaning.
The three levels of interpretation are connected as follows. Expression meaning is the
level of interpretation which results if the only information we use is the mere linguistic
material. Expression meaning forms the basis for utterance meaning, but does not
determine it. For, as we could see, a sentence with its fixed expression meaning will take
on different utterance meanings if it is used in a particular context. Utterance meaning,
in turn, forms the basis of communicative meaning, without, again, determining it. For
utterances with the same utterance meaning can serve the performance of different
types of speech acts, depending on the ongoing social interaction. Table 1.1 gives a
survey of the three levels of meaning and how they are defined.
Table 1.1
Three levels of meaning
Level of meaning
Definition
expression meaning
the meaning of a simple or complex expression taken in isolation
utterance meaning
the meaning of an expression when used in a given context of utterance
resulting from fixing reference
communicative meaning
the meaning of an utterance as a communicative act in a given social
setting
1.2 SENTENCE MEANING AND COMPOSITIONALITY
1.2.1 Lexical vs compositional meaning
We will now take a closer look at sentence meaning. It is a trivial fact that the
meanings of words and sentences differ in one important point. Meanings of words
must simply be known and therefore learned. In our minds, we host a huge lexicon
where all the words we know and their meanings are stored and at our disposition.
Stored meanings are therefore called lexical meanings.
Words can be combined into sentences. We are usually able to understand the
expression meaning of a sentence without any conscious effort. Nevertheless, this
ability is based on complex cognitive processes which take place in our minds
automatically and unconsciously. The process by which we calculate the meaning
of a sentence is called composition, and the resulting meaning is known as
compositional meaning. In some cases, sentences may have lexical meaning, for
example proverbs such as The early bird catches the worm. This does not mean that
their meanings are merely non-compositional. Rather, such sentences have a regular
compositional non-proverbial meaning plus a special meaning which we have to
learn and store in our lexicon.
Understanding semantics
8
Words can also be used to create new words. This is called word formation,
and its products complex words. For example, you could coin the word joggable,
deriving it from the verb jog by adding the suffix -able, which can be added to a very
large number of verbs. Or you could form the compound carp food. The products of
word formation need not all be stored in the lexicon and yet they can be as easily
and straightforwardly understood as sentences. Obviously, the semantics part of the
brain also knows how to interpret new words formed out of stored lexical material.
Crucially, this holds only for products of word formation that follow general patterns.
There are, on the other hand, very many complex words that carry irregular or
special meanings. Such cases have to be stored in the lexicon in order to receive their
proper interpretation. We’ll return to the semantics of word formation in 1.2.5. For
now, I just want to fix in your mind a terminological convention: the notion of ‘lexical
meaning’ is also to be taken to subsume the meanings of complex words, whether
semantically regular or irregular. This is in line with the ‘dynamic’ view of the lexicon
which sees it not just as a store of entries, but also as comprising components for the
formation of new words and their meanings.
Let us now take a closer look at the composition of sentence meaning in order to
see what is involved.
1.2.2 Grammatical meaning
For a change, we will consider a new example.
(2)
The dog ate the yellow socks.
Let us assume that we have assessed the lexical meanings of the words in (2): the, dog,
eat, yellow and socks. There are no larger units in the sentence with lexical meaning;
the rest of the interpretation is composition. The words in (2) occur in particular
grammatical forms. The verb form ate is past tense – more precisely: simple past
tense rather than progressive (was eating); it is in the indicative mood rather than
in the conditional (would eat), it is active rather than passive (was eaten), it is not
negated (did not eat). The noun socks is plural; dog is singular, by absence of plural -s.
The adjective yellow is neither comparative (yellower) nor superlative (yellowest), but
is given in its basic form called ‘positive’. The forms of the words matter directly for
their meaning, and consequently for the meaning of the whole sentence. The singular
noun dog has a different meaning from the plural noun dogs: dog refers to a single
creature of this kind, and dogs to more than one. Likewise, the meaning of present
tense eat(s) is not the same as that of past tense ate, and the meaning of the simple
form ate is different from the meaning of the progressive form was/were eating. The
meaning of the simple past form with verbs such as eat is the combination of past
tense and perfective aspect. (Tense and aspect will be discussed in chapter 6 on
verbs.)
The meanings of different word forms are not all stored in our lexicon. Rather the
lexicon contains the meaning (or meanings, cf. chapter 3) of only one form of the
word. For a verb, this is a tenseless and aspectless active meaning, for a noun it is
Meaning and semantics
9
4
its singular meaning and for an adjective its positive meaning. The meanings of
particular word forms are derived from the basic lexical meanings by general rules.
These are part of the apparatus we use in composition. We will call the meaning of
grammatical forms grammatical meaning. The meaning of a word in a particular
form is the combination of its basic lexical meaning and the grammatical meaning
of its form.
Word forms are often built by adding an affix to the invariable part of the word, its
stem. Affixes appended at the end of the word are called suffixes, those added at the
beginning are prefixes. It must be noted that not all differences in the grammatical
forms of words matter for their meaning. A certain form may be necessary just for
grammatical reasons. For example, a certain verb form may be obligatory depending
on the subject of the sentence (cf. I am vs she is, she sees vs you see). By contrast, the
grammatical number (singular or plural) of the nouns dog and sock in (2) can be
chosen independently of the construction in which they occur. Therefore, differences
in form only matter for meaning if – in a given construction – different forms can be
chosen freely. In addition, the different forms available must have different meanings.
If, for example, a verb has two different admissible past tense forms like spoiled and
spoilt, the choice of the form would not matter for its meaning. Also, it may be that
one form has a neutral meaning which subsumes the meanings of the alternative
forms. For example, in Korean nouns can take a plural suffix; it is, however, not
necessary to make use of this option in order to refer to a multitude of instances.
Therefore, the basic form of the noun without a plural suffix has a neutral meaning
and does not carry grammatical meaning.
We can sum up these considerations in the following definition of grammatical
meaning:
DEFINITION 2 Grammatical meaning
The form of a word carries grammatical meaning, if
(i) in the given construction alternative forms are admissible
(ii) different forms yield different meanings
(iii) the form chosen does not have a neutral meaning.
Languages differ considerably in what meanings they encode in grammatical forms.
Many languages do not obligatorily mark number on all nouns; others do not have
5
non-positive forms of adjectives or lack tense.
4 In some cases, certain forms of words may have a special lexical meaning, such as glasses as
opposed to glass, or used to as opposed to to use. Some nouns are only used in the plural (trousers);
for certain others the distinction does not matter (logic, logics).
5 See Corbett (2000) for a survey on systems of grammatical number including missing number;
Dryer (2011) offers an online map of forms of number marking for more than 1000 languages
(http://wals.info/chapter/33). For tenseless languages see Comrie (1985: 2.5).
10
Understanding semantics
1.2.3 Syntactic structure
As the last step of composition, the meanings of the words in their given forms are
combined into a whole, the meaning of the sentence. This process is guided by the
syntactic structure of the sentence (this is, for the most part, what grammar is good
for: to guide the interpretation of complex expressions). Let us first determine which
words in (2) belong together. The words the dog form a syntactic unit, a constituent
in syntactic terminology. This kind of constituent is called a noun phrase, NP
for short. The words the yellow socks form another NP; it contains an adjective
in addition to the article and the noun. Actually, the adjective and the noun form
another constituent within the NP. The combination of words into larger syntactic
constituents is governed by the rules of grammar. There is a rule for combining
adjectives with nouns, and another rule for combining a noun, or an adjective-noun
combination, with an article (the article comes first).
Given such rules for forming larger syntactic units we need corresponding
composition rules, in this case:
∑ a rule for deriving the meaning of an adjective-noun combination (yellow socks)
from the meaning of the adjective and the meaning of the noun
∑ a rule for deriving the meaning of an article with a noun (the + dog) or a noun
with adjective(s) (the + yellow socks) from the meaning of the article and the
meaning of the noun, or the noun with an adjective
We will not try to specify such rules now; they will be dealt with in chapters 4, 5
and 13.
Having assessed the dog and the yellow socks as complex constituents, we turn
to the total structure of the sentence. It consists of these two NPs and the verb ate.
These three parts are grammatically related as follows: the verb is the predicate of the
sentence, the NP the dog is its subject and the yellow socks its direct object. The verb
and the direct object form a constituent, known as verb phrase, or VP, which is then
combined with the subject NP to form the complete sentence. We therefore need two
more composition rules:
∑ a rule for deriving the meaning of a VP (ate + the yellow socks) from the meaning
of the verb and the meaning of the direct object NP
∑ a rule for deriving the meaning of a sentence (the dog + ate the yellow socks) from
the meaning of the subject NP and the meaning of the VP
Again, these rules are not trivial. Roughly speaking, the composition works as follows:
the verb eat in its given ‘active’ form means an event which necessarily involves two
elements, an eater and something that is eaten; the direct object NP contributes a
description of the object that is eaten, and the subject NP a description of the eater.
1.2.4 The Principle of Compositionality
Let us sum up the general results we can draw from this example. The syntactic
rules of a language allow the formation of complex expressions from what will be
Meaning and semantics
11
Figure 1.1
The process of composition
GRAMMAR
SEMANTICS
complex
expression
compositional meaning
of the complex
combine expressions
compose meanings
basic expressions
in a particular form
compositional meanings
of the basic expressions
in their given forms
choose grammatical
forms
add grammatical
meaning
basic expressions
lexical meanings
of basic expressions
called basic expressions. (Basic expressions are expressions with a lexical meaning.)
The meaning of complex expressions is determined by semantic composition. This
mechanism draws on three sources:
1 The lexical meanings of the basic expressions
2 The grammatical forms of the basic expressions
3 The syntactic structure of the complex expression
The general scheme in Fig. 1.1 shows that semantic composition is thought of as a
so-called bottom-up process: it proceeds from the basic units to the complex ones.
The lexical meanings of the smallest units serve as input for the rules of grammatical
meaning, whose output is the input for the composition rules. The converse of a
bottom-up process is a top-down process. If semantic interpretation were a topdown process, this would mean that the meanings of words are derived from the
meanings of sentences.
That complex expressions receive their meaning by the process of composition is
the central idea underlying semantics. This is called the Principle of Compositionality.
The principle is attributed to the German philosopher, logician and mathematician
Gottlob Frege (1845–1925), and sometime called Frege’s Principle. Although he
obviously applied the principle, there is no passage in his publications that could
serve as a quotation.
12
Understanding semantics
DEFINITION 3 Principle of Compositionality
The meaning of a complex expression is determined by the lexical meanings
of its components, their grammatical meanings and the syntactic structure of
the whole.
The principle implies that the meanings of complex expressions are fully determined
by the three sources mentioned, i.e. by the linguistic input alone. The principle,
therefore, does not apply to utterance meaning and communicative meaning because
at these levels non-linguistic context knowledge comes into play.
1.2.5 The semantics of word formation
Just like sentence formation, word formation has two sides: a form side and a meaning
side. Along with the formation of grammatical word forms, the form side of word
formation belongs to the domain of morphology, the area of linguistics concerned with
the forms of words. The two main mechanisms of word formation are derivation and
compounding. In the case of derivation, a given word is modified, usually by adding
an affix. Examples are the adjective drink-able, the noun drink-er or the verb en-slave.
There is also conversion – the use of a word in a different word class without changing
its form, for example the noun drink converted from the verb drink or the verb bottle
converted from the noun bottle. In English, conversion is very frequent. The other major
mechanism, compounding, is the combination of two or more words into one. Some
examples in this chapter are word meaning, content word, speech act and composition
rule. The products of word formation are called complex words, including the products
of conversion, even though nothing is added to the word form.
The semantics of word formation is a field that is much less clear than sentence
semantics. There certainly are patterns, however: certain conversions, derivations
and compounds are regularly interpreted in a particular way. Let me just give four
examples:
∑ suffixing a verb V with -able produces an adjective meaning ›one can V it‹: drinkable, read-able, compar-able, deriv-able
∑ suffixing a verb V with -er produces a noun meaning ›someone who Vs‹ : driv-er,
los-er, speak-er, bak-er
∑ prefixing a verb V with un- produces a verb which denotes the reversal of what is
done by V-ing: un-do, un-zip, un-plug, un-tie
∑ prefixing a verb V with out produces a verb which means ›doing V better, or
more‹: out-run, out-play, out-number, out-weigh
Such patterns can be defined as pairings of formation rules and corresponding
interpretation rules. If they can be applied to an open input class, they are called
productive. Productive patterns crucially rely on the fact that there is a general
interpretation rule for their productions; otherwise they could not be freely applied.
Meaning and semantics
13
However, the semantics of word formation is complicated by the fact that there
are often several interpretation rules for the same formation rule. This is a situation
quite different from the composition of sentence meaning where, as a rule, there
is only one composition rule for each syntactic rule. Compounds in particular are
notoriously manifold in how the meanings of the parts are related: a ‘book title’ is
a title that names a book, a ‘book shelf ’ is a shelf for books, a ‘book shop’ is a shop
where books are sold, and so on.
Another reason for the complexity of word formation semantics is the existence
of so many cases with unpredictable, lexicalized meanings. These blur the picture of
the semantic rules of word formation. If you take a look into a standard dictionary
you’ll find that very many entries are complex words. They are listed in the dictionary
because their meanings are irregular – otherwise there would be no need to include
them for explanation. Extreme cases are words like butterfly; the word looks like a
compound of butter and fly, but its meaning has nothing to do with ›butter‹ and
little to do with ›fly‹. Very frequent are complex words with special meanings that
are narrower than their regular meaning, e.g. drinker, which does not just mean
›someone who drinks‹, but rather ›someone who regularly drinks too much alcohol‹.
When two nouns occur in a sentence, like book and shelf or book and shop in (3),
often the sentence tells us how they are to be linked semantically:
(3)
a.
Ann put the book onto the shelf.
b.
She had bought it in the shop around the corner.
In (3a) we see that the shelf is something the book is put onto, in (3b) that the
shop is something where the book was bought. The respective verb specifies a
semantic relation between the two nouns. With compounds, there is nothing like
this to help us to know what the two parts have to do with each other. They are
just jammed together. There are authors who suggest that compounding is a ‘relic
of protolanguage’ (Jackendoff 2009:111). According to the theory of protolanguage
(Bickerton 1992), there was a stage in the evolution of language where there were
words with meanings but not yet grammar. Words were just somehow combined; no
functional elements, no word classes, no rules for sentences. According to Bickerton,
protolanguage still exists, for example as an early stage in language acquisition or
in the speech of aphasics who have lost grammar. Jackendoff adds compounding as
another case.
If the way in which two elements are to be semantically combined is not explicitly
expressed, there is more room for interpretation. Nevertheless, the interpretation of
compounds is not just mere guessing. It cannot be, for we obviously agree on how to
interpret regular compounds. The reason for this is that the way semantically to link
a noun to another noun is often built into their meanings. It is part of the meaning
of shop that a shop is a place where things are sold, and it is part of the meaning of
book that it is some kind of artefact which can be bought. Putting this information
together leads us to the interpretation of book shop. We will turn to the semantics of
word formation in various places in this book (8.2.2, 11.6.2, 12.2.2 and 12.3.3).
14
Understanding semantics
1.3 SEMANTICS: ITS SCOPE AND LIMITS
Obviously, we are able to understand sentences in isolation, for example if we find
them written on a piece of paper or appearing on a screen of our computer. We
know word meanings; we know the meanings of grammatical forms; we know how
to combine the meanings of the words in a complex expression. We need not have
a concrete CoU in order to interpret a sentence. We can just stop interpreting at the
meaning level of expression meaning.
One way of explaining what semantics is about is to say that it describes this part of
our linguistic knowledge: the cognitive equipment that enables us to interpret regular
linguistic expressions out of context, simply as linguistic forms. We acquire this
cognitive equipment along with grammar when we learn our native language. And
we have practical evidence for assuming that this equipment is, by and large, uniform
among our speech community. We normally understand what others say and we are
normally understood – at least at the level of expression meaning.
Of course, there are considerable differences in lexical knowledge among speakers
of the same language. Other people may know certain words and their meanings
which we don’t, and vice versa, and sometimes people disagree on word meanings.
But even so, we know that these differences can be removed by explaining or
discussing word meanings. As to grammatical meanings and composition rules, we
very rarely seem to disagree with others. The cognitive equipment for interpreting
linguistic expressions is essentially shared within a speech community. Agreement
within the community is the result of a permanent synchronization of language
usage, and thereby linguistic knowledge.
This mentalist approach to semantics is adopted here. It takes the so-called
‘I-language’ as its essential object. ‘I-language’ means the mental language apparatus
of the individual speakers of a language. Making I-language the point of departure
allows one to pose the relevant questions: What are the meanings which we have
stored in our minds or compute compositionally? What kind of semantic operations
do we apply to meanings, such as the combination of meanings, or meaning shifts?
Which kind of information is involved at the three levels of interpretation? Such
questions help to say more precisely what belongs to the scope of semantics and what
is beyond its scope.
1.3.1 The scope of semantics
1.3.1.1 Explaining sentence meaning
According to the Principle of Compositionality, the following sub-disciplines of
semantics are required in order to account for our ability to interpret sentences in
isolation:
∑ lexical semantics: the investigation of expression meanings stored in the mental
lexicon (dog, sock)
Meaning and semantics
15
∑ semantics of regular word formation: the investigation of the regular meanings
of words that are formed by the productive rules of word formation (joggable,
carp food)
∑ semantics of grammatical forms: the investigation of the meaning contribution
of grammatical forms
∑ compositional sentence semantics: the investigation of the rules that determine
how the meanings of the components of a complex expression interact and
combine
Often semantics is subdivided into two sub-disciplines only: lexical semantics is
then understood as also comprising regular word semantics, and the semantics of
grammatical forms is subsumed under sentence semantics.
1.3.1.2 Interfaces of expression meaning with other levels of interpretation
The domain of semantics is not confined to sentence meaning. It also concerns the
links of sentence meaning to utterance meaning and communicative meaning. The
example in (1) illustrated two things that we do with the meaning of a sentence when
we try to make sense of an utterance of it.
First, we will try to fix references for those elements of the sentence that require
this. Elements such as the pronoun I call for fixing their reference and guide us in
doing so. Remember that the meaning of I is an instruction for determining the
referent of the pronoun as the one who has produced the utterance of this particular
‘I’. One can think of such an element as an ‘anchor’ in the sentence meaning which
tries to find a purchase in the context for anchoring the sentence meaning in the CoU.
Other such anchors are provided by definite articles, by tense or by specifications of
time or location like here and then. It is a task for semantics to investigate the nature
and function of such anchors. These anchors are technically called indexicals. They
will be dealt with in more detail in chapter 4. Indexicals are part of what is called
the interface between expression meaning and utterance meaning. This interface
consists of those components which link the level of sentence meaning to the level of
utterance meaning. It, too, belongs to the domain of semantics.
Second, there is the aspect of what expressions can mean when used in context,
in addition to what they literally mean. It may happen that we are led to shift lexical
meanings when we try to fix the reference of a sentence in a satisfactory way. This is
what we saw happen to the word bicycle in the second scenario, where it was taken in
the sense ›card with a picture of a bicycle‹. Such shifts take place before the reference
is fixed. They, too, belong to our cognitive semantic equipment. They constitute
another component of the interface between expression meaning and utterance
meaning. We will deal with them in the second part of chapter 3.
There is also an interface of expression meaning with communicative meaning. It
is constituted by elements that guide the interpretation of an utterance as a speech
act. These elements include grammatical means of indicating different types of
sentences such as declarative, interrogative or imperative. Also to be included are
16
Understanding semantics
elements such as the indication of formality or politeness or elements that directly
serve certain speech acts, like please or goodbye (see 2.3 on social meaning).
We can now define the discipline of semantics much more precisely:
DEFINITION 4 Semantics
Semantics is the study of the meanings of linguistic expressions, either
simple or complex, taken in isolation. It further accounts for the interfaces of
expression meaning with utterance meaning and communicative meaning.
1.3.2 Beyond semantics
The approach to semantics taken here draws a clear boundary. Semantics ends
where contextual knowledge comes in. Thus, utterance meaning and communicative
meaning are beyond semantics; they fall into the domain of pragmatics.
Another important point of demarcation concerns the lexical meaning of content
words. Obviously, people vary greatly in how much knowledge they have about certain
things. Is all we can know about avocados, for instance, to be considered the lexical
meaning of the word avocado? This cannot be. Otherwise the lexical knowledge of
most content words would vary too much between experts and plain language users.
Rather, the meaning of a word is something like the common denominator of the
content shared in a speech community. The actual conventional meaning is a lean
bundle of information, just enough to secure efficient communication. Therefore,
word meaning needs to be distinguished from general world knowledge. This issue
will be taken up when we discuss connotations in the next chapter (2.5), and later in
11.6, when we will distinguish between semantic knowledge and cultural knowledge,
in the context of cognitive semantics.
EXERCISES
1. Explain the distinction between the three levels of meaning in your own words.
Part of the exercises will always be to define central terms introduced in the chapter
in your own words. In order to acquire new words, it is very important that you
really make active use of them. Do not copy the definitions from the text, but try to
explain them out of your own understanding. It may be very useful to work on the
answers together with others.
2. Sue tells Martin: ‘I don’t care if you use my microphone.’
Which one of the three levels of meaning is addressed in the following questions?
Give your reasons for the answer.
a. What is the meaning of microphone?
b. Which microphone is meant?
c. Is this an offer?
d. Does the sentence relate to one or several microphones?
Meaning and semantics
17
3. The square brackets in the following sentence indicate constituent structure.
(The words within a pair of brackets form a constituent.)
[I [[don’t [know]] [where [[the kids] are]]]]
a. How many composition rules are needed for this sentence? (You need not
formulate the composition rules.)
b. Determine word by word if the word form carries grammatical meaning. (You
need not specify the grammatical meanings.)
4. Describe the meanings of the following compounds. Try to explain how the
meanings come about:
a. schoolboy
b. schoolmate
c. schoolhouse
5. Read chapter 1 of Isac & Reiss (2008) on I-language.
FURTHER READING
1.1 Lyons (1995, ch. 1) on levels of meaning. Verschueren (1999) ch. 1 on inferences
and speech acts, ch. 3 on the role of context in interpretation. Levinson (1983) on
Austin’s speech act theory and Grice’s theory of implicatures; Huang (2007) on
speech act theory and relevance theory.
1.2 Van Valin (2001) and Tallerman (2011) on elementary syntax. On word formation
see Plag (2003) and ch. 19 by Laurie Bauer and Rodney Huddleston, in Huddleston &
Pullum (2002), Olsen (2012) on the meaning of compounds.
1.3 Isac & Reiss (2008, ch. 1) for an introductory definition of I-language; Huang
(2007, ch. 7) on the distinction between semantics and pragmatics.
2
?
Dimensions of meaning
This chapter will try to convey a more precise idea about expression meaning. In
the first part about descriptive meaning, we will consider the relationship between
meaning, reference and truth. The second part is concerned with non-descriptive
meaning, i.e. dimensions of lexical meaning that are relevant on the level of social
interaction or for the expression of subjective attitudes and evaluations.
2.1 MEANINGS ARE CONCEPTS
In order to understand what kind of entities word and sentence meanings are, the
best thing we can do is consider the role that meanings play in actual communication.
We will discuss another concrete example and assume a CoU that takes up scenario 1
from 1.1.2: Mary, just back from her trip, finds her daughter Sheila quite upset. Sheila
has spent the time with Mary’s dog Ken, and the two do not like each other. When
asked what happened, Sheila answers:
(1)
The dog has ruined my blue skirt!
Let us suppose that what Sheila says is true and that Mary believes what Sheila
says. Mary will then know something she did not know before: that Ken has ruined
Sheila’s blue skirt. She knows this because Sheila said (1) and because this sentence
has the meaning it has. Let us take a closer look at how the transfer of information by
such a sentence works, first for a single word and then for the whole sentence.
2.1.1 The meaning of a word
We assume that Sheila is referring to Ken. What enables Mary to recognize that?
Sheila used the words the dog: the definite article the and the noun dog. Both play
an important role. The main information is conveyed by the noun. It specifies the
referent as an entity of a certain kind, namely a dog. What entitles us to say so? It
is the fact that the word dog means what it means. When you were asked to explain
what the word dog means, you will probably say that dogs are a certain kind of
medium-sized animal with four legs and a tail, that they are often kept as pets, that
they bark, that they may bite, etc. In other words, you will most likely give a general
description of dogs. This is an adequate reaction: giving a general description of dogs
may well count as an explanation of the meaning of dog. At least roughly, the meaning
Dimensions of meaning
19
of such words may safely be regarded as a description of the kind of thing the word
can be used for.
Now, a very important point to realize is this: the word does not carry this
description with it. This can be seen from the trivial fact that words which we do not
know do not have any meaning to us. What a word in fact carries with it when it is
spoken and heard is its sound form (or its spelling, if it is written). When Sheila says
the word dog, she produces a certain sound pattern. The respective sound pattern
is stored in Mary’s mind as part of her linguistic knowledge and enables her to
recognize the word when she hears it.
The meaning of the word dog, i.e. the description of dogs, must also be something
residing in Mary’s mind. It must be knowledge directly linked to the sound pattern of
the word. The meaning is therefore a mental description. For mental descriptions in
1
general, the term concept will be used. A concept for a kind, or category, of entities
is knowledge that allows us to discriminate entities of that kind from entities of other
kinds. A concept should not be equated with a visual image. Many categories we have
words for, like mistake, thought, noise, structure, mood, are not categories of visible
things. But even for categories of visible things such as dogs, the mental description
is by no means exhausted by a specification of their visual appearance. The dog
concept, for example, also specifies the behaviour of dogs and how dogs may matter
for us (as pets, watch dogs, guide-dogs, dangerous animals that may attack us, etc.).
We can now give a partial answer to the question of how Mary is able to recognize
that Sheila is referring to Ken: Sheila acoustically produces the word dog; Mary
recognizes the sound pattern; in her mind the pattern is linked to the meaning of the
2
word dog, the concept ›dog‹; the concept is a mental description of a potential referent.
So due to the use of the word dog, Mary knows what kind of entity Sheila is referring to.
That Mary has the concept ›dog‹ linked to the sound pattern of dog in her mind
is, of course, only part of the story. Sheila must have the same concept in her mind
linked to the same sound pattern. More generally, a word can only be considered
established if its form and meaning are linked in the minds of a great number of
language users.
Still, we have not explained how Mary is led to assume that Sheila refers to this
particular dog. The crucial clue to an explanation is the definite article the. Had Sheila
used the indefinite article a instead, Mary would not have concluded that Sheila was
referring to Ken. What is the meaning of the definite article? It does not provide a
direct cue to Ken, but it signals that the description supplied by the following noun
applies to an entity in the given CoU which the addressees are supposed to be able to
sort out. Therefore the article will cause Mary to ask herself which entity in the given
3
CoU fulfils these conditions.
1 In Chapter 11 on cognitive semantics the notions ‘concept’ and ‘category’ will be treated in more
depth and detail. For the present purposes you may take a category as a set of entities of the same
kind.
2 This kind of quote will be used for concepts, and meanings in general: ›x-y-z‹ is the concept that
constitutes the meaning of the expression x-y-z.
3 We will have a much closer look at the meaning of the definite article in 4.4.2.
20
Understanding semantics
This is how far the meanings of the words the dog take us in reconstructing the
communication between Sheila and Mary with respect to the reference to this
dog Ken. For the conclusion that it is Ken which Sheila is referring to, Mary needs
extra-linguistic context information. The fact that Sheila is using the definite article
restricts the choice of candidate dogs to those Mary and Sheila both know. The
family’s dog Ken is a privileged candidate, since he is firmly established as ‘the dog’
in the given context.
2.1.2 The meaning of a sentence
In her mind, Mary has the forms and meanings of all words in (1) at her disposal.
She also knows the grammatical meanings of the singular form, of the positive form
of the adjective and of the indicative present perfect form of the verb (recall 1.2.1).
Applying all this and her knowledge of grammar to the linguistic input, she will be
able to compose the expression meaning of the whole sentence (1.2). The result is one
complex concept which combines the meanings of all elements of the sentence. Let
us call this a concept for a situation. The main component of the situation concept
is the concept ›ruin‹ contributed by the verb. It is of central importance because
4
it connects all other elements. As a concept for an event of the kind ‘x ruins y’ it
involves three elements: the event itself, the ruiner x and the ruined object y. In the
complete situation concept, the event is described as one of ruining, the ruiner is
described as a dog that is identifiable in the given CoU, the ruined object is described
as a skirt, a blue one, linked to the speaker (recall 1.1.1 for the meaning of possessive
pronouns like my and your); the present perfect tense contributes the specification
that the situation at the time of utterance results from a previous event of the kind
indicated. Thus the meaning of the sentence as a whole is a concept for a specific kind
of situation. It can roughly be described as in (2). The description does not contain
an explanation of the content-word meanings, but it makes explicit the contribution
of the functional elements (1.1.1).
(2)
›the situation at the time of utterance results from a previous event in which a
dog that is uniquely determined in the CoU ruined a blue skirt which is uniquely
determined by its being linked to the speaker‹.
What was said about the meanings of words and sentences can be summed up as
follows:
DEFINITION 1 Meanings of content words and sentences
The meaning of a content word (noun, verb, adjective), is a concept that
provides a mental description of a certain kind of entity.
The meaning of a sentence is a concept that provides a mental description of
a certain kind of situation.
4 See chapter 5 on predication.
Dimensions of meaning
21
2.2 DESCRIPTIVE MEANING
In the previous section it was established that expression meanings are concepts.
Actually the discussion here and in chapter 1 was confined only to a certain
dimension of meaning, namely the dimension that bears on reference and truth. It is
called descriptive meaning, alternatively ‘propositional meaning’. We will elaborate
on descriptive meaning now, making more explicit how it is related to reference and
truth. We will turn to non-descriptive meaning in the second half of the chapter.
2.2.1 Descriptive meaning and reference
2.2.1.1 Reference and the descriptive meaning of words
When dealing with reference, the first thing to be observed is that, strictly speaking, it
is usually not simply words that have referents. If the sentence in (1) is true, it involves
reference to five things: the dog (an object, in the widest sense), the speaker’s blue skirt
(another object), the speaker herself, the ruining of the skirt (an event) and the time of
utterance (a time). Table 2.1 shows which elements of the sentence refer to these five
things. The subject NP and the object NP each have a referent; the possessive pronoun
within the object NP has a referent of its own. The finite verb contributes reference to
a certain kind of event and, due to its tense, to a certain time. The adjective blue has no
referent of its own, but it contributes to the description of the referent of the NP my blue
skirt. The example shows that the referring elements of the sentence can be phrases (e.g.
NPs), words (the verb) or grammatical forms (tense).
Table 2.1
Five referents of sentence (1)
Type
Referent
Referring element
object
the dog
NP
the dog
object
the speaker’s blue skirt
NP
my blue skirt
object
the speaker
poss. pronoun my
event
ruining
verb
time
utterance time
tense has ___ed
ruin
All this notwithstanding, it makes sense to talk of the potential referents of
content words. Since the referent of an NP is essentially described by the noun, we
may loosely speak of it as the ‘referent of the noun’. Analogously, we can talk of the
‘referent of a verb’. Adjectives never have a referent of their own, but they always
describe the referent of some NP (see 5.4 for details). Thus, still more loosely
speaking, we may extend the notion of referent to adjectives, keeping in mind that
their ‘referents’ are borrowed, as it were. In simple words, a potential referent of
a content word is whatever can be called so. If the notion of potential referent is
22
Understanding semantics
extended to all content words, descriptive meaning can be defined in a way that
relates it directly to reference:
DEFINITION 2 Descriptive meaning of content words
The descriptive meaning of a content word is a concept for its potential
referents.
When a sentence is used in a particular CoU, the addressees will try to fix concrete
referents that match the descriptions. However, and this is a very important point, it
may be impossible to fix referents, if the sentence is not true. Consider the sentence
in (3):
(3)
There is a letter for you.
Let us assume that Sheila says so to her mother, but that she is not telling the truth:
there is no letter for Mary. There may be a letter, but not for her, or no letter at all. In
any event, if the sentence is not true, the NP a letter for you lacks a referent. Usually,
the finite verb of the sentence has a concrete event referent only if the sentence is true.
For example, if (1) is false in some CoU, then the dog has not ruined the speaker’s
blue skirt and hence the verb ruin, in that CoU, fails to have a referent.
2.2.1.2 The descriptive meaning of sentences: propositions
There is no generally accepted notion for what a sentence as a whole refers to in a
given CoU. For lack of a better term it will be called the situation referred to. The
referents of the referring elements of the sentence are components of the situation
referred to; together, in their particular constellation, they form a complex state of
affairs. For the sentence in (1), the situation referred to in the given CoU comprises
the five referents listed in Table 2.1: (i) the dog Ken, (ii) the speaker Sheila, (iii) her
blue skirt, (iv) its being ruined by the dog, (v) the time referred to. The particular
constellation of these five referents is as the sentence has it: at the time referred to, the
blue skirt is Sheila’s and its condition is the result of it being ruined by the dog Ken.
The notion of the situation referred to only makes sense if the sentence is true: as
we have seen, some elements of the sentence may lack a referent if it is not true. Thus,
only if a sentence is true in a particular CoU does it properly refer to a situation of the
kind it describes. Therefore, whenever the term situation referred to is used, it will be
assumed that the sentence is true.
In analogy to the notion of potential referents we can talk of the situations
potentially referred to. These are all those situations that fit the mental description
provided by the meaning of the sentence, i.e. all the situations for which the sentence
is true. The descriptive meaning of a sentence can now be defined as in Definition 3.
In accordance with common terminology, the descriptive meaning of a sentence is
called its ‘proposition’. Alternatively, the proposition of a sentence will be referred to
as the ‘situation expressed’, or the ‘situation described’.
Dimensions of meaning
23
DEFINITION 3 Proposition
The descriptive meaning of a sentence, its proposition, is a concept that
provides a mental description of the kind of situations it potentially refers to.
As we have seen, it is not only content words that shape the descriptive meaning
of the sentence. Functional elements such as pronouns and articles or tense, a
grammatical form, contribute to the proposition as well (recall the description of the
meaning of (1) given in (2)). Making use of Definition 3, we can give the following
general definition:
DEFINITION 4 Descriptive meaning of a word or a grammatical form
The descriptive meaning of a word or a grammatical form is its contribution
to the descriptive meanings of the sentences in which the word or grammatical
form may occur.
To sum up, the descriptive meaning of a sentence is a concept for a certain kind of
situation. If the sentence is true in a CoU, such a situation actually exists and can
be considered the referent of the sentence. The situation referred to contains the
referents of all referring elements of the sentence. Table 2.2 gives a survey of different
types of potentially referring expressions, their respective descriptive meanings and
types of referents.
Table 2.2
Descriptive meanings of the elements of sentence (1)
Expression
(type)
Descriptive meaning
(* definitions adopted from The New Oxford Dictionary
of English)
Referent
(type)
skirt
(noun)
›a woman’s outer garment fastened around the waist
and hanging down around the legs‹ *
object
eat
(verb)
›put (food) into the mouth and chew and swallow it‹*
event
blue
(adjective)
›of a colour intermediate between green and violet, as
of the sky or sea on a sunny day‹*
object [borrowed]
the [noun]
(article)
the referent of the noun is uniquely determined in the
given CoU
–
I
(pronoun)
the referent is the speaker
object
The dog has ruined my
blue skirt.
(sentence)
see (2)
situation
24
Understanding semantics
2.2.2 Denotations and truth conditions
2.2.2.1 Denotations
The descriptive meaning of a content word is a concept for its potential referents. As
such it determines, or mentally describes, a category of entities. The meaning of dog
5
is a concept that determines the category DOG of all dogs, the verb concept ›ruin‹
determines the category RUIN of all events of ruining. The category determined by
the meaning of a content word is called its ‘denotation’; a word is said to ‘denote’ this
category.
DEFINITION 5 Denotation
The denotation of a content word is the category, or set, of all its potential
referents.
The denotation of a word is more than the set of all existing entities of that kind. It
includes fictitious referents as well as real ones, usual exemplars and unusual ones,
6
maybe even exemplars we cannot imagine because they are yet to be invented.
The relationship between a word, its meaning and its denotation is often depicted
in the semiotic triangle, a convenient schema which will be used in this volume in
a variety of forms. Figure 2.1 gives the semiotic triangle for the descriptive meaning
of content words as such. The arrow that connects the word with its denotation is
drawn with a broken line. This is to indicate that a word is not directly linked to its
denotation, but only indirectly via its descriptive meaning.
Figure 2.1
The semiotic triangle for a content word in general
content word
descriptive meaning:
a concept
denotation:
a category
5 SMALL CAPITALS are used for categories.
6 Yet the totality of existing exemplars of a category certainly is representative of the category, and
primarily shapes our concept for this kind of thing. We will occasionally use the term actual
denotation for the subset of the denotation that is formed by its real existing members. In 11.6
the notion of cultural category will be introduced for actual denotations. There we will see more
clearly what the relation is between the total denotation of a word and the corresponding cultural
category – and why the distinction is important.
Dimensions of meaning
25
When a content word is actually used in a concrete CoU, we deal with a token of the
word, i.e. a particular spoken or written realization. The semiotic triangle then yields
a relationship between the word token, its meaning and its referent: the meaning
describes the referent to which the word, in the given CoU, refers (Fig. 2.2).
Figure 2.2
The semiotic triangle for a content word token
content word
descriptive meaning:
a concept
referent in the
world
2.2.2.2 Truth conditions
There is no established term for what would be the denotation of a sentence.
In analogy to the denotation of a content word it would be the set, or category, of
all situations to which the sentence can potentially refer, i.e. the category of
all situations in which the sentence is true. There is, however, another notion that
is quite common and directly related to the would-be denotation of a sentence: its
7
so-called truth conditions.
DEFINITION 6 Truth conditions
The truth conditions of a sentence are the conditions under which it is true.
We will say that a sentence has the truth value ‘true’ if it is true, and the truth value
‘false’ if it is false. In order to see what is meant by ‘truth conditions’, let us consider
sentence (1) once more. Obviously, this sentence taken as such is not merely true
or false; rather, its truth value depends on the CoU in which it is used: the question
8
of its truth or falsity arises only in relation to a given CoU. Sentence (1) is true in
a given CoU if there is a uniquely determined dog and a uniquely determined blue
7 The definition given here is sufficient for more general purposes. It will be made more precise in
4.7.3 in order to integrate presuppositions.
8 There are exceptional sentences that have the same truth value in all possible CoUs: these are called
‘logically true’ or ‘logically false’ (see 7.2); examples are sentences like ducks are birds, which is
logically true, or two times three is seven, a logically false sentence. Only for this kind of sentences,
can the truth value be determined independently of a given CoU.
26
Understanding semantics
skirt belonging to the speaker. Additionally, the dog must have done something to the
skirt such that at the time of utterance the skirt is ‘ruined’. If all these conditions are
fulfilled in a given CoU, the sentence is true in this CoU, and vice versa: if the sentence
is true in a CoU, then all these conditions are fulfilled.
This is the reason why it is possible to communicate things by saying a sentence. We
utter a sentence like (1) or any other declarative sentence, and communicate, by the
way in which we speak, that what we say is true in the given CoU. The addressee(s)
will take the truth of the sentence for granted – assuming that they have no reason
to doubt what is being said – and conclude that its particular truth conditions are
fulfilled in the given CoU.
If sentence (1) were uttered by someone else and/or at some other time and/or
under different circumstances, it might well be false. Of course, this does not mean
that there is only one CoU where (1) is true. Many blue skirts belonging to somebody
have been, or will have been, ruined by some dog and in all these cases the owner
may truly utter sentence (1).
For an explicit formulation of the truth conditions of (1) we can resort to the
description of the meaning of (1) given in (2) above.
(4)
Truth conditions of sentence (1):
The sentence The dog has ruined my blue skirt is true in a given CoU if and only
if the following holds:
‘The situation at the time of utterance results from a previous event in
which a dog that is uniquely determined in the CoU ruined a blue skirt
which is uniquely determined by its being linked to the speaker.’
A proper definition of the truth conditions of a sentence S always takes this form:
‘S is true in a given CoU if and only if …’
‘If and only if ’ – abbreviated ‘iff ’ – means ‘if ’ in both directions: ‘S is true if the truth
conditions hold, and if the truth conditions hold, then S is true.’
We can now see that the notion of truth conditions is equivalent to the notion of
the denotation of a sentence. The denotation of a sentence would be the category, or
set, of all the situations the sentence can refer to. If one knows the truth conditions of
a sentence, one knows which situations the sentence can refer to: to those which are
part of a CoU where the sentence is true. Conversely, if one knows which situations a
sentence can refer to, one is able to define its truth conditions: the sentence is true in
a CoU iff this CoU comprises a situation that the sentence can refer to.
In analogy to Fig. 2.1, the connection between a sentence, its proposition and its
truth conditions can be put as follows: the descriptive meaning of the sentence is its
proposition, and the proposition determines the truth conditions of the sentence.
The resulting picture is given in Fig. 2.3, another variant of the semiotic triangle.
Dimensions of meaning
27
Figure 2.3
The semiotic triangle for a sentence
sentence
proposition:
a complex concept
truth conditions
2.2.3 Proposition and sentence type
So far in this discussion of sentence meaning one aspect has been neglected: the
grammatical type of the sentence too contributes to its meaning, and this contribution
is non-descriptive. Compare, for example, sentence (1) to its interrogative counterpart
(5):
(1)
The dog has ruined my blue skirt.
(5)
Has the dog ruined my blue skirt?
The question describes exactly the same sort of situation. Hence it is considered to
have the same proposition as (1). Yet the total meaning of (5) is, of course, different
from the meaning of (1): (5) renders a question while (1) renders an assertion. The
difference in meaning is due to the grammatical forms of the sentences or, technically
speaking, to differences in grammatical sentence type. (1) is a so-called declarative
sentence. Declarative sentences in English have a certain word order: the finite verb is
in the second position of the sentence, usually after the subject. (5) is an interrogative
sentence of the yes-no-question type: in English, the finite verb is in the initial
position and has to be an auxiliary verb.
The semantic contribution of the grammatical sentence type is not part of
the proposition. For declarative sentences, it consists in presenting the situation
expressed as actually pertaining. This sentence type is therefore used for making
assertions, communicating information, etc. The interrogative sentence type, by
contrast, leaves open whether or not the situation pertains. It is therefore the
standard option to be chosen for asking questions.
Imperative sentences represent a third grammatical sentence type:
(6)
Don’t ruin my blue skirt!
In English imperative sentences, the finite verb fills the first position of the sentence
and normally there is no explicit subject. Imperative sentences are used for
commands, advices and similar speech acts. The proposition of (6) would be that the
addressee(s) do not ruin the speaker’s blue skirt.
28
Understanding semantics
The meaning contribution of grammatical sentence type is a first example of
non-descriptive meaning. We will now consider two more types: social meaning and
expressive meaning. The meaning of sentence type belongs to neither of them.
2.3 MEANING AND SOCIAL INTERACTION:
THE DIMENSION OF SOCIAL MEANING
Talking to others is social interaction, i.e. an activity exerted in coordination with
others. Any verbal utterance will receive an interpretation as a communicative act
(1.1.3) in the current social network, and in this sense it always has a social function.
Language as such can be said to serve first and foremost social functions. (This
does not contradict the view that language is primarily a means of communication:
communication, in particular the communication of information, is of course a very
important type of social interaction.)
2.3.1 Expressions with social meaning
The term social meaning does not refer to this general aspect of verbal interaction,
and is thereby not to be confused with the communicational meaning of a verbal act.
Rather, social meaning is on a par with descriptive meaning: it is a dimension of the
lexical meaning of certain words, phrases or grammatical forms. If an expression
has social meaning, it has so independently of the particular CoU. Most expressions
and grammatical forms do not have social meaning, but some do. If expressions with
social meaning are built into sentences, these inherit a dimension of social meaning
due to the principle of compositionality. Let us consider an example. Sheila is on the
train in Germany and is addressed by the ticket inspector:
(7)
a.
Ihre Fahrkarte, bitte! – Danke.
(German)
b.
Deine Fahrkarte, bitte! – Danke. (German)
c.
‘Your ticket, please! – Thank you.’
(7a) would be appropriate if Sheila is an adult and no acquaintance of the inspector.
9
The 3rd person plural form of the possessive pronoun, Ihre, literally ‘Their’, is
required for the formal, or ‘polite’, style of speech used for addressing adults. By
contrast, (7b) contains the simple 2nd person singular possessive pronoun dein and
would be the proper, informal, form of address if Sheila were a child, a relative or a
close acquaintance of the ticket inspector. Otherwise, using (7b) would be considered
rude. If the inspector addressed Sheila in English, (7c) would be adequate in all cases.
But when speaking German, the inspector is forced to choose between the formal and
the informal way of address (or to avoid the use of pronouns altogether).
The formal German pronoun Sie, when used as a term of address, has the same
descriptive meaning as English you: it indicates reference to one or more addressees.
9 In their ‘polite’ use the 3rd person plural pronoun and the corresponding possessive pronoun are
written with a capital letter: Sie for you, Ihr for your.
Dimensions of meaning
29
But, in addition, it has a non-descriptive dimension of meaning which English you
lacks: the social meaning of indicating a certain kind of social relationship between
the speaker and the addressee. Correspondingly, the informal variants du (singular)
and ihr (plural) have the same descriptive meaning as Sie (if one disregards the
differentiation in number) but differ in social meaning. The distinction between the
two kinds of relationship relevant for choosing either Sie or du in German is also
relevant in other respects: it coincides with the use of surnames with title (along
with Sie) vs first names as vocative terms of address. For example, the unmarked
way of address and reference to the addressee would be (8a) or (8b), while the mixed
variants in (8c) and (8d) are marked under normal circumstances:
(8)
a.
Ist das Ihr Fahrrad, Herr Schmidt?
formal
b.
Ist das dein Fahrrad, Helmut?
informal
c.
?? Ist das Ihr Fahrrad, Helmut?
mixed
d.
?? Ist das dein Fahrrad, Herr Schmidt? mixed
‘Is that your bicycle, Helmut/Mr Schmidt?’
(7a) and (7b) above contain two further expressions with social meaning: bitte
›please‹ and danke ›thank you‹. Unlike the German terms of address, these two
expressions have exclusively social meaning; they are the first expressions we
encounter which lack descriptive meaning altogether. (English thank you, containing
you, might be considered as referring to the addressee(s). To this extent it also has
descriptive meaning.) The sentences in (7) are elliptical: they lack anything but
the direct object and could be spelt out as Geben Sie mir Ihre Fahrkarte, bitte (lit.
‘give you me your ticket, please’). At least they will be interpreted in this sense
and can thus be considered elliptical imperative sentences with the proposition
›addressee gives speaker addressee’s ticket‹. The addition of bitte does not change
the proposition. Rather it marks the request as modestly polite. Bitte, in this use, is a
mere formality marker, indicating, similar to the forms of address, a certain kind of
social relationship between speaker and addressee(s).
Danke ›thank [you]‹ has no descriptive meaning either. It just serves as a slightly
formal response that counts as a recognition of a favour or service. Given different
social rules, the ticket inspector might have bowed instead of saying anything. Since
the utterance lacks a proposition, the question of truth does not arise. One might
think that, instead, the question arises whether or not the speaker is in fact grateful.
But their emotions are not relevant. If in a relevant situation people say ‘Danke!’ or
‘Thank you!’ they have automatically performed the act of thanking, regardless of
what they actually feel. Verbal acts of thanking obey specific social rules, and these
are observed if one merely utters one of the conventional phrases of thanking.
All languages have set phrases with a clear-cut social meaning and no other
dimension of meaning: phrases of greeting (Hi) or saying goodbye, phrases of
apologizing (sorry), recognizing (thank you), or answering the phone. For each such
phrase in each language, there is a social rule that defines the circumstances under
30
Understanding semantics
which it is properly used and what it then means. For example, the use of goodbye
phrases is only adequate if both speaker and addressee(s) are in some form of social
contact; saying goodbye then means a certain step in ending the current contact.
Generally, the defining criterion for an expression carrying social meaning is this:
DEFINITION 7 Social meaning
An expression or a grammatical form has social meaning if and only if it
conventionally serves the indication of social relations or the performance of
conventionalized social interaction; there must be rules for social interaction
that govern its use.
By contrast, the use of expressions with descriptive meaning is governed by rules
of factual adequacy; for example, a declarative sentence is used correctly if its truth
conditions are fulfilled in the given CoU.
In 1.1, three levels of meaning were introduced. Expressions like please or goodbye,
with exclusively social meaning, belong to the interface between expression meaning
and communicative meaning (1.1.3, 1.3.1). They are elements in the utterance that
provide the crucial clues for the kind of speech act which is performed. The social
rules that govern their use are treated in Austin’s speech act theory under the label
‘felicity conditions’ (i.e. conditions necessary for the speech act actually to come
about; see Further Reading in chapter 1 for references). Table 2.3 displays examples
of different kinds of expressions with social meaning.
Table 2.3
Social meaning
Expression (type)
Social meaning
Descriptive meaning
German du
(pronoun of address)
informal relationship
the person addressed
German Sie
(pronoun of address)
formal relationship
the person or persons addressed
English you
(pronoun of address)
–
the person or persons addressed
Sheila
(proper name as term of address)
informal relationship
the person called Sheila
Mr Murple
(proper name as term of address)
formal relationship
the person called Mr Murple
please
(adverb)
formal demand
–
Hi!
(complete utterance)
informal greeting
–
Dimensions of meaning
31
In the linguistic literature, phenomena of social meaning are mostly treated under
the label of honorifics. These include forms of address and self-reference, special
grammatical forms such as the Japanese masu-form (see next section) and other
linguistically conventionalized forms of paying respect.
2.3.2 Social meaning in Japanese
In European and European-influenced societies, social differentiation of speech
is only moderately reflected in the language system. Apart from set phrases,
social meaning is essentially restricted to terms for reference to the addressee.
Furthermore, the system of differentiation is usually restricted to two levels of
formality, an informal and a more formal way of speaking. In other cultures, however,
the social differentiation of expression pervades language to a much greater extent.
Such languages include Japanese, Korean, Vietnamese, Thai, Burmese and former
imperial Chinese. We will take a closer look at Japanese.
In English or German, you do not find any formality markers in sentences unless
the speaker refers explicitly to the addressee. There would be no simple way of
reformulating an utterance like The dog ate the yellow socks to express your relationship
with the addressee along the lines relevant for the use of surname vs first name or
the use of Sie vs du in German. (The only possibility would be to add a vocative to
the sentence: The dog ate the yellow socks, John / Mr Murple.) In Japanese, however,
one would have to choose among two or more levels of formality in any sentence
whatsoever, because formality is obligatorily marked on the finite verb. In Japanese, the
sentence the dog ate the yellow socks would correspond to either (9a) or (9b):
(9)
a.
inu
wa
kiiroi
sokkusu
o
tabe-
ta.
dog
TOP
yellow
socks
ACC
eat
PT
10
‘the dog ate the yellow socks’
b.
inu
wa
kiiroi
sokkusu
o
tabe-
mashi- ta.
dog
TOP
yellow
socks
ACC
eat
FORM
PT
‘the dog ate the yellow socks’
The plain past tense form of the verb in (9a) represents the informal way of talking
which is only appropriate in rather intimate relationships like those within a family,
between lovers or good friends. Normal formal talking, which is the standard between
people of the same status, is expressed by inserting a formality marker, here in the form
10
= topic marker (‘as for … ’), ACC = accusative case marker, FORM = formality marker, PT =
past tense, PS present tense. In Japanese, particles follow the NP; there are no articles. We use
the Hepburn transcription for Japanese which is oriented at the pronunciation (consonants are
roughly pronounced as in English, vowels as in Spanish or German).
TOP
32
Understanding semantics
-mashi-, between the stem of the verb and its tense ending (9b). The resulting form is
called the masu-form. The marker -mashi- has no descriptive meaning.
In addition to marking the level of formality on the finite verb, Japanese has rich
differentiation among pronouns, or pronoun-like expressions, for the addressee as
well as for the speaker. Instead of the single 1st person singular pronoun I in English,
a Japanese speaker has to choose among plain omission (usual and neutral), boku or
ore (informal, men only), watashi (formal, standard), watakushi (very formal) and
other, more special expressions. Furthermore, there are different nouns and verbs
for things belonging to, or actions done by, oneself vs others. Let us consider another
example in order to get an idea of how the differentiation of formality works:
(10) a.
Uchi
ni
i-
ru.
b.
Uchi
ni
i-
mas-
u.
c.
Uchi
ni
ori-
mas-
u.
home
in
be
FORM
PS
‘… am/is/are at home’
The sentences in (10) are typical colloquial Japanese sentences, lacking a subject
term. The word uchi for someone’s home is an informal term. The verb iru ›to be
[temporarily]‹ is used in its plain, informal form in (10a). (10b) is more formal,
replacing plain iru with its masu-form imasu. In addition to the grammatical form,
the choice of the verb itself matters for social meaning. The verb iru as such has no
social meaning attached to it. In (10c), however, it is replaced by the verb oru in its
masu-form orimasu. Oru also means ›to come/go‹, but is ‘humble’: by using it, its
subject is socially lowered. There is also a variant of ›to be/come/go‹ that raises its
subject referent in social status, the verb irassharu with its masu-form irasshaimasu:
(10) d.
Otaku ni
irasshai- mas-
u.
home
be
PS
in
FORM
‘… is/are at home’
If, for the sake of simplicity, the meanings ›to go/to come‹ of oru and irassharu
are disregarded, the four sentences in (10) all have the same descriptive meaning:
›[someone not specified] is at home‹. The term uchi for someone’s home is informal,
while otaku is a formal term that marks the house as belonging to someone superior.
It is now interesting to see how the differences in social meaning restrict the use and
interpretations of the sentences. The employment of terms with social meaning in
Japanese is governed by two principles: (i) the addressee is never inferior, and (ii)
the speaker is never superior. Given these principles, (10d), although it contains no
subject, cannot be used in the sense ›I am at home‹. (10c) amounts to ›I am at home‹
or that somebody is at home who, in face of the addressee(s), ranks equal with the
speaker, e.g. members of the speaker’s family. By contrast, (10a) and (10b) can mean
Dimensions of meaning
33
›I/you/he/she/it/we/they am/is/are at home‹ since the verb is neutral with respect to
the social status of the subject.
2.4 MEANING AND SUBJECTIVITY:
THE DIMENSION OF EXPRESSIVE MEANING
Anything we say will also be taken as the expression of a personal emotion, opinion
or attitude. When you imagine Sheila saying ‘The dog has ruined my blue skirt’, you
will probably have assumed a manner of speaking that displays anger, indignation
or frustration. But the sentence can also be spoken in a neutral manner or in a
way that exhibits different emotions, for example desperation or concern, but also
delight, amusement or relief. It could be pronounced in a mean or a kind manner. Or
hastily, or particularly slowly. All these ways of shaping an utterance would serve the
expression of Sheila’s feelings and attitudes towards the dog, or Mary, or the entire
situation. But even if she chose a neutral way of speaking, Sheila would inevitably
display certain traits of her personality, by choosing the term ruin for what the
dog did to her skirt, by not calling the dog by its name, by reporting damage of her
property to the dog owner, etc. In this sense, every utterance serves, consciously or
not, the expression of personal feelings, evaluations and attitudes.
2.4.1 Expressive meaning
Again, this general aspect of language use, its expressive function, is not what is
meant with the notion of expressive meaning. On a par with descriptive and social
11
meaning, expressive meaning is part of the lexical meaning of certain expressions,
a semantic quality of words and phrases independent of the CoU and of the way
they are being spoken. Again, there are two kinds of expressions with this special
meaning: some have only meaning and others have both descriptive and expressive
meaning. There do not seem to be expressions with both social and expressive
meaning. We will first turn to the former.
DEFINITION 8 Expressive meaning
An expression or a grammatical form has expressive meaning if and only if
it conventionally serves the immediate expression of subjective sensations,
emotions, affections, evaluations or attitudes.
Let us call expressions with expressive meaning simply expressives. The most
typical instances are words and phrases such as ouch, wow, oh. Such interjections
11 Be careful not to confuse the terms expression meaning and expressive meaning. Expression
meaning is a cover term for the semantic meaning of words, phrases and sentences, since all
these are expressions. Expressive meaning is a dimension of expression meaning, alongside with
descriptive meaning and social meaning.
34
Understanding semantics
are language-specific. Languages may differ in how many such expressions they
have, and an interjection may have different meanings in different languages. Here
are some interjections from Hungarian: fuj [fuj] (disgust), au [u] (sudden pain),
jaj [jj] (sudden pain or fright), jajaj [jjj] (sadness or concern), hüha [hyh]
(admiration, warning, fright), hú [hu:] (admiration), ejha [ejh] (astonishment).
Other examples of expressives are exclamations of various sorts, such as Gosh!,
Goddammit!, Jesus!, Oh my goodness!, and so on.
At least some feelings, sensations, attitudes and evaluations can thus be expressed in
two ways: subjectively and immediately by means of expressives, and propositionally
by forming sentences with the respective descriptive meaning. For example, the
difference between saying ‘Ouch!’ and ‘That hurts!’ is this: That hurts! is a sentence
with a regular descriptive meaning, consisting in the proposition that ‘that hurts’.
‘Ouch!’ expresses sudden pain, as would a certain grimace or moaning. You can react
to someone saying ‘That hurts!’ with ‘Really?’, ‘That’s too bad’ or ‘I don’t believe it!’,
which are replies relating to the proposition of the sentence and its being true. You
cannot reply to ‘Ouch!’ in the same way because there is no proposition.
The rules governing the use of expressives are simple. Since all expressives serve
to express personal feelings, attitudes or sensations, which are perceptible only to the
holder, their correct use is just a matter of personal judgement. For instance, saying
ugh! is generally taken as an expression of disgust. In order to decide whether the
use of this expression is correct in a given CoU, the speaker only has to make up their
mind whether or not they want the addressee(s) to believe that they find something
disgusting.
Interjections and exclamations can be used as complete utterances. Other
expressives such as hopefully, (un)fortunately or thank God can be inserted into
a sentence in order to add a personal attitude to the situation expressed. These
additions do not contribute to the proposition, as the following examples may
illustrate:
(11) a.
b.
Fortunately, Bob will arrive tonight. – Really?
I’m glad Bob will arrive tonight. – Really?
The questioning reply ‘Really?’ is always directed at a proposition asserted before.
In (11a), the remark can only be related to the proposition that Bob will arrive that
night (‘Will Bob really arrive tonight?’), not to the speaker’s attitude expressed by
fortunately. In (11b), however, the same attitude is described by the descriptive
formulation I’m glad ... and is hence part of the proposition of the sentence.
Therefore, the reply ‘Really?’ can be used to question the speaker’s claim that they are
glad that Bob will arrive (‘Are you really glad that Bob will arrive tonight?’), i.e. their
attitude described.
More drastic cases of expressives are swear words. Consider the following
12
passages from the film Pulp Fiction:
12 Quentin Tarantino, Pulp Fiction (London: Faber and Faber, 1994), pp. 140 and 13, respectively.
Dimensions of meaning
(12) J:
35
But me, my eyes are wide fuckin’ open.
V:
What the fuck does that mean?
J:
That’s it for me. From here on in, you can consider my ass retired.
(13) Any of you fuckin’ pricks move and I’ll execute every motherfuckin’ last one of
you.
The adverbial insert the fuck in (12) serves the expression of an emotion that is
probably neither necessarily positive (like joy) nor negative (like indignation) but
just strong: it adds an emotional charge to the whole utterance. Other expressives,
like bloody or the highly productive fucking add the same kind of emphasis to
single components of the proposition. None of these expressives make additions
to the propositional content. However, in the examples two other expressions with
expressive meaning do: my ass in (12) is an emotionally charged term for reference to
the speaker. It must have descriptive meaning because it is an NP in object position.
Its descriptive meaning is just ›I/me‹. Similarly, the expression you fuckin’ pricks in
(13) has the same descriptive meaning as the plain pronoun you, plus the specific
expressive content of the rest of the phrase. This expression is not just emotionally
charged but clearly offensive when used with reference to others.
In English and presumably most other languages, there exist many emotionally
charged terms for persons, such as idiot, bastard, motherfucker or ass-hole for the
offensive part, or darling, baby, honey as terms of affection. They all have the same
unspecific descriptive meaning ›person‹. The phenomenon extends to expressions for
men, women, children, body parts or other things of central importance in everyday
life, such as certain animals, vehicles, housing, food, clothes, as well as to the most
common activities: walking, talking, working, eating, drinking, sleeping or sexual
activities. Most expressive terms add a negative attitude to the meaning, others are
just emotional. Positive expressive meanings are rare. Many of these expressions have
a neutral, purely descriptive, meaning variant besides the expressive one. For example,
paw in its neutral meaning denotes the feet of certain animals. The expressive variant
of the word means ›hand‹ for its descriptive meaning part while its expressive
meaning part is an unspecific emotional emphasis.
Table 2.4
Expressive meaning
Type
Expression
Expressive meaning
Descriptive meaning
interjection
ouch
sudden pain
–
adjective
stupid
dislike, contempt
–
adverb
fortunately
preference
–
noun
bastard
dislike
›person‹
noun
paw
emotional emphasis
›hand‹ (of a person)
verb
cram
emotional emphasis
›eat‹
36
Understanding semantics
2.4.2 Social vs expressive meaning
Many semanticists consider expressive meaning and social meaning as not clearly
separated. The distinction is, however, not as difficult as it might appear. The use of
terms and forms with social meaning is governed, for this part of their meanings,
by rules of social interaction. They define what kind of social circumstances make
suitable occasions for using the expression and they define what its use is taken
for: a greeting, an apology, a polite or intimate way of referring to other persons,
etc. By contrast, the use of terms with expressive meaning is governed (again, for
this part of their meanings) by criteria which concern only subjective adequacy
with respect to expressing one’s personal feelings, attitudes etc. Sure enough, there
are rules of conduct constraining the expression of feelings or attitudes under
certain social conditions and the use of expressive terms, in particular swear
words as terms of address, may have severe social consequences. But while there
are clear social rules for the use of, say, the first name vs the surname plus title
as terms of address, there is no such rule for addressing someone with idiot or
motherfucker.
Some expressions with predominantly social meaning can be considered as means
of the socially ritualized expression of feelings and attitudes, e.g. terms of thanking,
wishing well, congratulation, condolence or apologizing. Phrases such as I’m sorry
or nice to meet you, which literally represent descriptions of attitudes, point in this
direction. Still, these phrases are primarily social and not expressive. One’s real
subjective feelings are not, and cannot be, socially relevant. What counts is the actual
expression of feelings, and behaviour consistent with having them.
Table 2.5
Dimensions of meaning
Dimension
Function
Criteria for correct use
descriptive meaning
description of referents and situations
agreement with facts
social meaning
indication of social relations and
performance of social acts
social rules of conduct
expressive meaning
immediate expression of personal
sensations, feelings, attitudes or
evaluations
subjective choice
2.5 CONNOTATIONS
If an expression has descriptive meaning, any mention of it will activate not only the
concept for its potential referents but together with it a host of further associations.
Among the associations, some are conventional. They are called connotations
and often considered to be something like a secondary meaning in addition to the
primary lexical meaning. However, connotations such as ‘dirty’ for pigs are neither
Dimensions of meaning
37
part of the descriptive meaning of pig (clean pigs can perfectly be referred to as
‘pigs’) nor do they constitute expressive meaning (the word pig can be used in
an expressively neutral way). Often the connotations of a word change, while its
meaning remains the same. For example, the connotations of the word computer
have changed considerably since the 1960s (when computers had the connotations of
dangerous super-intelligent machines threatening to escape human control and take
over), but the word still means the same. What has changed dramatically in the last
fifty years is the actual denotation (recall fn. 6) of the word, and it is these changes
which have given rise to the change of connotations. It is therefore more appropriate
to consider connotations to be connected not to the word itself (like meaning) but
rather to the actual denotation.
While a distinction is drawn in this volume between a word’s meaning and the
connotations associated with its actual denotation, it should be mentioned that
expressive meaning, e.g. the negative attitude expressed by derogative terms, is
called connotation by other authors. Indeed, connotations play a role for the semantic
motivation of swear words. For example, the conventional attribute of dirtiness is
the basis of the use of pig or the German equivalent Schwein as an offensive term for
people. But the attitude expressed by the swear word is not part of the meaning of pig
as a mere animal term, nor is the conventional attitude towards pigs identical with
the expressive meaning conveyed by the swear word.
Negative connotations, together with social taboos, are responsible for what
is called euphemisms: roughly, good or indirect terms for bad or taboo things.
Political language is full of euphemisms (just take the vocabulary of warfare),
as are the semantic fields of death (pass away for die) or sexuality (cf. indirect
terms like intercourse, sleep with s.o., make love, etc.) Negative connotations are
also at issue in matters of political correctness. Certain labels like homosexual
are considered discriminatory due to their connotations and replaced by
connotatively neutral terms or terms used by the members of the social group
themselves. But as long as the discrimination itself persists in society, the new
expressions will soon take on the old connotations. Due to this inflationary
process, we are observing a rapid succession of ‘politically correct’ expressions
for certain social groups (such as handicapped being replaced by disabled being
replaced by challenged).
2.6 DIMENSIONS OF MEANING
We have had a detailed look at three dimensions of meaning: descriptive, social and
expressive meaning. In addition, we saw in 2.2.3 that the grammatical meaning of
sentence type is another non-descriptive dimension of meaning: it is neither social
nor expressive meaning. These four dimensions of meaning, if given, are equally
parts of expression meaning. The distinction of dimensions of meaning is orthogonal
to the distinction of levels of meaning which we made in 1.1 (see Fig. 2.4).
38
Understanding semantics
Figure 2.4
Dimensions and levels of meaning
Dimensions
Levels
expression
descriptive social expressive (others)
utterance
speech act
The levels of utterance meaning and communicative sense build on the level of
expression meaning. Therefore, these levels inherit descriptive, social and expressive
meaning (and further dimensions of meaning) from the basic level of expression
meaning (if it is present there). Consider an utterance like this:
(14) ‘Jack, take your damn coat off my seat!’
The sentence carries the descriptive meaning ›addressee take their coat off the
speaker’s seat‹, the grammatical meaning of an imperative sentence, the social
meaning of informal address by the given name and the expressive meaning
conveyed by damn. At the level of utterance meaning, this amounts to an imperative
with informal reference to the addressee and pejorative reference to their coat. The
communicative sense amounts to a rather rude request.
Semantics has always been mainly concerned with descriptive meaning. The
exploration of social and, even more so, expressive meaning is not yet very advanced.
For this reason, these dimensions will not be further elaborated in this book.
Mainstream semantics, in particular in the field of sentence semantics, investigates
linguistic meaning by means of logical analysis (see chapters 7 and 13). Applied to
sentences, it aims at describing their truth conditions and the rules of composition by
which they derive; for content words and complex expressions, it is the denotations
that matter. Widely disregarding other dimensions of meaning, expression meaning
is identified with descriptive meaning, and this in turn is reduced to truth conditions
and denotations.
In this book we take a mentalist approach to meaning: we deal with meanings as
concepts. These concepts determine truth conditions and denotations (as depicted
in the semiotic triangles), but they cannot be identified with them. For as we shall
see in 7.6 and 13.5, it is possible that words with different meanings have identical
denotations, and sentences with different propositions may have the same truth
conditions. There may be conceptually different ways of denoting the same set of
cases or expressing the same kind of situation.
If we want to understand the connection between language and cognition, it is
important to investigate the conceptual level of meaning as this is the interface
between language and thinking. Of course, the dimension of descriptive meaning
Dimensions of meaning
39
is the one that is by far the most differentiated and interesting one; it constitutes
the way in which we represent the world. But if we also want to understand the way
in which the use of language is embedded in social interaction and in which it is
connected to our feelings and emotions, we should spend more effort understanding
social and expressive meaning.
EXERCISES
1. Try to define the descriptive meaning of the following words and compare your
definition to the definition given in a monolingual English dictionary, e.g. Oxford
English Dictionary: fish, milk, red, pregnant, follow.
2. What are the appropriate terms in English for (i) asking somebody to repeat
something you did not understand, (ii) for accepting an apology, (iii) for
answering the phone, (iv) for New Year’s greetings? What kind of meaning do
these expressions have?
3. Try to find three interjections other than ouch and ugh and determine what they
express.
4. Try to find expressives with the descriptive meaning ›mouth‹, ›car‹, ›child‹, ›walk‹,
›work‹.
5. Try to find five examples of modern euphemisms.
6. Try to determine the presently politically correct terms for people who are
politically incorrectly called ‘blind’, ‘black’ or ‘fat’.
7. Discuss the roles of words uttered in communication: how is it possible that
words ‘convey’ meaning?
8. Discuss the connection between descriptive meaning, reference and truth.
9. Most sentences of a natural language like English can be used in different CoUs
for communicating different information, although their meaning remains the
same. How is this possible?
10. Try to determine word by word the descriptive, social and/or expressive meaning
of the expressions in the following dialogue:
A:
Hi, Velma. Are you going to that stupid lecture?
B:
Well, yes. I’m sorry, Sweety.
11. Discuss the relationship between expression meaning and connotations for
words such as pig.
40
Understanding semantics
FURTHER READING
Lyons (1977, ch. 7) on reference, sense (his term for meaning) and denotation.
Lyons (1995, chs 6.6–7) on sentence type and its meaning. Palmer (2001) has a
comprehensive account of non-descriptive sentence meaning. Levinson (1983,
ch. 2.2.5) on social deixis. Brown and Gilman (1960) for a classical study about
pronouns of address. Brown and Levinson (1978) for a general account of social
meaning. Suzuki (1978, ch. 5) for a comparison of forms of address and self-reference
in Japanese and English; Sohn (1999: 9.14) for social meaning in Korean; Cooke
(1968) on the rich choice of terms for speaker and addressee in Thai, Burmese and
Vietnamese; Kummer (2005) on honorifics in Thai. Watts (2003) offers a more recent
general work on politeness from a sociolinguistic perspective. Lyons (1995, ch. 2.3),
Cruse (1986, ch. 12.2) for expressive vs descriptive meaning. Andersson and Trudgill
(1990) and Hughes (1992) for comprehensive studies of swear words in English, also
Bryson (1990, ch. 14) on swearing and taboos in English.
3
Ambiguity
In dealing with meaning in the previous chapters, expressions were treated as though
they had only one meaning (though possibly composed of different dimensions).
This is, of course, not the case. Many, if not most, words have more than one meaning
and even complete sentences may allow for several readings. The technical term for
this phenomenon is ambiguity: an expression or an utterance is ambiguous if it
can be interpreted in more than one way. The notion of ambiguity can be applied to
all levels of meaning: to expression meaning, utterance meaning and communicative
meaning. In dealing with ambiguity, we will first turn to the level of lexical meaning.
This is in accordance with the general bottom-up perspective on interpretation taken
in this book (1.2, 1.3).
3.1 LEXEMES
In 1.2.1 we distinguished lexical meaning from compositional meaning. Lexical
meaning must be learned and stored in our mental lexicon; compositional meaning is
the meaning of composite expressions and is derived ‘on demand’ from the meanings
of the components by applying general semantic rules. Those linguistic units which
carry lexical meanings are called lexemes. Typically, lexemes are single words, but
there are also composite expressions with a special lexicalized meaning. Therefore
‘lexemes’ cannot just be equated with words. Examples of composite lexemes are
so-called idioms like throw in the towel meaning ›give up‹ or fixed adjective-noun
combinations such as white lie, broad bean or little finger. Composite lexemes need
not be idioms like the ones mentioned. Less spectacular cases are particle verbs such
as give up, fill in, look forward to, put on, figure out, etc.
Lexemes are stored in the lexicon of the language, a huge complex structure in
the minds of the language users. Lexical meaning is not to be confused with the
meaning you may find in a dictionary. Dictionaries describe the meanings of their
entries by means of paraphrases. For example, you may find the meaning of bird
described as ‘feathered animal with two wings and two legs, usually able to fly’ (New
Oxford Dictionary of English). In order to understand the description, you have to
know what these other words, feathered, animal, etc. mean. If you look these up in
turn, you will find yet other words used for their description. Dictionaries are in this
sense circular. No matter how carefully they are compiled, they will always contain an
irreducible set of words the meaning of which is not, in fact cannot be, explained in
this way. By contrast, the lexical meanings we have in our mental lexicons are not just
42
Understanding semantics
paraphrases. They are concepts. Whatever these are (this is a question for cognitive
psychology), they are not words.
To a lexical item, there is more than its form and its meaning. Above all, lexemes
are linguistic units within the language system; they can be built into phrases and
sentences according to the grammatical rules of the language. They are assigned to
1
different grammatical categories which differ in their grammatical behaviour. For
example, a noun can be combined with an adjective, and the whole with an article to
form an NP. The NP in turn can be combined as a direct object with a verb to form a
VP, and so on (cf. the discussion of the example in 1.2.3).
In English, many expressions can be used as members of more than one category:
for instance, light is used as a transitive verb (light the candle), a noun (a bright light)
and an adjective (a light colour); walk may be a noun (take a walk), an intransitive
verb (walk in the park) or a transitive verb (walk the dog); too may be a particle
(›also‹) or an adverb (too much), and so on. Although these expressions may be very
similar in meaning, they are considered different lexemes. There are (at least) three
different lexemes light, three lexemes walk and two lexemes too. In general the same
word in different grammatical categories constitutes as many different lexemes.
The members of certain grammatical categories in a language may exhibit
inherent grammatical properties, such as gender in languages like German,
Russian or Latin (see 4.1.1 on gender). The gender of a noun is not a grammatical
form it may take freely: rather, it is an inherent property of the lexeme that may also
determine the form of the article and preceding adjectives (cf. German masculine
der Computer, feminine die Maus (Engl. mouse), neuter das Motherboard). In Bantu
languages such as Swahili, nouns belong to one of five to ten noun classes, e.g. terms
for persons, for animals, for plants or for artefacts. Russian and the Slavic languages
have inherent aspect for verbs, distinguishing between perfective and imperfective
verbs (see 6.3 on aspect).
The grammatical category determines the range of grammatical forms a lexeme
can take. Some categories of lexemes have just one form, e.g. adverbs (here, then),
particles (already, too, only) or prepositions (on, after, without), while the forms of
expressions belonging to other categories may vary. English nouns have a singular
and a plural form (child, children) as well as genitive forms of the singular and the
plural (child’s, children’s). Adjectives have a positive, comparative and superlative
form (light, lighter, lightest or precise, more precise, most precise) and an adverbial
form (lightly, precisely). Verbs exhibit a fairly wide variety of forms which mark
(among other things) grammatical person (sings vs sing), tense (sings vs sang)
and aspect (sings vs is singing). Composite forms such as have been singing are also
considered forms of the main verb sing. For each grammatical category, there are
morphological rules for building the forms in the regular cases. But certain lexemes
may be exceptional. Irregular verbs have special past tense and past participle forms
(sang, sung instead of singed); some adjectives have special comparative, superlative
and/or adverbial forms (better, best, well instead of gooder, goodest, goodly). A few
1 Another term for grammatical categories is parts of speech. We will occasionally use the more
informal term word class.
Ambiguity
43
nouns have non-standard plural forms (child–children, mouse–mice, foot–feet, leaf–
leaves, sheep–sheep). Exceptional forms are also part of the definition of a lexeme.
If a lexeme has different forms, one of them will be used as its citation form or
dictionary form, i.e. the form in which it will be listed in a dictionary or cited when
it is spoken about. Usually it is the simplest form of the lexeme. For example, the nongenitive singular form of a noun is used as its citation form.
Each grammatical form of a lexeme has a spoken form and an orthographic form
(if there is a written standard for the language). Let us use the terms sound form
and spelling, respectively. The sound form of the three grammatical forms kids, kid’s
and kids’ is the same, while their respective spellings differ.
To sum up, a lexeme is a linguistic item defined by the following specifications,
which constitute what is called the lexical entry for this item:
DEFINITION 1 Lexeme
A lexeme is defined by the following constitutive properties:
∑ its sound form and its spelling (for languages with a written standard)
∑ the grammatical category of the lexeme (noun, intransitive verb, adjective,
etc.)
∑ its inherent grammatical properties (for some languages, e.g. gender)
∑ the set of grammatical forms it may take, in particular irregular forms
∑ its lexical meaning
These specifications apply to both simple and composite lexemes. Composite
lexemes too, such as throw in the towel or red light, have a fixed sound form, spelling
and lexical meaning. They belong to a grammatical category (intransitive verb for
throw in the towel and noun for red light), they have inherent grammatical properties
and the usual range of grammatical forms. For example, the grammatical forms of
throw in the towel are those obtained by inserting the grammatical forms of the verb
throw: throws in the towel, threw in the towel etc.
In principle, each of the specifications of a lexeme is essential: if two lexical items
differ in one of these aspects, they are considered different lexemes. There are,
however, exceptions. Some lexemes have orthographic variants, e.g. rhyme/rime,
others may have different sound forms, e.g. laboratory may be stressed on either the
first or the second syllable. The American and the British varieties of English differ
in pronunciation and spelling for many lexemes. As long as all the other properties
of two orthographic or phonetic variants of a lexeme are identical, they will not be
considered different lexemes but lexemes with a certain limited degree of variation.
As far as the component of lexical meaning is concerned, there may be variation of
different extent. There are cases of the same form being associated with completely
unrelated meanings, e.g. bank in the sense of a financial institution and in the sense
given in river bank. These are considered cases of two lexemes which happen to have
the same form. The phenomenon is called homonymy. Sometimes these differences
in meaning come with minor differences in form; for example, the German word
44
Understanding semantics
Bank, too, has two unrelated meanings, namely the financial institution meaning
and the meaning ›bench‹; the plural of the first is Banken, of the second Bänke.
Homonymy contrasts with cases where the different meanings of a form are clearly
related; for instance, body may denote the whole physical structure of a human being
or an animal, or just the trunk; it may denote a corpse, or a group of people working
or acting as a unit (like a physical body). These are considered cases of one lexeme
with different variants of meaning, i.e. as cases with ‘minor’ variation in the meaning
part of the lexeme. This phenomenon is called polysemy.
To sum up, the definition of a lexeme will tolerate, and incorporate, minor variation
in some of its components, while substantial differences for one or more aspects will
give rise to the distinction of different lexemes.
3.2 LEXICAL AMBIGUITY
3.2.1 Homonymy
The adjective light can be used with two meanings. Let us talk of lightA1 if the
adjective is taken as the opposite of dark, and of lightA2 if it is the opposite of
heavy. LightA1 and lightA2 have not always had the same form. LightA1 derives from
a historical source which in German developed into the present-day adjective licht
(meaning, in one of its meaning variants, approximately the same as lightA1). Words
with the same historical origin are called cognates. LightA2 is a cognate of a different
German word, the adjective leicht (›light, easy‹). Due to their different origins, lightA1
and lightA2 are considered two different words by most linguists. In general, different
meanings are assigned to different lexemes if they have different historical sources.
The idea is that, as long as their meanings remain distinct, different words do not
develop into one, even if their sound forms and/or spellings happen to coincide for
2
independent reasons. In addition to lightA1 and lightA2 there is a noun lightN which
is related to lightA1 and means the kind of visible radiation as well as certain sorts of
objects that emit light. A verb lightV is also related to lightN and lightA1.
The two adjectives lightA1 and lightA2 are an instance of what is called total
homonymy: two lexemes share all distinctive properties (grammatical category and
grammatical properties, the set of grammatical forms, sound form and spelling) yet
have unrelated meanings. One would talk of partial homonymy if two lexemes with
unrelated meanings coincide in some but not all of their grammatical forms, e.g. the
verbs lie1 (lay, lain) and lie2 (lied, lied). Partial homonyms can give rise to ambiguity
in some contexts (don’t lie in bed!) but can be distinguished in others (he lay/lied in
bed).
2 It has often been questioned if this historical criterion is really relevant. Average speakers do not
know the origin of the words they use. All that matters for them is some ‘feeling’ as to whether
or not the two meanings have anything to do with each other. For instance, to many speakers of
English the words ear for the body part and in an ear of corn appear to be the same word, although
historically they are of different origins.
Ambiguity
45
DEFINITION 2 Homonymy
Two lexemes are totally homonymous if they have unrelated meanings, but
share all other constitutive properties.
Two lexemes are partially homonymous if they have unrelated meanings, but
coincide in some of their grammatical forms.
Homonymy can be related either to the sound forms of the lexemes or to their
spellings: homonymy with respect to the written form is homography; if two
lexemes with unrelated meanings have the same sound form, they constitute a
case of homophony. The nouns bow1 [b] (cf. bow and arrow; German cognate
Bogen), bow2 [ba] (›front of a ship‹, German cognate Bug), bow3 [ba] (›bending‹;
German cognate Ver-beug-ung) are all homographs, but only bow2 and bow3 are also
homophones. Examples for words that are total homophones but not homographs
would be the noun pairs tail/tale, story/storey or cue/queue. Partial homophones are
numerous: threw/through, write/right, there/their, whole/hole, to/two/too and so on.
In Japanese, homophones are very frequent, while homographs are rare: most
homophones are disambiguated by using different characters. For example, the sound
form seichō [seio:] has more than ten unrelated meanings, each written differently
when written with Chinese characters; among them there are ᡂ㛗 ›growth‹, ᨻᗇ
›government office‹, Ύ⃈ ›clear, serene‹, ኌㄪ ›tone (of voice), style‹, ᡂ㫽 ›adult
bird‹, ᩚ⭠ ›medicine for internal disorders‹, 㟼⫈ ›listening quietly‹ and ṇᮅ
3
›legitimate dynasty‹. Each word is written with two Chinese characters, the first of
which is pronounced sei, and the second chō. If you look at them, each sei character
and each chō character is different. Since Chinese characters carry meaning, each of
the eight two-character combinations has a different meaning. The high frequency in
Japanese of homophones written with different characters is one reason why writing
in Chinese characters is considered a necessity by most Japanese. One might think
that the high number of homophones would cause problems in oral communication.
But if you look at the different meanings of seichō (and similar cases), you will see
that they are so different that they are very unlikely to occur in the same context.
3.2.2 Polysemy
While homonymy is a rare and accidental phenomenon, polysemy is abundant. It is
rather the rule than the exception.
DEFINITION 3 Polysemy
A lexeme is polysemous if it has two or more interrelated meanings.
3 Writings and translations are taken from http://jisho.org/, 15 June 2012.
46
Understanding semantics
4
We will refer to the interrelated meanings as meaning variants. Each of these
meaning variants has to be learnt separately in order to be understood. The
phenomenon of polysemy results from a natural economic tendency of language.
Rather than inventing new expressions for new objects, activities, experiences, etc.
to be denoted, language communities usually opt for applying existing terms to new
objects, terms hitherto used for similar or related things. Scientific terminology is
one source contributing to polysemy on a greater scale. Some scientific terms are
newly coined, but most of them will be derived from ordinary language use. Among
the terms introduced here, lexeme, homonymy, polysemy are original scientific terms,
while others, such as meaning, reference or extension, are ordinary expressions for
which an additional technical meaning variant was introduced. There are, of course,
lexemes which are not polysemous, e.g. the terms for the months or for the days of
the week; these are called monosemous and the phenomenon, monosemy.
As an example of polysemy let us consider the noun light: it means a certain sort
of visible radiation, but also electric lamps, traffic lights or illuminated areas (cf.
light and shadow). Clearly, these meanings are interrelated. Likewise, the different
meanings of lightA2 which correspond to the opposites heavy and difficult are
somehow interrelated although the relation is harder to define. Note that heavy itself,
and with it its opposite light, is again polysemous (cf. a heavy stone, heavy rain, a
heavy meal).
3.2.3 The relationship between homonymy and polysemy
It is only from one particular perspective that homonymy and polysemy present
an alternative: namely, when we ask whether a particular form that is used with
more than one meaning is to be assigned to one or to more lexemes. Apart from this
question, polysemy and homonymy are independent phenomena; two homonymous
lexemes may well at the same time be themselves polysemous. For example, lightA1
and lightA2 are homonymous, and both homonyms are in themselves polysemous.
Different meaning variants of lightA2 are involved in combinations like a light bike, a
light meal, a light exercise or a light breeze. LightA1 has two meaning variants, too, the
opposite of dark and something like ›pale‹ in expressions such as light blue.
The relationship between polysemy and homonymy is illustrated in Fig. 3.1. The
difference is this: in the case of homonymy, multiple meanings are assigned to as
many lexemes; in the case of polysemy, to the same one. Homonymous lexemes
may themselves be polysemous. The fact that the multiple meanings (i.e. meaning
variants) of polysemous expressions are related is indicated by depicting them as
clusters.
3.2.4 Vagueness
There are no gradual transitions between the meaning variants of a polysemous
expression. The meaning variants are distinct. Polysemy therefore is not to be
4 The different meanings of homonyms or of polysemous lexemes are often referred to as ‘senses’.
The term ‘meaning variant’ chosen here is restricted to cases of polysemy.
Ambiguity
47
Figure 3.1
Homonymy and polysemy
Homonymy
lexeme1
Polysemy
lexeme2
lexeme
is realized as
is realized as
sound form/
spelling
sound form/
spelling
means
meaning1
(variants)
means
meaning2
(variants)
means
meaning
variants
confused with flexibility of use. For very many lexemes, their proper application to a
given case is a matter of degree. For example, whether or not we will refer to a child as
a ‘baby’ depends on criteria such as the age of the child and its developmental stage.
Both criteria are gradual. What one person considers a baby need not be considered
so by another person. As a consequence, the denotation (2.2.2) of the word baby
has flexible boundaries. This does not mean that the word baby has infinitely many
meanings that differ in how the borderline is fixed between babies and ex-babies, as
it were. Rather the concept ›baby‹ is in itself vague: it allows for adaptation to the
given CoU.
DEFINITION 4 Vagueness
A lexical meaning is vague if it allows for flexible adaptation to the given CoU.
Vagueness can be observed with all concepts that depend on properties varying on a
continuous scale. Colour terms like red have a vague meaning, because we conceive
the range of colours as a continuum with fuzzy transitions. Whether something is
‘big’ or not, or ‘good’ or not is a matter of degree. In general, all gradable adjectives
(i.e. adjectives with a comparative and superlative form) are vague.
Widespread vagueness in the lexicon should be considered another economic
trait of language. For example, with the pair tall/short, language provides us with
a rough distinction on the scale of body height. This is much more efficient for
everyday communicative purposes than expressions with a more precise meaning,
48
Understanding semantics
say ›between 6 and 7 feet‹. The issue of vagueness and the important role it plays will
be taken up again in 11.5.
Vagueness may occur in combination with polysemy. For example, the meaning
variants of lightA2 are a matter of different underlying scales (of weight, difficulty,
etc.). These scales can be distinguished quite clearly. But for each scale, the meaning
of light describes just a low degree on this scale, whence each meaning variant in
itself is vague.
3.3 COMPOSITIONAL AMBIGUITY
It is not only lexemes that may be ambiguous at the level of expression meaning, but
also complex expressions, in particular sentences. As was stated in 1.2, the meaning
of a sentence is derived in the process of composition and is thereby determined by
its lexical components, their grammatical meaning and the syntactic structure of
the sentence. Each one of these three ingredients can give rise to ambiguity of the
sentence. If a sentence contains an ambiguous lexeme, the process of composition
will yield as many meanings of the sentence as the ambiguous item has, for example
two readings (at least) in the case of sentence (1), since window may also mean a
window on a computer screen. If the sentence contains more than one ambiguous
lexical item, the meanings will multiply.
(1)
Open another window.
As we will see below, not all these meanings will reach the level of utterance meaning.
But strictly speaking all these are possible expression meanings. Since this type of
ambiguity concerns compositional meaning, I would like to call it ‘compositional
ambiguity’.
Independently of lexical ambiguities, the syntactic structure of a sentence may be
ambiguous. Consider the following examples.
(2)
a.
She met the man with her friend.
b.
Flying planes can be dangerous. (Chomsky)
c.
John and Mary are married.
In (2a) the PP with her friend can be related to the verb meet (meaning roughly the
same as she and her friend met the man), or it can be taken as an attribute of the NP
the man (›the man who was with her friend‹). In (2b), the phrase flying planes can
be read as ›flying in planes‹ and as ›planes that are flying‹. (2c) can mean that John
and Mary are married to each other or that they are both married, each to a different
spouse. Such sentences are syntactically ambiguous. Their structure can be analysed
in different ways, and consequently the composition of sentence meaning can follow
different paths. Syntactic ambiguity usually results in semantic ambiguity, i.e. in
different meanings.
Ambiguity
49
Grammatical forms, too, may be ambiguous. For example, the English past tense
forms of verbs have an indicative past tense reading referring to past events along
with a conditional reading with present or future time reference. The form were has
the past tense meaning in (3a) and the conditional reading in (3b):
(3)
a.
I knew you were here.
b.
I wish you were here.
3.4 CONTEXTUAL AMBIGUITY
3.4.1 Interpretation in context
The process of composition yields one or more compositional meanings of the
sentence. When it comes to interpreting words and sentences in their context, i.e.
when one proceeds from the level of expression meaning to the level of utterance
meaning (1.1.2), the expression meanings of words and sentences may be modified.
A sentence actually uttered in a CoU must fulfil certain requirements in order to
qualify as a reasonable message. First, as a minimal requirement, it must not be selfcontradictory, i.e. false in all possible CoUs, because in this case, it cannot be applied
5
to any concrete situation whatsoever. Second, it must in some way be relevant in the
given CoU. These three conditions will be embodied in the Principle of Consistent
Interpretation to be introduced in 3.4.4. Utterance meanings of a word or a sentence
that pass these conditions are called possible readings.
Due to these additional constraints, the set of compositional expression meanings
of the sentence may undergo considerable changes. Three things can happen to a
particular compositional expression meaning:
(a) The expression meaning may be taken over as it is and enriched with
contextual information, e.g. by assigning it a concrete referent.
(b) The expression meaning may be refuted and eliminated if it is contradictory
or does not fit the CoU.
(c) The expression meaning may be modified by some kind of meaning shift in
order to fit the CoU, and subsequently enriched with contextual information.
As we will see right now, this not only holds for sentence meanings but also for
the expression meanings of all components of the sentence, i.e. lexical meanings,
grammatical meanings and the compositional meanings of complex components of
the sentence.
Option (b) may lead to a disambiguation of the sentence at utterance level, i.e. to
a reduction of the number of possible readings. Polysemous lexemes are often such
that their respective meaning variants occur in different types of contexts. As a result,
mechanism (b) allows us to keep the number of lexical forms lower by packing them
5 In 7.2 the term self-contradictory will be replaced by the technical term logically false.
50
Understanding semantics
with clusters of meaning variants that will normally be reduced to a single one when
used in a specific context.
The meaning shifts involved in option (c) create new expression meanings and,
out of them, utterance meanings. For example, when the sentence I don’t need your
bicycle in 1.1.2 was interpreted in the second scenario, bicycle was taken to refer to a
playing card with a picture of a bicycle. This interpretation rests on a meaning shift
of the word bicycle, by which its lexical meaning is replaced with a closely related
new expression meaning. Thus for lexical items, the application of meaning shifts
is another source of ambiguity. The resulting additional meanings only occur in
contexts that trigger the respective meaning shifts. Therefore, this type of ambiguity
shall be dubbed ‘contextual ambiguity’. Meaning shifts follow general patterns which
belong to the semantic repertory of the language users. Three basic patterns will be
described in the next subsection. Meaning shifts constitute another mechanism that
keeps our lexicon economical. Thanks to them, lexemes can actually be used in many
more senses than need to be stored in the lexicon.
In actual communication, however, there may be instances in which the
compositional meaning of a sentence does not fit the context and does not lend itself
to sense modification in order to make it fit. If we assume that our interlocutors are
not talking nonsense, there is a fourth possibility:
(d) The context is revised or enriched as to make the meaning fit into it.
We will return to this issue in 4.7.5, under the label of ‘accommodation’.
3.4.2 Disambiguation
The following sentences contain the ambiguous lexeme letter (›alphabetic character‹
vs ›written message‹):
(4)
a.
Johnny wrote a letter.
b.
Johnny wrote a letter to Patty.
c.
Gamma is the third letter of the Greek alphabet.
(4a) has two readings because the rest of the sentence, Johnny wrote a ____, allows
for both meanings of letter. (4b), however, has only one reading. The addition to
Patty requires the message meaning of the word. Likewise in (4c), the rest of the
sentence would not make sense unless letter is understood as ›character‹. Thus
the immediate sentential environment of a word may call for particular meaning
variants and eliminate others. Along with the meanings of the parts concerned, the
respective compositional meanings of the sentence will be eliminated. A sentence
may also have no possible sensible reading at all if its parts do not fit together. For
example, (5) below is self-contradictory and therefore disqualified at utterance
level. Due to its lexical meaning, the verb shiver requires a subject referent that is
animate and has a body. But the lexical meaning of age does not allow for this sort
of referent.
Ambiguity
(5)
51
Johnny’s age shivered.
Such sentences will hardly be used (except as examples for self-contradictory
sentences, or maybe in poetry). But they serve to illustrate an important point:
observing the rules of grammar and semantic composition does not guarantee that
the result ‘makes sense’.
3.4.3 Meaning shifts
3.4.3.1 Metonymical shift
The following example is borrowed from Bierwisch (1982):
(6)
James Joyce is difficult to understand.
The sentence has at least four readings. If you relate the sentence to James Joyce the
writer, you may first of all take it as meaning that (i) the writings of James Joyce are
difficult to understand. But if you imagine a context where the author is still alive, the
sentence might also mean that (ii) the way he talks, (iii) the way he expresses himself
or (iv) the way he acts is difficult to understand. In the first reading, the name James
Joyce refers to Joyce’s work. In the other readings, it refers to the writer himself. Yet the
proper name James Joyce is not polysemous: we do not have to learn about the lexical
meaning of this particular name to know that these interpretations are possible. In
principle, all names of persons can be used for referring to their published work
(if there is any). The interpretation in context is due to a semantic shift generally
available for all names of people.
Similar shifts are very common. Consider the readings of university in the
following examples:
(7)
a.
The university lies in the eastern part of the town.
b.
The university has closed down the faculty of agriculture.
c.
The university starts again on April 15.
The subject the university refers to the campus in (7a), to the institutional body in
(7b), and to the courses at the university in (7c). Again, this is not a case of polysemy.
The word university lends itself naturally to the meaning shifts that create these
readings. We do not have to know each of them. Many other words with similar
meanings exhibit the same kind of variation: school, theatre, opera, parliament and
so on. The same kind of variation is also paralleled in other languages. This kind of
variation is not rooted in lexical ambiguity. Its source is more general.
If we take a closer look at the meaning shifts involved, we see that in each case the
term the university refers to something that somehow belongs to a university. Let
us assume that the word lexically denotes a certain kind of educational institution.
Such an institution (unless it is a virtual university on the web) must be located
somewhere, it must have an administration, and it must offer courses. It is in this
52
Understanding semantics
sense that its premises, its committees and administration, and the courses offered
belong to the university. Apparently, a term that denotes objects of a certain kind can
also be used to refer to certain things that usually belong to such objects. The term,
then, is felt to ‘stand for’ those things which belong to its referents proper: in (7) the
university in this sense ‘stands for’ the campus, its administration and the courses. In
(6) James Joyce stands for his work. This use of terms is called metonymy: a term
that primarily refers to objects of a certain kind is used to refer instead to things that
belong to objects of this kind.
DEFINITION 5 Metonymy
An expression is used metonymically if it is used to refer to things that belong
to the kind of objects to which the expression refers in its literal meaning.
The corresponding type of meaning shift will be referred to as metonymical shift.
The crucial condition of ‘belonging to an object of this kind’ can be made more
precise if we use the notion of a concept. The word university is linked to a concept
for universities as its lexical meaning. The concept specifies that a university is an
educational institution with premises, teaching, teachers, students, an administration
and so on. A metonymical shift shifts the reference of the word from a standard
referent, a university, to an essential element of the underlying concept. We will get a
clearer picture of metonymy in the chapter on frames (12.3.1).
3.4.3.2 Metaphorical shift
The four sentences in (8) are the opening lines of an article in an American news
magazine (Newsweek, 19 October, 1998, p. 30):
(8)
a.
They were China’s cowboys.
b.
The swaggering, fast-talking dealmakers threw around grand projects and
big figures as if the money would never stop flowing.
c.
Then the sheriff came to town.
d.
Last week Beijing said it was shutting down one of the flashiest investment
institutions, [name of the institution].
The sentences are about Chinese investment institutions and they are full of
metaphorical language. Although there is a literal reading for the first sentence, it will
not be taken in that sense. Rather the next sentence tells us that (8a) refers to certain
‘dealmakers’. We will therefore take the expression China’s cowboys in a metaphorical
sense: the persons referred are not claimed to be cowboys, but to be in some way
like cowboys. In this case, according to (8b), they resemble cowboys in that they
are swaggering, fast-talking and throwing things around. The metaphor is further
developed in (8c) with the appearance of the sheriff, another typical ingredient of a
Ambiguity
53
Wild West setting. Sentence (8d) explains who the ‘sheriff ’ is: Beijing (the name of the
Chinese capital metonymically stands for the Chinese government). This sentence
takes us back from the metaphor to literal interpretations.
Let us define more explicitly what a metaphor is: concepts, notions, models,
pictures for things from one domain, the source domain, are borrowed for the
description of things in another domain, the target domain.
DEFINITION 6 Metaphor
An expression is used metaphorically if it is used to refer to things that are in
crucial aspects similar to the kind of objects to which the expression refers in
its literal meaning.
In (8) the source domain is the Wild West and the target domain is the international
investment scene of China at the time when the article was published. To the majority
of the magazine’s readers, the source domain is better known than the target
domain. Hence, concepts taken from the Wild West domain may help to describe to
this particular readership what’s going on in China. (A Wild West metaphor would
probably be of less help to Chinese readers.) Every metaphor is the construction
of a parallel: the dealmakers are likened to cowboys in certain respects, mainly
their public behaviour, and the Chinese government takes the role of the sheriff in
exerting its authority. In general, metaphorical language can be characterized as
talking about things in the target domain in terms of corresponding things in the
source domain.
A metaphor yields a new concept in the target domain, a concept that is similar
to the original concept of the source domain in that it contains certain elements,
although not all, of the source concept. Metonymy is quite different from metaphor.
When we talk metonymically, we remain within the same domain. We borrow an
element from the original concept, but the links to the other elements remain.
University in the ›campus‹ meaning remains immediately related to university in its
›institution‹ meaning, James Joyce’s work remains related to the person James Joyce.
The relations between the general objects and the things, or aspects, belonging to it
are only possible within one domain.
3.4.3.3 Differentiation
The James Joyce example (6) is relevant in one more respect. The four readings
mentioned differ in the way the verb understand is interpreted in context: it may
relate to the author’s work, to his articulation, his way of expressing himself and the
way he behaves, respectively. Likewise, in the first reading of the sentence, relating to
his work, the verb will be taken in one particular meaning. It is reasonable to assume
that the verb understand in all these cases just means ›understand‹. If we attributed
the different readings of understand to polysemy, we would end up in countless
distinctions of lexical meaning variants of the majority of words. Note, for example,
that understanding a sentence may relate to its articulation when uttered, its syntactic
54
Understanding semantics
structure, its descriptive meaning or its utterance meaning. The different readings can
be better explained if one assumes that to understand means to understand someone
or something in a certain respect that is determined by the context.
The following example (taken from Bierwisch 1982: 11) can be explained in the
same way:
(9)
a.
John lost his friend in the overcrowded subway station.
b.
John lost his friend in a tragic car accident.
c.
John lost his friend, as he could never suppress bad jokes about him.
The common part John lost his friend has three different readings due to the
respective sentence context. In (9a) lose means a loss of contact, in (9b) John’s friend
stops being his friend because the friend no longer exists, and in (9c) the friend is
supposed to live on but stops entertaining a friendly relationship with John. In each
case, the verb lose can be taken to mean something like ›stop having, due to some
event‹. What the context contributes to this is the meaning in which the ›having‹
component is interpreted and the kind of event that causes the loss.
The examples illustrate a third common kind of meaning shift. Bierwisch calls it
conceptual differentiation. In this book, the simple term differentiation is preferred.
DEFINITION 7 Differentiation
Differentiation adds content to a given concept.
It can be defined in general as a meaning shift which results in a special case of what
the expression denotes in its lexical meaning.
There are several more types of meaning shifts, but we will not go further into the
matter.
Table 3.1
Kinds of meaning shifts
Lexical meaning
Metonymy
the university starts in April
›educational
institution‹
Metaphor
→
›courses at the
university‹
they were China’s cowboys
›man who herds
cattle‹
Differentiation
Shifted meaning
→
›person behaving
like a cowboy‹
James Joyce is hard to understand
›perceive the
meaning‹
→
›interpret the text
meaning‹
building a new concept out of an
element of the original concept
building a new concept in the target
domain by borrowing parts of the
concept in the source domain
adding conditions to the original
concept
Ambiguity
55
3.4.4 The Principle of Consistent Interpretation
The driving force of the meaning modifications due to interpretation in context is a
very basic principle:
DEFINITION 8 Principle of Consistent Interpretation
At the level of utterance meaning, a composite expression is always interpreted
in such a way that its parts fit together and that the whole fits the context.
This principle, if appropriately generalized, probably governs all interpretation
whatsoever, because interpretation usually concerns a complex input and is always
interpretation in some relevant context. As we have seen, its application to sentence
interpretation at utterance level may lead to elimination of compositional meanings as
well as to the creation of new ones. The principle generally rules out self-contradictory
readings: they are always due to parts within a sentence that do not fit together (recall
(5)). It also rules out irrelevant readings: these do not fit the wider context.
The notion of ‘context’ here is not identical with the notion of CoU. Parts of a
composite expression, e.g. a sentence, are embedded into the whole expression, and
the whole is embedded into the CoU as well as into larger, more general contexts.
In order to see what is involved, let us return to one of the examples above. The
immediate context of a lexical item is primarily the syntactic phrase it is a part of.
For example, the whole NP a letter to Patty is the phrase context of the noun letter
in (4b). Since the PP to Patty is attached to the noun, it must denote something that
can be directed to somebody; this is the reason why the reading ›written character‹
of letter is rejected. Every further syntactic combination of a sentence component
containing a word embeds it into an additional context level. In (4b), a letter to Patty
is combined with the verb wrote to form the VP of the sentence; since letters in the
sense of written messages are things that are written, the message interpretation of
the NP is accepted in the context of the VP wrote a letter to Patty, and along with it
the respective meaning of letter. When the VP is combined with the subject of the
sentence, in this case Johnny, the reader checks whether this meaning of the VP fits
the expression meaning of the subject. This is a matter of the logical requirements
of the verb write for its subject referent. These requirements are fulfilled in the given
context if the name Johnny refers to a person (rather than a pet or a teddy bear).
In a wider context, the interpretation of the sentence may be modified further. For
example, Johnny might be a four-year-old boy not yet able really to write. He might
be playing that he is writing a letter to his friend Patty. In such a context, the sentence
might be used to describe the situation ‘in quotes’, as it were.
In general, when an expression is built into a sentence, it is combined with other
expressions in accordance with certain syntactic rules, and the result may be
combined with other parts of the sentence until the sentence is complete. Each step
of syntactic combination embeds the original expression in a new, and wider, context,
and each context may impose new requirements on the meaning of the expression.
56
Understanding semantics
In addition, even grammatical forms require certain meanings of the expressions
that carry them, and consequently they may trigger meaning shifts, too. For example,
the English progressive form ‘be + V-ing’ requires a verb that denotes some process
or activity that is going on; it cannot be used with verbs that denote just a state.
Therefore, if the progressive form is used with state expressions such as be polite
(for example in John is being polite), the lexical meaning of the expression is shifted
to something like ›act in a polite way‹, which does not describe a state but some sort
of active behaviour. The context which a sentence provides for an expression can
be called its sentential context. It is the sentential context of the subject in (5) that
eliminates any available meanings and the sentential context of the word letter in
(4b) and (4c) that leads to the elimination of one or the other meaning of the word.
In other cases, the sentential context triggers meaning shifts in order to make all
parts of the sentence fit together. In (7a, b, c) the metonymical shifts of the subject
NP the university are due to the logical requirements of VPs (discussed as ‘selectional
restrictions’ in 4.7.2.2 and 5.7). The metaphorical shifts in (8) come about in different
ways. In (8a) it is the combination with China’s which leads to a reinterpretation of
cowboys (because, supposedly, there are no real cowboys in China); in (8b) threw
around grand projects requires a metaphorical interpretation of threw around in
order to come up with new selectional restrictions that fit the kind of object (grand
projects) given here; the analogue holds for the combination of money and flowing.
In (9a–c) it is obviously the sentential context which leads to the differentiated
interpretations of the VP lost his friend.
Finally, the context of the whole sentence, the context of utterance and the wider
context in which it is embedded, influence the interpretation of the sentence at
utterance level. For example, the actual CoU of the James Joyce sentences in (6)
will determine whether we choose one interpretation or the other. In (8c) the
whole sentence Then the sheriff came to town is taken metaphorically to fit the
extrasentential context; without doing so the context would not provide a referent for
the definite NPs the sheriff and town.
In 1.2 semantic composition was characterized as a bottom-up process, in which
the expression meaning of the whole is derived step by step from the expression
meanings of its elements (lexical items, grammatical forms and syntactic structure).
In other words, the output of the process, the sentence meaning(s), is determined by
the input. When a sentence is interpreted in context, i.e. when its possible utterance
readings are determined, meaning shifts and meaning eliminations interfere with
the process of composition. This interference constitutes a top-down element of
the interpretation: the input may be re-interpreted in terms of appropriate outputs.
Thus, to a certain extent, the output determines the input – at the level of utterance
meaning. It is important to see that the meaning modifications at the interface
of expression meaning and utterance meaning do not contradict the Principle of
Compositionality. Rather, the modifications of meaning that are triggered top-down
by the context guarantee that the bottom-up process of composition works even if
special output is required from the process. When the required shifts of the input
expression (!) meanings are arranged, the process of composition produces its
output in a perfectly regular way. This shows that the rules of composition are not
Ambiguity
57
simply sacrificed if they produce an unwelcome result, but, on the contrary, they are
preserved even at the cost of sacrificing (to a certain extent) the expression meanings
of their input.
3.5 MEANING SHIFTS AND POLYSEMY
3.5.1 Contextual vs lexical ambiguity
The fact that almost all lexemes can undergo meaning shifts in certain contexts
can be considered a further dimension of their abundant ambiguity. However, this
type of ambiguity is fundamentally different from polysemy. In principle, polysemy
is a matter of single lexemes in single languages. Consider the colour adjectives in
English. Many of them are polysemous, with meaning variants not primarily relating
to colour properties. For instance, green may mean ›unripe‹. This is motivated by the
fact that green colour of many fruits indicates that they are not yet ripe. From this,
in turn, derives the meaning variant ›immature‹ due to a metaphor that establishes a
parallel between the development of personality and the process of ripening of fruits.
This meaning variation is an accidental matter of English green. Due to the same
motivations, it might, but need not, occur in other languages provided they have a
word for the colour green. But there is no parallel for exactly this kind of variation
in the case of the other colour words. Although the colour of very many fruits is red
when they are ripe, red cannot mean ›ripe‹ or ›mature‹. Likewise in German, blau
(›blue‹) also means ›drunk‹, but English blue does not; nor does any other colour
adjective in German or English mean ›sober‹. Sometimes, words given as translation
equivalents in different languages may have parallel meaning variants, but usually
their variation will not match.
All this does not hold for interpretations resulting from meaning shifts triggered
by observing the Principle of Consistent Interpretation:
∑ Interpretations that result from meaning shifts triggered by the context need not
be written into the lexicon.
For example, we will not assume that (8c) provides evidence that the lexical meaning
of sheriff exhibits a variant ›Chinese government‹; likewise uses of bicycle such as
in Scenario 2 in 1.1.2 do not force us to assume that bicycle can mean ›playing card
that carries a picture of a bicycle‹. Consequently, we cannot resort to our stock of
lexical meanings in order to arrive at the interpretations that result from metaphor,
metonymy, differentiation or other types of meaning shifts. These shifts can only be
successfully employed in communication if they constitute general patterns available
to all language users.
∑ Contextual meaning shifts are based on general patterns.
These general patterns can be uniformly applied to very many cases; they are
available for all lexical meanings that fulfil the input conditions for these conceptual
operations. The respective types of shifts belong to the interface between expression
58
Understanding semantics
meaning and utterance meaning (recall 1.3.1). They form part of our general
linguistic competence.
∑ The patterns of meaning shifts can be observed systematically with open classes
of lexemes.
The patterns are not even language-specific; the examples discussed in 3.4 can be
literally translated into other languages without getting stripped of the meaning
shifts they undergo.
∑ The same patterns of meaning shift can be observed to operate cross-linguistically.
Psychologists and cognitive linguists consider conceptual operations like metonymy,
metaphor and differentiation to constitute not just semantic faculties, but more
generally fundamental mechanisms of concept formation. This would explain the
fact that the same types of shifts can be observed cross-linguistically.
The phenomenon of systematic meaning variation is currently dealt with in
various frameworks from different theoretical perspectives. You will find it discussed
under labels such as ‘systematic polysemy’, ‘principled polysemy’, ‘facets of meaning’
and ‘microsenses’, as well as ‘coercion’.
3.5.2 Polysemy and meaning shifts
The types of meaning shifts mentioned correspond to types of meaning relations.
For example, the relationship of the shifted meanings of university to the lexical
(‘institution’) meaning of the word is metonymical and the reading ›lose by death‹ of
lose is a differentiation of ›lose‹ in the general sense. The same kind of relationships
exist between the meaning variants of polysemous lexemes: in very many cases
of polysemy, meaning variants are interrelated by way of metonymy, metaphor or
differentiation.
As to metonymy, recall the case of green with its secondary meaning ›unripe‹. Here
green colour is taken as metonymically standing for a certain stage of biological
development. Other cases of lexicalized metonymy are the following:
(10) a.
The asshole did not even apologize. part for the whole, ‘pars pro toto’
b.
He talked to celebrities.
property for a person with the
property
c.
His last date was in a bad temper.
event for person involved
d.
I wrote a paper.
carrier for content
e.
the Green Berets
clothing for wearer
Parts of the body (10a) belong to a person or animal. A property belongs to the
one who has the property (10b): celebrities are persons who have the property of
celebrity. To a date belongs the person one dates (10c). To a piece of paper with
Ambiguity
59
something written on it belongs what is written on it (10d). Pieces of clothing have
their wearers; they are associated with the people wearing them (10e).
The metaphorical use of the verb flow for money that we encountered in (8b) is so
common that it can be considered a lexicalized meaning variant, in addition to the
literal meaning of the flow of liquids. Other examples of lexemes with metaphorical
meaning variants are terms like mouse for a computer mouse, lightA2 in a light meal,
or the majority of idiomatic expressions like throw in the towel, kick the bucket or
make a mountain out of a molehill. Most proverbs are metaphorical, e.g. Birds of a
feather flock together or A rolling stone gathers no moss.
Differentiation too is a common source of polysemy: a lexeme may have a meaning
variant that applies to a special case of what the basic meaning of the lexeme
applies to. Car denotes some sort of vehicle in general, but nowadays preferably an
automobile; glass may mean a kind of material, but one meaning variant is ›glass‹
in the meaning of a certain kind of drinking container made of that material (as in
beer glass).
Although polysemy is a matter of single lexemes in single languages, we see that
the variation of meaning follows general principles. Apparently, we command an
intuition that allows us to recognize that the meaning variants of a polysemous
expression are interrelated. This is evidence for the existence of general patterns in
variation, and these are the very patterns to be found in contextual meaning shifts.
One difference is that the meaning variation encountered as polysemy often
involves more than one meaning shift. A good example is provided by the word film.
Originally it denoted a thin layer or membrane (cf. a film of sweat). From this derives
the metonymic variant that denotes a thin flexible transparent strip of celluloid or
similar material that carries such a layer of light-sensitive emulsion (i.e. the type
of ‘film’ that can be put into a camera, exposed, developed and fixed). A second
metonymy produces the meaning variant ›film (exposed to light, developed and
fixed) that carries photographic pictures‹. If the pictures form a movie, this can again
be called ‘film’, now due to a third metonymy of the type ›carrier medium for the
visible projection‹; this ‘film’ can be projected on a screen and watched in the cinema.
The projection tells a story and this, again, is called a ‘film’ (metonymy level 4). In a
‘film’ in this sense, characters appear and act their parts and a story takes its course.
A fifth metonymy derives the meaning of ‘film’ in the sense of an art or a medium,
a sixth produces the reading ›film industry‹ (go to the film, etc.). Note the amazing
semantic distance between the original meaning present in a film of sweat and the
last variants mentioned.
EXERCISES
1. Which properties determine a lexeme?
2. Find three composite lexemes (idioms) of each of the categories noun, intransitive
verb and transitive verb.
3. What is the difference between homonymy and polysemy?
60
Understanding semantics
4. Discuss the ambiguity of the following words with the meanings indicated: do
they constitute a case of polysemy or homonymy? Try to determine the historical
sources.
a. fraud ›act of deceiving‹ vs ›person who deceives‹
b. calf ›young of cattle‹ vs ›fleshy back part of the leg below the knee‹
c. sole ›bottom surface of the foot‹ vs ›flat fish‹
d. point ›sharp end of something (e.g. a knife)‹ vs ›dot used in writing‹
e. character ›mental or moral qualities of a person‹ vs ›letter‹ (e.g. Chinese
character)
f. palm ›inner surface of the hand‹ vs ›palm tree‹
g. ring ›circular band of metal‹ vs ›telephone call‹
5. What is the relation between the meaning of a word in its lexical meaning and
the word in a metaphorical meaning? What is the relation between a word in its
lexical meaning and in a metonymical meaning?
6. Find out the instances of metaphor and metonymy in the following passage
(from the Newsweek article cited above):
Sound like Asian contagion? So far, China has escaped economic disaster. But
even in China, the mighty can fall. … Can China reform its financial system,
but avoid the social unrest that has crippled the rest of Asia?
7. Find examples where two meanings of a polysemous lexeme are related by
metaphor, metonymy or differentiation (three of each kind).
8. Discuss the meaning shifts underlying the use of bean for the head, paw for the
hand, snotnose for a child.
9. Does the polysemy of so many words constitute an advantage or a disadvantage
for communication?
10. Discuss the difference between polysemy and the variation of meaning due to
metaphorical shift, metonymical shift or differentiation.
11. Discuss the ways in which the Principle of Consistent Interpretation affects the
interpretation of a sentence in context.
12. Think of examples where university is used metonymically for more things
connected to a university than its premises, courses and administration.
FURTHER READING
Tallerman (2011, ch. 2) on lexemes, grammatical categories and their connection with
syntax. Cruse (1986, ch. 2 and ch. 3) on lexemes and ‘lexical units’. Lyons (1995, ch. 2)
and Lyons (1977, ch. 1) on ambiguity. Lakoff (1987) and Ungerer & Schmid (2006)
and Kövecses (2006, chs 7–10) on textbook level, on metaphor and metonymy in
Ambiguity
61
the context of cognitive semantics; Tylor and Takahashi (2011) is a recent handbook
article. For a different account of homonymy, polysemy and contextual ambiguity, see
Croft & Cruse (2004, ch. 5).
4
Meaning and context
This chapter is relatively large. It deals with three phenomena which are normally
1
not treated together in other textbooks on semantics or pragmatics: deixis, the
determination of NPs and presuppositions. These three phenomena are, however,
more closely connected than is commonly recognized. They all contribute to the
indexicality of language and belong to the interface between expression meaning and
utterance meaning. In addition, they are partially embedded in each other: deictic
reference is a special case of definite NP determination; definite NPs are, in turn, a
paradigm case of expressions carrying presuppositions.
PART 1: DEIXIS
We have already seen in 1.3.1 that a sentence taken as such does not tell you what
it refers to. It is only when a sentence is actually used in a concrete CoU that we
establish its reference. You say, for example, ‘This sushi doesn’t taste good!’, referring
at that very moment to a particular dish of sushi in front of you. It is only by concrete
reference that the sentence expresses anything at all: one fixes the reference of the
subject NP, and thereby the predicate of the sentence is applied to an object in the
world and tells the recipient something about it. Fixing the reference is the basis for
any verbal communication.
For efficient communication it is most important that speaker and addressee agree
on what is being referred to. Therefore, the fixation of reference has to be based on
givens that are equally apparent to all interlocutors. The most obvious at hand is
the immediate context of utterance. It consists of things directly and automatically
given for speaker and addressee: who it is who produces the utterance, to whom
it is addressed, at which time and at which place it is produced. At least this is the
case in the standard situation of face-to-face communication. All languages have
means of directly referring to immediate elements of the CoU. Among these means
are pronouns such as I and you or expressions like here and now. Linguists call such
expressions deictic and the general phenomenon deixis.
The use of deictic expressions not only anchors an utterance in the world, it also
imposes the perspective of the speaker on the utterance. The speaker forms the
so-called deictic centre; the ‘I’ is the one who decides who is being addressed; where
1 Pronunciation [dakss], also [dekss].
Meaning and context
63
the speaker is is ‘here’; when they speak is ‘now’. In the first part of this chapter,
we will consider the three most important deictic relations: relation to the persons
involved in the utterance (4.1), to the spatial situation (4.2) and to the time of
utterance (4.3).
4.1 PERSON DEIXIS
Person deixis is deixis based on the linguistic category of person. The category of
person relates to the roles that the participants take in an individual utterance: in
the singular, the ‘1st person’ is the speaker, or producer, of the utterance, S for short;
2
the ‘2nd person’ is the addressee, or A; the term ‘3rd person’ subsumes everybody
3
who is neither first nor 2nd person. These roles are called discourse roles. In most
languages the discourse roles play a role in grammar; these are languages that have
the category of ‘grammatical person’. English is one of these languages: grammatical
person plays a role in the choice of personal pronouns and the forms of finite verbs
(she says; I am; you are; etc.).
The most salient linguistic means of person deixis are systems of personal
pronouns. They may differ in the grammatical categories of person, gender, number
and formality. Table 4.1 displays the personal pronouns in German. They form a
paradigm, i.e. a closed system in which each slot is filled with a particular form. It
often happens that there is the same form for more than one slot (this is known as
syncretism) as is the case for the form sie/Sie in the German paradigm: sie is both 3rd
person feminine singular (‘she’) and 3rd person plural (‘they’); Sie (with capital S)
is the 2nd person pronoun of formal address (recall the discussion in 2.3.1), in both
singular and plural function. It is grammatically 3rd person plural, but semantically
2nd person, since it is a pronoun of address.
Table 4.1
German paradigm of personal pronouns (nominative case)
Singular
Person
Plural
Masc.
1
Fem.
Neut.
ich
wir
du
ihr
Social
meaning
informal
2
formal
Sie
3
er
sie
es
sie
2 You will often find the notions ‘hearer’ or ‘recipient’, but these are imprecise because what matters
is who is addressed. Someone may hear or receive an utterance without being addressed.
3 For plural reference, e.g. with we and you, the correspondence between grammatical person and
S and A is less direct (see the discussion of number at the end of 4.1.1 The meaning of personal
pronouns).
64
Understanding semantics
4.1.1 The meaning of personal pronouns
The descriptive meaning of personal pronouns can be described in terms of
definiteness, person, number and gender; in addition, personal pronouns may carry
social meaning, like German du, ihr and Sie.
Definiteness. All personal pronouns are definite. Their referents are always uniquely
4
determined in the given CoU. I refers to the speaker, YOUsg to the unique addressee,
not just to ›someone addressed‹. HE, SHE or IT is always used for someone or
something certain. This is also the case with plural pronouns: WE, THEY and YOUpl
always jointly refer to a certain group of persons or things. In such cases, too, one
talks of one referent.
Person. For 1st or 2nd person singular pronouns, the descriptive meaning is clear:
they refer to S and A, respectively. For 3rd person pronouns, the only restriction is
that their reference excludes S and A. Since 3rd person pronouns are definite, the
uniqueness of reference relies on the CoU. They are mainly used anaphorically: they
refer to something which has been mentioned before. As they do not provide much
descriptive content (except as regards number and gender), the previous mention
of their referent needs to be immediately before the utterance, if not in the same
sentence. An anaphoric expression is called an anaphor, the kind of reference,
anaphora. The so-called antecedent is the expression in the preceding discourse by
which the referent was previously mentioned. Let us have a look at a sentence and
two possible continuations. Both contain the pronoun it in anaphoric use. In (1a), the
antecedent is the video, in (1b), it is the big tsunami.
(1)
a.
The video shows the big tsunami.
A friend of mine took it himself.
b.
The video shows the big tsunami.
A friend of mine survived it when he was in Malaysia.
In both cases it would be impossible to determine the utterance meaning of the
second sentence without the first. In this regard, 3rd person pronouns differ
fundamentally from 1st and 2nd person pronouns: I and YOUsg can be interpreted
without any preceding text. One only needs to know the immediate context of the
utterance that contains the pronouns. In order to construe an anaphoric 3rd person
pronoun (or any other anaphor) one also needs to know the CoU of the preceding
utterance that provides the antecedent. The only exceptions are sentences that
contain the anaphor along with its antecedent as in (2):
(2)
If you watch this video [antecedent], you’ll never forget it [anaphor].
4 I use italic capitals when I refer to words independent of a particular language. For example, SHE
stands for English she, French elle, Japanese kanojo, and so on. The context will help to decide if I
means 1st person singular pronoun in general, or the English one. For the distinction of singular
and plural YOU we use YOUsg and YOUpl.
Meaning and context
65
Gender. English does not have the grammatical category of gender. The distinction
between he and she relates to biological sex. In (3a) she refers to a female person, in
(3b) he to a male one. In the interpretations that first come to mind, it is a female
5
teacher, or male teacher, respectively, speaking of themselves.
(3)
a.
My teacher said she would be off for a week.
b.
My teacher said he would be off for a week.
The situation is different in gender languages such as Spanish, French, Russian or
German. In German and Russian, every noun has one of three genders, masculine,
feminine or neuter. For most nouns, there is no apparent semantic motivation for
the gender. The only exception is person terms (and sex-specific terms for animals,
like Kater ›tomcat‹ or Stute ›mare‹): nouns for females are feminine and nouns for
6
males, masculine. The converse does not hold: masculine or feminine nouns for
persons do not necessarily refer to male and female persons, respectively. If a noun
that denotes persons is unspecific as to the sex of the person, it may carry any
grammatical gender: Gast ›guest‹, Nachkomme ›descendant‹ are masculine, Person
›person‹, Flasche ›twerp‹ (epithet, lit. ›bottle‹) are feminine, Kind ›child‹ and Mitglied
›member‹ are neuter. These data show that grammatical gender as such does not
have descriptive meaning in a gender language like German. If a person or animal
term specifies the biological sex of a potential referent, this is part of the meaning
of the noun. In such cases, grammatical gender will match with biological sex. This
produces the illusive impression that grammatical gender carries the meaning of
biological sex.
This also holds for personal pronouns in gender languages: gender does not
carry descriptive meaning. When 3rd person pronouns are used anaphorically, their
gender matches the gender of the antecedent NP. For this reason, the gender of a
pronoun can be an important clue for determining the antecedent.
Number. Plural has a different meaning with 1st person pronouns than with nouns.
When one talks of ‘children’, one means several instances of what is described by the
singular form child. The plural multiplies, as it were, the reference of the singular.
However, when you use WE instead of I, you do not refer to several ‘I’s who would
jointly produce the utterance. Rather, WE means: ›S and those who belong to S‹. Who
it is that belongs to S depends on the given CoU. Those belonging to S may have the
status of 2nd or of 3rd person. Similarly, plural 2nd person pronouns may refer to A
and those belonging to A. Let us assume that Mary asks John:
5 Actually, the situation in English is more complicated because there are remnants of grammatical
gender from earlier stages of English, such as feminine pronouns being used for vessels, motor
bikes, or nations.
6 The only exceptions are the neuter nouns Mädchen (›girl‹) and Weib (›woman‹ with an expressive
meaning component). The neuter gender of Mädchen is due to the fact that the word is a
diminutive. Diminutives formed with the suffixes -chen or -lein are neuter, independent of the
gender of the noun they are derived from.
66
(4)
Understanding semantics
What are you doing tonight? We wanted to ask you out for a drink.
Using we she refers to herself plus 3rd persons belonging to her in the given context,
maybe her husband; using you, she may refer to John alone, or to John plus 3rd
persons belonging to him.
This special plural in the case of WE and YOUpl is called associative plural. Plural
2nd person pronouns in addition have a normal plural reading which refers to a
plurality of simultaneously addressed persons. The plural of 3rd person pronouns
is always the normal plural: THEY refers invariably to a set of persons that does not
include the speaker or any addressees.
Social meaning. In addition to their descriptive meaning, personal pronouns may
carry social meaning, for example the meaning of informal or formal interaction (see
2.3.1 and the next subsection).
4.1.2 Paradigms of personal pronouns compared
The languages of the world differ considerably in the structure of their systems of
personal pronouns. The differences concern the categories of person, number, gender
and social meaning.
Person. Some languages have fewer distinctions, some more. There are languages
with only 1st and 2nd person pronouns. On the other hand, some languages
distinguish between ‘exclusive’ and ‘inclusive’ WE: inclusive WE includes A in
addition to S, while exclusive WE refers to S and 3rd persons, excluding A. In (4),
Mary would have to use an exclusive WE; inclusive WE would have been adequate for
a question like ‘What shall we do tonight?’
Number. Almost all languages distinguish between singular and other grammatical
numbers. English only has singular and plural; other languages also have dual,
for reference to two cases, or further numbers for small sets of referents (trial for
reference to three, paucal for reference to a small number of cases).
Gender. Gender distinction is more frequent in the singular than in the plural
and more frequent in the 3rd person than in the 2nd or even 1st person. There are
languages like Hungarian which have no gender distinction at all in their personal
pronoun system.
Social meaning. Many languages have personal pronouns that carry social meaning.
These are most commonly pronouns that distinguish between informal and formal
address, such as German informal du and ihr vs formal Sie. These pronouns often
historically developed as special uses of other personal pronouns. For example,
German Sie was originally a 3rd person plural pronoun (‘they’), which, over the
course of history, also evolved a 2nd person usage. Therefore, in present usage the
plural pronoun form sie/Sie has two functions: 3rd person plural without social
meaning and 2nd person (both singular and plural) with the social meaning of
formality. In general, there are three strategies for using existing pronouns in a
special way for achieving a social meaning of formality:
Meaning and context
67
Table 4.2
Strategies of pronominal formal address
One addressee
More than one
informal
2nd singular
2nd plural
Strategy 3
3rd singular
3rd plural
Strategy P
2nd plural
2nd plural
Strategy 3+P
3rd plural
3rd plural
Strategy 3
3rd person pronouns for formal address, with 3rd person singular for a
single addressee and 3rd person plural for more than one
Strategy P
2nd person plural pronoun for formal address
Strategy 3+P
3rd person plural pronoun for formal address
Strategy 3 is applied in Hungarian and Italian; French and Russian apply Strategy
P, German and Danish Strategy 3+P. Strategy P cannot distinguish between formal
and informal pronominal address if more than one is addressed, since the formal
pronoun of address coincides with the informal 2nd person plural pronoun. The
3+P strategy of German and Danish does not allow any distinction between formally
addressing one or more people.
For a while, English made use of Strategy P: the language featured the informal
2nd singular thou as well as the formal you. In the course of time, the use of formal
you was extended to more and more situations and finally became the appropriate
pronoun of address for all occasions while thou dropped out of use except for
certain religious contexts. The result is the present situation in which there is no
distinction between formal and informal 2nd person pronouns and, concomitantly,
no distinction between singular and plural 2nd person pronouns. Some varieties
of English have developed new 2nd person plural pronouns, e.g. you all or you guys
(Wales 1996: 73). In these varieties, the meaning of you is shifting to 2nd person
singular.
Sometimes pronouns of formal address developed from expressions other than
pronouns. For example, Spanish usted (formal 2nd person singular) is a contraction
of the original vuestra merced (‘your mercy’). Being derived from a common noun,
usted and its plural ustedes is grammatically 3rd person. Spanish can thus be
considered a variant of strategy 3.
Languages without paradigms of personal pronouns. Some languages have
expressions for indexical reference to discourse roles, but lack a rigidly structured
paradigm with a single entry for each case. Typically, the personal pronouns are
more like ordinary nouns. There are a large number of expressions for speaker and
addressees, and these differ strongly in social meaning. These languages include
Japanese, Korean, Vietnamese, Thai and Burmese; Imperial Chinese, too, used to be
among these. Thai speakers choose from about thirty expressions when referring to
themselves. The choice depends on the gender and status of the speakers themselves
68
Understanding semantics
and those of their addressees. For example, there is a special 1st person expression
khâa´phraphúd´thacâaw to be used when addressing the king; the literal meaning is
›majesty’s servant‹.
4.1.3 Person deixis in verb inflection
Person deixis is not only accomplished with personal pronouns. Often verbs are
inflected for person and number and the verb forms form a similar paradigm to the
personal pronouns. If the form of the verb agrees with the subject in number and
person and if a pronoun subject can be omitted, person deixis is exerted by the verb
form alone. This is the case in Spanish, as in many other languages (cf. Table 4.3).
Table 4.3
Form paradigm of the Spanish verb (indicative present active)
beb-o
I drink
beb-emos
we drink
beb-es
you drink (sg, informal)
beb-éis
you drink (pl, informal)
beb-e
he/she/it drinks
you drink (sg, formal)
beb-en
they drink
you drink (pl, formal)
4.1.4 Possessive pronouns
In most European languages there is a parallel paradigm of possessive pronouns
(MY, YOUR, etc.). Other languages do not have extra possessive pronouns; for
example, Japanese uses the personal pronouns with the genitive particle instead,
literally ‘I’s’, ‘you’s’, ‘he’s’, etc. instead of my, your, his. Hungarian has a paradigm of
noun suffixes instead of possessive pronouns (cf. Table 4.4).
Table 4.4
Possessive suffixes in Hungarian
hajó-m
my ship
hajó-nk
our ship
hajó-d
your (sg, informal) ship
hajó-tok
your (pl, informal) ship
hajó-ja
his/her ship
your (sg, formal) ship
hajó-ju
their ship
your (pl, formal) ship
Possessive pronouns and affixes extend person-deictic reference to other things
than the discourse participants themselves. I can only refer to the speaker, but with
MY plus noun one can refer to all sorts of things that belong to the speaker. For
example, the descriptive meaning of my ship or hajóm combines the concept ›ship‹
with the deictic relation to S. The result is the concept ›ship that belongs to S‹.
7
The meaning of possessive pronouns. Possessive pronouns express that the noun
referent belongs to the person or thing indicated by the pronoun. This person or
thing is called the possessor and the thing or person that belongs to the possessor,
7 The following equally applies to possessive affixes.
Meaning and context
69
the possessum (Latin, ›possessed‹). In the case of my ship, the possessor is S and
the possessum is S’s ship. The meaning of possessive pronouns has two aspects: the
specification of the possessor, and the relation between possessor and possessum.
Specification of the possessor. Possessive pronouns determine the possessor in the
same way as personal pronouns determine their referent. This meaning component
is identical with the meaning of the corresponding personal pronouns. Everything
said in 4.1.2 carries over to this part of the meaning of possessive pronouns. It
may involve social meaning along with descriptive meaning. For example, German
Ihr and French votre have the descriptive meaning component of the possessor
being the addressee(s) and the social meaning component of formality toward the
addressee(s).
Relation between possessor and possessum. Talking of ‘possessor’ and ‘possession’
and of ‘possessive pronouns’ makes one think of ownership. However, the relation
between possessor and possessum is by no means always of this kind. ‘My son’, ‘my
leg’, ‘my age’, ‘my mistake’, ‘my name’ are not things which I possess in the ordinary
sense of the word; rather, they are things which somehow can be connected to me.
The meanings of these five nouns describe the potential referents as standing in a
certain connection to a possessor, but the respective relations are very diverse. The
son’s relation to S is a kinship relation; the leg’s relation to S is a part-of relation; the
age is a temporal measure of its possessor; S’s mistake is something S did; his name
is what S is called by. For each of these nouns, a specific relation between possessor
and referent is directly written into its descriptive meaning. Such nouns are therefore
called relational nouns. They not only specify a relation between the referent and
a possessor, their referent also depends on the possessor: in order to determine the
referent of a relational noun, one needs to know the possessor to which it relates.
Of course, there are many nouns which do not specify a relation of their referent
to some possessor, e.g. terms for arbitrary sorts of things, such as ship, dog, bicycle,
chewing gum or stone. Most nouns are of this type. They are called sortal nouns.
They describe their potential referents in terms of what they are like. If one uses them
with a possessive pronoun, this does not necessarily express possession. As we have
noted in 1.1.1, S can use my bicycle to refer to a bicycle which is connected to S in
one way or other: it may be the bicycle which S owns, or uses all the time (without it
belonging to S), it may be the bicycle which S was assigned in some connection, or it
may be the bicycle which S is notoriously talking of. If the noun denotes an article of
daily use, there may be a privileged relation to a particular user. For example, without
a special context, one will assume that ‘my bicycle’ is the bicycle which S uses. But
this is not necessarily so. Ultimately, the relation to the possessor is a matter of the
particular CoU.
We can summarize the meaning component of the relation to the possessor as
follows: the possessive pronoun indicates that the referent of the NP stands in some
relation to the possessor, but it does not indicate which kind of relation this is. If the
possessive pronoun is combined with a relational noun, the relation is provided by
the meaning of the noun. If it is combined with a sortal noun, the relation is a matter
of the CoU.
70
Understanding semantics
These considerations concerning the relation between possessum and possessor not
only hold for possessive pronouns and affixes, but also for possessive constructions
in general, e.g. Anna’s studio or the meaning of possessive pronouns.
4.2 DEMONSTRATIVES AND PLACE DEIXIS
Place deixis relates to the spatial situation in which an utterance takes place. The
points of reference are the location of S and, in some cases, also the location of A.
The immediate categories of place deixis are HERE and THERE. HERE serves as
the reference to the location of S, THERE refers to where S is not. Place deixis is not
restricted to reference to places. Arbitrary objects of reference can be determined
by place deixis. The linguistic means for this purpose are demonstratives such as
English this and that: this denotes something that is where S is (except S herself) and
that refers to something that is where S is not.
When I use English here or this, I only give the information ›where I am‹ or ›thing
that is where I am‹. This is a very vague localization: ‘where I am’ can extend to
a very different radius around myself (depending on context) and there may be
several possible referents within this radius. That is why demonstratives are often
8
accompanied by a gesture of pointing, meant to help the identification of the referent.
4.2.1 Japanese demonstratives
Let us take a look at the Japanese system of demonstratives. It is richer than the
English system (Table 4.5).
Table 4.5
Japanese demonstratives
Close to
speaker
Close to
addressee
Close to
neither
pronominal
kore
sore
are
‘this’
adnominal
kono
sono
ano
‘this’
place noun
koko
soko
asoko
‘this place’
adjective
konna
sonna
anna
‘such’
adverb
k
s
‘so’
Japanese demonstratives distinguish three deictic categories: close to speaker, close
to addressee and close to neither speaker nor addressee. Thus, the deictic orientation
is person deictic. The forms kore, sore and are are nouns; they can be used to refer
to things, not to persons. For demonstrative reference to persons, one uses the forms
8 The terms demonstrative and indexical are of Latin origin, deixis is from ancient Greek. They all
contain the meaning of ‘pointing’.
Meaning and context
71
kono, sono and ano combined with a noun meaning ›person‹ or some kind of person,
e.g. kono hito ‘this person’. The -no forms are combined with nouns like articles; they
are what is called determiners. Demonstratives like kore which can be used as full
NPs are called pronominal, those combined with nouns, adnominal. Koko, soko
and asoko are demonstrative place terms. They are nouns, unlike English here and
there which are adverbs; Japanese koko means ‘this place’, not ›here/at this place‹.
Konna, sonna and anna are adjectival demonstratives which combine with nouns.
Konna hon means approximately ‘a book such as this one here, a book like this’. Kō,
sō and ā are adverbs like English so; ‘kō yatte’ (yatte ‘do it’) would mean ‘do it this
way’. The case of Japanese illustrates two dimensions of systems of demonstratives:
the semantic dimension of deictic relation (e.g. proximity to speaker) and the
grammatical dimension of their function as noun, determiner, adjective or adverb.
4.2.2 Systems of demonstratives compared
Most languages distinguish two or three deictic categories of demonstratives. In
cases where it is two, the distinction is between proximal (close to the deictic centre)
and distal (not close to it). Since the deictic centre is defined as the location of S,
‘proximal’ and ‘distal’ can also be defined as ‘close/not close to S’. English is of this
type: here and this are proximal, as opposed to distal there and that.
Spanish has a system with three levels of distance: proximal este/esta, medial (i.e.
not in the deictic centre, but also not far from it) ese/esa and distal (far from the
deictic centre) aquel/aquella. Thus, its organization is different from the Japanese
system.
German is a rare exception. There are no deictic distinctions for pronominal or
adnominal demonstratives; there is just one form dies(er). This demonstrative can,
however, be combined with demonstrative adverbials hier (proximal), da (distal) and
9
dort (distal): dieses Buch da (›this book there‹). French is quite similar.
4.2.3 The meaning of demonstratives
Demonstratives have three semantic dimensions: deictic relation, type of reference
and definiteness. The deictic dimension has just been discussed. The type of
reference concerns the function of the demonstratives. If they are used as full NPs,
they refer by themselves. In adnominal use, the deictic information is combined
with the descriptive meaning of the noun. For example, the referent of this dog is
conceptually described as a dog by means of the noun, while the demonstrative adds
the information that it is a certain something, close to S. In this way, it is possible
to refer to a particular dog. The function of adnominal demonstratives is thus
analogous to the function of possessive pronouns, where the descriptive meaning of
the noun is combined with a person-deictic relation.
9 The WALS survey on distance contrasts in demonstratives, which covers 234 languages, has only
seven with no contrast, among them German and French (Diessel, chapter 41 in WALS, http://wals.
info/chapter/41).
72
Understanding semantics
Adjectival demonstratives like such or Japanese konna, etc. add to the nominal
concept a particular quality which is determined deictically. Demonstrative spatial
adverbs or nouns refer to a place that is deictically determined.
Definiteness is a common component of all demonstratives. With this book we do
not just refer to any book close to S, but to the specific book close to S, and likewise
for pronominal and spatial demonstratives. Adjectival demonstratives, too, are
definite in referring to the quality deictically indicated.
4.2.4 Anaphoric use of demonstratives
In most languages, demonstratives also have anaphoric uses, others have separate
anaphoric pronouns. In the second sentence in (5), both that and there are used
anaphorically.
(5)
I saw a hedgehog in Frank’s garden the other day. Do you think that hedgehog is
living there?
In anaphoric use, the deictic distinction between proximal, distal or medial is not
relevant in most cases. If a language has a system with two demonstratives, usually
the distal one is used for anaphora; languages with a three-part system use the
medium one. Anaphoric uses of demonstratives loosen the tie to the immediate CoU.
By using anaphora one can refer to things that are not immediately present, but were
mentioned earlier on. Anaphors are bound to the given discourse; they do, however,
constitute an important first step towards a kind of reference which is purely
definite. In fact, definite articles in most languages developed from demonstratives.
For example, the definite articles in Spanish (el/la), Italian (il/la) and French (le/la)
emerged from reductions of the Latin distal demonstrative ille/illa. English the and
German der/die/das, too, derive from demonstratives.
4.3 TIME DEIXIS
Time deixis relates to the time when an utterance is produced, the ‘time of utterance’.
For a CoU, the time of utterance is the present: the time before it is the past, and
the time after it, the future. There are two phenomena in the domain of time deixis:
grammatical relation to time by means of verb tense, and temporal reference by
lexical means such as temporal adverbs, specifications of time, etc. Many languages,
for example Chinese, do not have a real tense system of verb forms; Chinese verbs
need not carry tense, and temporal reference is mainly a matter of the given context.
Tense. The grammatical category of tense will be dealt with in chapter 6.4. Just for
the sake of illustration, consider three examples from English:
(6)
a.
past tense:
She knew the e-mail address.
b.
present tense:
She knows the e-mail address.
c.
future tense:
She will know the e-mail address.
Meaning and context
73
The past tense form knew of the verb to know expresses a state in the past: the state
expressed obtained at a particular time before the time of utterance. The present
tense relates the same predication to the ‘present’ time, i.e. the time of utterance,
while the future tense relates to a particular time after the time of utterance.
Frequently, the tense forms in a language have more than one use. For example, the
past tense forms in English can also be used for counterfactual reference in present
time, and the form will plus infinitive is alternatively used for expressing probability
(of a present state).
Lexical means of time deixis. The central expression of time deixis is NOW. It refers
to the time of utterance. Just as HERE may refer to a place of rather vague extension,
NOW may refer to time intervals of very different length. English has another timedeictic adverb, then, which is distal and can refer to a certain time either in the past or
in the future. There are a number of adverbs that enable more fine-grained temporal
localization: just, recently, formerly, once, at once, soon, later, etc. Other adverbs refer
to days: today, tomorrow, yesterday; ‘today’ is the day which contains the time of
utterance, ‘tomorrow’ the day after today, etc. There are also adnominal expressions
of time deixis, such as next, last, former or future. When combined with a noun, they
add a temporal localization of the referent.
Stopover
We have taken a look at the central means of deixis. They serve to bring about
reference by relating the utterance to immediate elements of the CoU. These
elements are those things that are automatically given with any utterance in
face-to-face communication: speaker, addressee(s), the time of utterance and
the place where it is produced. Part of the deictic expressions serves the direct
reference to these elements. These expressions are independent parts of the
sentence with their own reference, e.g. pronouns or adverbs. Along with these
independent expressions, there are adnominal deictic means. They allow us
to combine the descriptive meaning of a noun with a deictic indication of the
referent. Table 4.6 gives a short survey of pronominal/adverbial and adnominal
deictic expressions.
Table 4.6
Pronominal and adnominal deictic expressions
Kind of deixis
Pronominal/adverbial
Adnominal
person deixis
I, you, …
my ship
place deixis
here, there
this, that
this ship
time deixis
now, tomorrow
the next ship
74
Understanding semantics
If we were confined to deictic reference, we could only talk about things
which are immediately given in the CoU. With anaphoric use of demonstratives,
we can expand our radius of reference considerably. Still, in order to free
ourselves from the limits of the given context of discourse, we need more
general linguistic means of anchoring utterances. The most important one is
general definite reference, for example marked by definite articles. As we shall
see, definite reference still anchors the utterance in the CoU, but only in a very
general way. A definite article only conveys the message ‘the referent is uniquely
determined in this context.’
We will proceed by further pursuing the adnominal track. In addition to
the deictic adnominals – demonstratives and possessives – more general
determiners will enter the stage: definite and indefinite articles and so-called
quantifiers.
PART 2: DETERMINATION
If one wants to use a noun for reference, one has to form an NP with it. There are
various ways to do this, and to indicate thereby the kind of reference intended. In
many languages one can choose the grammatical number of the noun, e.g. singular or
plural, expressing if one refers to one or more than one instance. In many languages
the noun can be combined with a definite or indefinite article to indicate whether the
referent is something given beforehand or not. The noun can also be combined with
other determiners: with demonstratives, possessive pronouns or elements such as
each, every, all, both or no. It is also possible to use an NP without concretely referring
to something given in the CoU. In all the examples we have considered so far, the NPs
referred concretely (to a particular dog, a skirt, a bicycle, etc.), but NPs can also be
used for relating to general cases without talking about particular things given in
the context. (7) is an example of this. The NP books does not refer to any particular
books. This use of NPs is called generic.
(7)
Books are getting cheaper and cheaper.
These aspects of NPs are subsumed under the notion of determination. The
following sections discuss the core phenomena: definiteness and indefiniteness (4.4),
quantification (4.5) and genericity (4.6).
4.4 DEFINITENESS AND INDEFINITENESS
4.4.1 An example
Let us assume Sheila did not say (8a) to Mary, but (8b) instead:
Meaning and context
(8)
a.
The dog has ruined my blue skirt.
b.
A dog has ruined my blue skirt.
75
When you compare the two sentences, it is easy to see the difference between the
definite and the indefinite article. By using the definite article, Sheila indicates that
she is referring to a particular dog given beforehand in the context of the utterance,
to the dog. For both Sheila and Mary, this is the only dog that comes into question in
the context described, the family dog Ken. The definite article points to this context.
It is indexical.
If Sheila had used the indefinite article, she would have expressed that she was
not talking about a dog which is uniquely determined beforehand. The addition
‘beforehand’ is crucial: objectively, the dog is uniquely determined by having ruined
Sheila’s skirt; but what matters for the choice of the article is the fact that this dog
comes into play between Sheila and Mary only with this very utterance. Before that,
the creature did not belong to the common context of the two of them. After having
said (8b), Sheila is able to talk anaphorically of ‘the dog’, continuing, for example, in
the following way:
(9)
A dog has ruined my blue skirt. It happened when I was walking through the
park. It suddenly came running from the back and snapped at me. I was really
frightened. I had no idea who the dog belonged to.
The example illustrates a frequent interplay of indefinite and definite article: first
a referent is introduced in the discourse with an indefinite article, later the same
referent is taken up with a definite article. Correspondingly, it is often said that the
function of the indefinite article is to introduce new referents into the discourse,
and that the definite article serves to anaphorically refer back to referents already
introduced. These descriptions of the functions of the articles are, however, much
too narrow.
4.4.2 The meaning of the definite article
The combination of a noun with the definite article is called a definite description.
A definite description can contain additional material such as adjectives or relative
clauses attached to the noun: the blue skirt, the skirt on the chair or the skirt that the
dog has ruined. In a definite description, the article is combined with material that
provides semantic information about the referent: the referent is a skirt, it is blue, it is
on the chair, it has been ruined by the dog, etc. What the definite article itself adds to
the description of the referent is the information that the given description is unique
in the given context: there can be only one thing it applies to.
The crucial question, then, is this: when does a definite NP constitute a unique
description in the given context? There are essentially two possibilities: the uniqueness
can be due to the very meaning of the NP, independently of the particular context,
or it can be due to the special circumstances in the given CoU. In the first case, we
have what is called semantic uniqueness, in the second, pragmatic uniqueness
76
Understanding semantics
The definite description the dog in the fifth sentence of (9) is pragmatically unique.
The preceding sentences create a special context in which one particular dog is
introduced and plays a prominent role. Without this context, the referent of the dog
would not be uniquely determined.
DEFINITION 1 Semantic and pragmatic uniqueness
An NP is semantically unique if the description of its referent is unique
independently of the given CoU.
An NP is pragmatically unique if the description of its referent is unique only
in the special circumstances given in the CoU.
4.4.2.1 Semantic uniqueness
In order to illustrate the notion of semantic uniqueness, we will take a look at a
newspaper text. The text is a translation of parts of a news article in the German daily
Tageszeitung from 17 October 2011. The article relates to the debate at that time about
the succession of the king of Saudi Arabia.
(10) […] ‘The king is suffering from chronic back pain,’ said one Western diplomat.
[…] The monarch’s health and age recently gave rise to speculations about his
succession. In Saudi Arabia, the entire ranks of the leadership are about eighty
and are all ill, out of the country or unable to govern.
10
The noun king is relational, the possessor being the respective kingdom. Since for
a country, if it has a king at all, it has only one, the description king is semantically
unique. Uniqueness of reference is built into the meaning of the noun king. One
therefore says: the noun king is inherently unique. Since the news article is about
Saudi Arabia and the country had a king when the article appeared, the possessor
of the king (the kingdom) is uniquely determined and with it the king at that time.
The noun monarch, however, though of a related meaning, is neither relational
nor inherently unique, but plainly sortal. The concept ›monarch‹ comprises all
queens, kings, empresses, tsars, pharaohs, etc. of all times and all countries under a
sortal notion of a particular kind of person, namely one who solely rules a country
(under certain conditions). The noun monarch is not relational because its potential
referents can be determined without fixing a possessor. It is not inherently unique
because in a given CoU there may be many monarchs (or, by chance, just one, or none
at all). For this reason, the description the monarch in the text is not semantically
unique, but pragmatically. In (10) the noun can be used with the definite article here
10 Note that the possessor for a relational noun like king is what appears in the of phrase after the
word: in King of Saudi Arabia, Saudi Arabia is the possessor of the king, in the grammatical sense
(rather than the king being the possessor of the country).
Meaning and context
77
only because a particular monarch, the king of Saudi Arabia, was established in the
context of the article (and no other monarch).
Like king, the nouns health and age (in the meanings given here) are relational
and inherently unique. Health and age are always health and age of some possessor,
and a given possessor, at a given time necessarily has one health and one age only. If
the possessor of health or age is uniquely determined, so is the single referent of the
nouns. The last definite description in the article, the entire ranks of the leadership, is
also semantically unique: for a given country or other institution, there is only one
group of persons which forms the leadership ranks.
Most inherently unique nouns are relational. The most frequent sub-types are role
terms, part terms and terms for all kinds of attributes. Role terms encompass nouns
such as king, president, director, etc. They denote the head of some organizational unit
which is their possessor – a country, a company, a university, etc. Other role terms
include words like mother or boss: the possessor is a person, and for a given possessor
the referent is a uniquely determined other person (usually a person has only one
mother and at most one boss). Examples of inherently unique part terms are bodypart terms for parts of which there is only one: head, nose, mouth, throat, back, etc.,
or terms for unique parts of objects: mouthpiece, lid, bottom, surface and so on. Terms
for attributes denote various aspects, or dimensions, by which their possessor can be
described: size, weight, width, structure, shape, colour, meaning, price, name, content,
character, age, profession, sex, etc. Inherently unique relational nouns are called
functional nouns. They are so called because there is a function in the mathematical
sense which assigns a referent to every possible possessor: her mother, her back, her
address, her height, etc. Relational nouns, in the narrow sense of the word, are the
remaining relational nouns that are not inherently unique. Relational nouns, too,
include role terms (friend, neighbour), part terms (finger, tooth, side) and attribute
terms (attribute, property, feature), the crucial difference being that for one possessor
there may be no, one or more referents.
There are also non-relational nouns that are inherently unique. They denote
‘institutions’ in the world, in the broadest sense: things of which – in a given type
of context – there is only one, playing a particular role in that context. Such nouns
include sun (in the sense of the sun of Earth), moon (in the corresponding sense),
pope (the head of the Catholic Church), weather (in a given CoU, which includes a
particular place and time, there is only one weather), date (day, month and year), etc.
Inherently unique non-relational nouns are called individual nouns.
The four types of nouns mentioned so far are arranged in Table 4.7. ‘Relational
nouns’ are to be taken in the narrow sense of the word.
Table 4.7
Types of nouns
not inherently unique
inherently unique
non-relational
sortal nouns
boy, dog, skirt, rock
individual nouns
pope, sun, date, weather
relational
relational nouns
son, foot, property
functional nouns
father, head, size
78
Understanding semantics
Sortal or relational nouns can be turned into inherently unique concepts by adding
certain expressions, such as superlatives (the fastest notebook), ordinal numerals and
similar expressions (the third/next/last chapter), or appositions that are themselves
inherently unique (cf. the year 1984, my daughter Emma, the word akimbo). The
nouns notebook, chapter, year, daughter and word are not inherently unique, but
when complemented by these adjectives and appositions they represent semantically
unique descriptions.
There are things or persons which are established in certain social contexts and
play a unique role there. For these, often functional or individual terms exist: for ‘the
dean’ of a faculty, ‘the driver’ in a bus, the ‘Mum’ in a family. But very often we use
just sortal nouns in such cases, in particular if the social context is rather restricted:
we talk about ‘the bus stop’, ‘the pizza shop’, ‘the playground’, ‘the pharmacy’ in a
neighbourhood, ‘the kitchen’, ‘the postbox’, ‘the dog’, ‘the car’ in a household, ‘the
waitress’, ‘the menu’, ‘the meal’, ‘the bill’ when we visit a restaurant. In all these cases,
the definite article can be used without establishing unique reference beforehand.
The referent is established in the given type of context and identified by the function
or role it plays.
When one uses an NP with an inherently unique noun or with a sortal noun in
a context where it denotes something with a unique role, its referent is uniquely
determined beforehand. The definite article only confirms this fact. Nevertheless,
the definite article is not superfluous: it is indexical in that it points to the given type
of context. It is the CoU that ultimately determines the respective referent of the
definite description: ‘the pope’ is the pope in office at the time referred to, his ‘health’
is the state of health at that time, ‘the weather’ is the weather at the given time and
place, ‘the dog’ is the dog presently kept in a household, ‘the playground’ is the kids’
playground in the neighbourhood where they live, etc.
4.4.2.2 Pragmatic uniqueness
Pragmatically unique definite NPs only refer uniquely due to the particular context
they are used in. This uniqueness, too, is established in a social context – in the current
discourse – but its lifespan is restricted to this situation. Once the discourse is over, or
has moved on, the ground for unique reference is no longer given. One way to create
pragmatic uniqueness is by deictic use of definite descriptions. Unlike adnominal
demonstratives, the definite article does not carry a deictic differentiation, but it still
indicates that the reference of the NP is determined by the CoU. For example, if you
encounter somebody on a walk and there is a dog around, you can say to her:
(11) Is the dog yours?
You may accompany your utterance with a gesture pointing to the dog you mean. The
dog may also be in the focus of attention of the interlocutors anyway, in which case
a gesture is not necessary.
A further case of pragmatic uniqueness involves definite NPs in anaphoric use. An
example is the dog in Sheila’s story in (9). Its antecedent is a dog in the first sentence.
Meaning and context
79
The three NPs a dog, it in the third sentence, and the dog all have the same referent;
they are coreferent. Sheila sets up the referent as a character in her story and thereby
establishes it in a unique role. In the news article in (10), the monarch is an anaphor
with the antecedent the king. The referent of the anaphor and its antecedent was
introduced previously in the specific context of the article.
4.4.2.3 A test for pragmatic vs semantic uniqueness
There is a simple test for distinguishing between semantic and pragmatic uniqueness.
In the case of pragmatic uniqueness, the definite article can be replaced by an
adnominal demonstrative. This is due to the fact that the article in these uses has
the demonstrative function of pointing to the immediate CoU. Replacement by a
demonstrative is not possible with semantic definites. One would not say, ‘I’ll take
this dog for a walk’ when just referring to the family dog, or, ‘This king of Saudi
Arabia is already old and frail’. At least, the use of the demonstrative would require
a special context, or express a special attitude. By contrast, Sheila in her story in (9)
could equally well say: ‘I had no idea who this dog belonged to’, and the news article
in (10) might have a demonstrative with monarch: ‘The health and the age of this
monarch recently gave rise to speculations about his succession.’
4.4.2.4 The range of application of the definite article
As this test shows, the range of application of definite descriptions overlaps with the
one of NPs with adnominal demonstratives. Both have deictic and anaphoric uses.
In the case of deictic demonstratives, the deictic content and/or a gesture of pointing
helps to identify the referent uniquely. Deictic and anaphoric uses essentially make
up the range of pragmatic uniqueness. In the domain of semantic uniqueness,
however, only definite descriptions can be used (Fig. 4.1). This is the basis for the
test just described. For demonstratives, the anaphoric use is an extension of their
original deictic use. Definite descriptions, in turn, extend unique determination to
semantic uniqueness.
Figure 4.1
Uses of demonstratives and the definite article
Demonstratives
deictic
anaphoric
pragmatic uniqueness
semantic uniqueness
definite article
We can fix the following as the general function of the definite article:
80
Understanding semantics
DEFINITION 2 Function of the definite article
(i) The definite article indicates that the NP refers to something uniquely
determined in the given context, either in the particular or in a wider
context.
(ii) The definite article is indexical in relating to the relevant context in which
the referent is uniquely determined.
4.4.3 Singular and plural, count and mass nouns
In order to better understand NP reference, we need to have a look at the function
of singular and plural and the distinction between count nouns and mass nouns.
These distinctions are not relevant for understanding definiteness, but they will play
a central role for indefiniteness, which we will deal with next.
Singular and plural. Most English nouns can be used in the singular and in the
plural. For a noun like skirt, the singular form skirt refers to a single object, the plural
form skirts to more than one. These multiple instances form one referent; a statement
like (12) refers to a complex referent consisting of several skirts:
(12) These skirts have become too tight.
Consequently, the denotation of the singular noun skirt is different from the
denotation of plural skirts. The denotation of skirt is the set of all single skirts,
while the denotation of skirts is the set of all cases of more than one skirt. In formal
semantics, such cases are called ‘sums’ or ‘groups’.
Count nouns and mass nouns. Count nouns are nouns which can be used in
the plural, and in most cases, with numerals; apart from a few exceptions (e.g.
pants), they can also be used in the singular. By contrast, mass nouns are used
in the singular to refer to an unspecified quantity of something: juice, flour, air,
garbage, metal, clothing, furniture are mass nouns, as are abstract nouns such as
anger, love, heat or grudge. Several instances of ‘apple’ are referred to as ‘apples’, but
several instances of ‘juice’ may form just one case of ‘juice’. One can therefore say, for
example, that the guests at a party ate ‘many apples’ (plural) and consumed ‘much
juice’ (singular). The denotation of a mass noun encompasses single instances of
juice as well as collections of several such instances.
The difference between mass nouns and count nouns is due to the fact that their
meanings are concepts of different type. The meaning of the count noun skirt is a
concept which clearly delineates its potential referents; it provides a criterion for
what is one skirt: one skirt is a single, whole skirt. Neither does a part of a skirt
constitute an instance of a skirt, nor do several skirts form a case of one skirt.
By contrast, a mass noun juice describes its potential referents in a way that is
appropriate for distinguishing juice from other kinds of stuff, but does not yield a
criterion for what would constitute one potential referent. It is not possible to decide
Meaning and context
81
if some quantity of juice is one case of ‘juice’, multiple cases or just part of one case.
Mass nouns do not ‘individuate’ their potential referents.
A special sub-type of mass nouns, so-called aggregate mass nouns, comprises
nouns like literature, furniture, equipment, staff, etc. Aggregate mass nouns collect
different kinds of items into one concept. For the single items there may be count
noun terms, such as chair, table and bed for items of furniture. Aggregate mass
nouns leave open how many items are referred to. You can add items to a referent of
furniture, or remove items, and it still remains a possible referent of furniture. Often
there are pairs of plural count nouns and aggregate mass nouns with roughly the
same meaning: clothes and clothing, shoes and footwear or employees and staff.
Since count noun concepts define clearly delineated single cases, these cases can
be counted, and count nouns can therefore be combined with numerals and used in
the plural. Mass nouns lack the preconditions for these uses.
Actually, mass nouns are sometimes used in the plural, but then they take on the
meaning of count nouns. ‘Juices’ are different kinds of juice (kinds are countable)
and ‘beers’ are servings of beers, or kinds of beer. Without being reinterpreted as
count nouns, mass nouns can only be used in the singular and not with numerals.
They combine with vague quantity specifications such as much, a little or more. For
a precise specification of quantity they must be combined with a measure unit word
such as litre, kilogram, etc. Unit words form count expressions with mass nouns –
litre juice, kilogram flour – which can be combined with numerals.
4.4.4 Indefinite NPs
English has an indefinite article a(n), but only for singular count indefinites. No
article is used for simple mass or plural indefinites. Sometimes they are said to
contain a ‘zero article’, but where there is nothing there is also no zero article. Thus,
one should prefer to talk of bare mass nouns and bare plurals. Consider the
following example:
(13) a.
I had an apple and __ strawberries and __ orange juice.
In French, indefinite NPs with mass or plural nouns are marked explicitly, with a
combination, or fusion, of the partitive particle de and the definite article forms le/
11
la/les. The French equivalent of (13a) is this:
b.
J’ai mangé une pomme et des fraises et j’ai bu du jus d’orange.
formally: ‘I have eaten an apple and of the strawberries and I have drunk of
the orange juice.’
The French indefinite markings for mass and plural NPs contain the definite article,
but this is not reflected by their meaning. The referent of the NPs need not be part of
11 The respective forms are: du > de le (masculine), de la (feminine) and des > de les (plural).
82
Understanding semantics
a quantity given beforehand; des fraises does not mean ›some of the strawberries‹, as
the formal one-to-one translation suggests but just ›(some) strawberries‹.
In colloquial English, there may be an indefinite article for mass and plural nouns
presently evolving: a reduced form of some, often described as sm in the semantic
literature and pronounced [sm/sm]. It is to be distinguished from full some [s m],
which constitutes a quantity specifier (see 4.4.5 and 4.5.2)
The meaning of indefinite NPs. The meaning of the indefinite article is the same
as the lack of an article, or sm, with mass nouns and plurals: it indicates that the
referent of the NP is not given beforehand. It is not uniquely determined, but only
given a qualitative characterization.
4.4.5 Definite and indefinite NPs in general
Definiteness and indefiniteness are not tied to the presence of definite or indefinite
articles; some of the cases have already been discussed. In English, the following are
definite NPs:
DEFINITION 3 Definite NPs
a. NPs with
- a definite article (definite descriptions) – the cup
- adnominal demonstratives – that cup
- possessive pronouns – her cup
- Saxon genitives – Fred’s cup
b. proper names
c. pronominal demonstratives
d. personal pronouns
NPs with a possessive pronoun or a Saxon genitive are only definite if they are not
used predicatively. In (14a) my brother and Kate’s brother are in predicative use; the
NPs do not refer and are not definite. (14a) means the same as ›Ben is a brother of
me/ a brother of Kate‹. In (14b), the respective NPs are used referentially; here they
mean the same as ›the brother of me/of Kate is coming for a visit‹.
(14) a.
b.
Ben is my brother/Ben is Kate’s brother.
My brother/Kate’s brother is coming for a visit.
Indefinite NPs, too, comprise more than the types just discussed:
Meaning and context
83
DEFINITION 4 Indefinite NPs
a. simple indefinite NPs: singular count nouns with indefinite article; bare
mass nouns, bare plurals; mass nouns and plural nouns with sm.
b. count nouns with quantity specifications:
numerals, no, many, (a) few, some [s m], any, several, …
c. mass nouns with quantity specifications:
measure specifications 3 litres of, no, much, (a) little, some [s m], any, …
d. indefinite pronouns: somebody/-one/-thing, no/-one/-thing, anybody/-one/thing, etc.
4.4.6 Articles across languages
Definite or indefinite articles, or equivalent markings on nouns, do not exist in all
languages. Some languages only mark definiteness, some only indefiniteness, many
do not mark either. Russian and most of the other Slavic languages have no articles,
likewise Japanese. Turkish and Chinese mark only indefiniteness (in certain cases);
Irish only definiteness. In addition, languages differ considerably in when they make
use of the markings available. For example, in Modern Greek and colloquial German,
personal proper names are used with the definite article, but not so in English and
written standard German. If a language does not mark definiteness, there will be
other ways to indicate unique reference. The most important means is sentence
structure. In Chinese, for example, a language with SVO word order, it is essentially
assumed that the subject has definite reference while the object is indefinite. If these
conditions are not given, special constructions are used.
4.5 QUANTIFICATION
4.5.1 Quantifier, domain of quantification and quantified predication
In theoretical semantics, no NP phenomenon has been studied as deeply as so-called
quantification. Some theories consider all types of determination as instances of
quantification, including definite, indefinite and demonstrative determination.
However, I consider it important to apply a more restricted notion of quantification,
in order to properly understand the phenomenon. Let us have a look at an example;
the sentences are to be construed coherently:
(15) a.
Next week there will be a big birthday party at John’s place.
b.
Sue is going to invite nine kids.
c.
Each kid is to get a present.
d.
For some kids, John has not got anything yet.
e.
No kid will be disappointed.
84
Understanding semantics
In the second sentence, certain ‘kids’ are introduced in the discourse, those nine kids
that Sue is going to invite. From this point on, one could refer anaphorically to these
kids as ‘the kids’. In (15c) these kids are referred to: each of them is to get a present.
The reference to ‘the kids’ is implicit, though. But we might as well have the following
in (15c, d, e):
(16) c.
Each of the kids is to get a present.
d.
For some of the kids, John has not got anything yet.
e.
None of the kids will be disappointed.
Such constructions with a pronoun and an of PP following are called partitive
constructions. The fact that in (15c, d, e) the three NPs each kid, some kids and
no kid can be replaced by partitive constructions reveals that they involve definite
reference to ‘the kids’. They relate to a given total of cases which could be referred to
with a definite plural NP.
The three NPs mentioned are quantifying NPs. All quantifying NPs relate to a
total of cases, the so-called domain of quantification (DoQ for short). In (15c, d, e),
the domain of quantification is formed by the nine kids introduced in (15b).
A quantifying NP is combined with a predication (usually the rest of the sentence)
about the single cases in the DoQ. If we use the variable x for the single cases in the
DoQ, the predications in the three sentences are: ‘x is to get a present’, ‘for x, John has
not got anything yet’ and ‘x will be disappointed’. The predication is simply obtained
by replacing the quantifying NP with a variable. Let us call the predication the
quantified predication.
Now, the contribution of the determiners each, some and no is the following: they
quantify the number of cases within the DoQ for which the predication is true. This
is why they are called quantifiers. In the case of each, the predication is true without
exception, in the case of some it applies to an unspecified part of the DoQ and for
no it is not true for any case in the DoQ. Table 4.8 displays the three components –
domain of quantification, quantified predication and the quantification expressed by
the quantifier – for the examples in (15).
Table 4.8
Components of quantification
Domain of
quantification (DoQ)
Quantified predication
(QP)
Quantification:
the QP is true …
(17c) the nine kids
x is to get a present
for each x in the DoQ
(17d) the nine kids
for x, John has not got
anything yet
for some x in the DoQ
(17e) the nine kids
x will be disappointed
for no x in the DoQ
What is at issue when quantification is applied is the fact that the quantified
predication might be true for some cases within the DoQ and might be false for
Meaning and context
85
others. Using each, every or all, we exclude that the predication is false for any cases;
by using no, we exclude that it is true for any cases; some expresses that there are
cases within the DoQ for which the predication is true – leaving open whether there
are also cases where it is false.
In English there are four determiners which are exclusively used as quantifiers:
each, every, all and both. In addition, certain forms of NPs with other prenominal
elements can be used for quantification, but only under certain conditions, as we
will see next.
4.5.2 Quantification and indefinite NPs
Indefinite NPs are sometimes quantificational, sometimes not. There is a simple
criterion to determine whether a given indefinite NP in a particular utterance is
quantifying or not: it is quantifying iff it is explicitly or implicitly partitive. Let us
return to the examples in (15). As we have seen in (16), the three quantifying NPs
in (15c, d, e) can be replaced by partitive constructions; they are therefore implicitly
partitive. The five sentences in (15) also contain two indefinite NPs that do not
quantify: nine kids in (15b) and a present in (15c). In the context assumed here, they
cannot be interpreted as implicitly partitive and therefore they cannot be replaced by
explicit partitive constructions (the sign § marks semantic inadequacy):
(17) b.
§ Sue is going to invite nine of the kids.
c.
§ Each kid is to get one of the presents.
In the case of (15c), a straightforward partitive paraphrase is not even grammatically
possible; the indefinite article would have to be replaced by the numeral one.
It is not a coincidence that the quantifying NPs in (15) have partitive paraphrases,
while the others do not. NPs which quantify over, and thereby refer to, a certain DoQ
can always be paraphrased with a partitive construction. Conversely, a (referential)
NP which can be construed as partitive always produces quantification. The definite
NP within the partitive paraphrase explicitly refers to the DoQ. Selecting some of the
cases of the domain has the effect of quantification: the predication about the NP is
true for as many cases in the DoQ as indicated, and may be false for the rest.
We can therefore tell precisely when an indefinite NP is quantificational. As we have
seen, this is not merely a matter of the form of the NP, but crucially depends on the
context.
DEFINITION 5 Quantificational indefinite NPs
An indefinite NP is quantificational in a given utterance iff it can be construed
as partitive. It can be construed as partitive iff it can be replaced by the
corresponding partitive construction.
86
Understanding semantics
It remains to be stated that not all types of indefinite NPs are eligible for partitive
interpretation – for some types there is no corresponding partitive construction. The
partitive construction corresponding to an indefinite NP has a pronominal element
which corresponds to an adnominal element of the NP. Usually the adnominal
elements have identical pronominal forms, but sometimes they do not; for example,
none is the pronoun corresponding to adnominal no. In order to have a corresponding
partitive construction, an indefinite NP needs to contain an adnominal element with
a pronominal counterpart. Bare mass nouns and bare plurals lack such an element,
and the indefinite article, too, has no pronominal counterpart. The same holds for sm
with mass nouns and plurals. The weak form sm has no pronominal use. Pronominal
some [s m] is the counterpart of adnominal some [s m], not of sm. Therefore, simple
indefinite NPs lack a corresponding partitive construction – and cannot be used for
quantification. Only indefinite NPs with a quantity specification (cf. Definition 4 b
and c) are eligible for quantificational use.
(18) Indefinite NPs with corresponding partitive constructions
Only indefinite NPs with a quantity specification have a corresponding partitive
construction.
When indefinite NPs are not implicitly partitive, they simply refer to something not
yet uniquely determined in the given CoU; if the NP contains a quantity specification,
the latter adds information about the quantity of the referent (cf. nine kids in the
example above).
Indefinite mass NPs are analogous to count NPs. They can be used for quantification
iff they contain a specification of quantity and the context allows for a partitive
interpretation. An example would be:
(19) Mary:
John:
‘Did you have enough cake?’
‘Oh yes, there’s even some cake left.’
The indefinite pronouns somebody, none, etc, are like indefinite NPs with a quantity
specification (in fact their form tells that they originated from such NPs). They can be
used in a partitive construction and therefore for quantification if the context allows
for this interpretation, for example as someone of us, none of the cake, etc. They also
have, of course, non-quantificational uses as in There’s somebody at the door.
4.5.3 Quantification and definite NPs
Unlike indefinite NPs, definite NPs never quantify. This is obvious with singular
count definites:
(20) The kid will be brought.
The definite subject NP refers to a single case and the predication – in this case ‘x will
be brought’ – applies directly to this single case.
Meaning and context
87
Very similar are so-called collective plurals. Actually, collective plurals are not
a special variant of plural, but a variant of predication; I therefore prefer the term
collective predication. A collective predication is applied to a complex object; it can
be applied only to the whole as such, not to the elements that make it up. Examples
are (21a, b) with plural NPs and (21c) with a mass NP:
(21) a.
The kids know each other.
b.
The kids gather in front of the school.
c.
The literature on the topic is very diverse.
A single kid cannot gather, or know each other, a single item of literature cannot
be diverse. These are predications which can apply only to complex things. Thus,
such sentences, too, are predications about just one case, the complex referent of the
definite plural or mass NP.
The case of plural and mass definites is less obvious with distributive predication.
Distributive predication, too, applies to a complex object. However, it is ‘distributed’
over the individual parts that make it up; if it is true, it is true for every single part,
and thereby for the whole in general. (22a,b) are examples:
(22) a.
b.
The kids will be brought.
The literature on the topic is in Russian.
(22a) is true if for each of the kids the predication ‘x will be brought’ is true. Some, or
all of them may be brought together, but this is left open here, and does not matter.
Likewise, (22b) says that every single item of literature on the topic is written in
Russian.
Crucially, distributive predication, though about a multitude of single cases,
is essentially one predication about the complex object. This can be seen if one
considers such sentences together with their respective negation. There is a simple
method for determining the negation of a sentence: transform the sentence into a
yes–no question, answer it with ‘No’ and form the sentence which exactly expresses
what the negative answer says. Applied to (22a, b), this yields the following:
(23) a.
b.
‘Will the kids be brought?’ – ‘No.’
Negation: The kids will not be brought.
‘Is the literature on the topic in Russian?’ – ‘No.’
Negation: The literature on the topic is not in Russian.
Obviously, the negation is achieved by negating the VP, i.e. the predication of the
sentence. Now, if you look at the negations, you realize that they constitute the same
kind of predication: distributing down to the single parts that make up the whole.
The point of distributive predication is therefore whether or not it applies uniformly
to all parts of the whole. Therefore, distributive predication and its negation form an
88
Understanding semantics
all-or-nothing contrast: either the kids will all be equally brought, or all equally not,
analogously for the Russian or non-Russian literature. Treating all elements of the
complex referent uniformly makes them form a unit, a single complex case. If the
parts of the whole are not uniform, distributive predication does not make sense. If,
for example, some of the kids will be brought and some will not, we can say neither
(22a) nor its negation (23a).
To sum up, sentences with predications about the referent of a definite NP always
constitute a predication about just one case. The case is complex if the definite NP is
mass or plural, but it is still just one case. Distributive predication carries down to the
elements of the complex referent, but it carries down uniformly and therefore, too, it
treats the complex referent as one uniform whole. For this reason, predications about
the referent of a definite NP are never quantifications.
If the definite NP is mass or plural, and the predication distributive, the
sentence shares with instances of quantification the reference to a complex domain.
But it differs crucially from quantification in that it excludes the possibility of
positive and negative cases in the domain. This very possibility, however, is what
quantification is all about. Every single case in the domain is to be considered
individually; quantification always involves parallel predications with potentially
different outcomes.
Table 4.9 summarizes the results of the discussion of quantifying, indefinite and
definite NPs.
Table 4.9
Quantificational and non-quantificational NPs
NPs of the form …
quantify …
every, each, all, both N
(genuine quantifying NPs)
… always
indefinite NP with
quantity specification
… if construed as partititve
simple indefinite NP
(a(n) N, bare mass N, bare plural)
… never
definite NP
… never
4.5.4 Quantification and negation
With quantifying NPs, negation yields a different picture than with definite NPs. The
negation of (24a) is (24b):
(24) a.
b.
Each kid will be brought.
Negation: Will each kid be brought? – No. Not each kid will be brought.
Meaning and context
89
The negation is not placed at the verb, but at the quantifying determiner (at least
12
it can be placed there). The result is an all-or-not-all contrast. Unlike the all-ornothing contrast we observed with (22), it covers all possibilities. The kids in the
DoQ are either all brought (24a) or not all (24b); the mixed case of some kids being
brought and some not is also covered by the negation. Recall that it is covered neither
by (22a) – the kids will be brought – nor by its negation (22b) if quantification is
absent.
The quantifiers every, all, and both behave the same way as each. With quantifying
indefinite NPs, we also get a contrast that covers all possibilities. For example, the
negation of a quantifying NP of the form ‘some N’ is formed by replacing some with
no:
(25) a.
b.
Some kids will be brought.
Negation: No kids will be brought.
Here, the mixed case is covered by the positive sentence. Figure 4.2 displays the
different contrasts under negation for sentences with definite NPs and quantifying
NPs. Quantifying sentences and their negations together cover the whole range of
possibilities, while sentences with a definite plural or mass NP leave a gap, due to the
all-or-nothing contrast between global truth and global falsity. These observations
explain the function of quantification: it serves to fill this gap which arises with
simple predication over complex referents.
Figure 4.2
Negation contrasts for definite and quantifying NPs
negative cases only
mixed cases
positive cases only
the N not
the N
not every N
no Ns
every N
some Ns
For sentences with a definite subject NP, negation is formed by negating the VP,
as you saw in (22) and (23) above. For sentences with a quantifying subject NP, VP
negation yields not the negation of the sentence, but what is called internal negation
as opposed to external negation in (24b) and (25b).
12 There is also the syntactic variant EACH kid will NOT be brought, with a special intonation contour
on each and not, which serves to indicate that the negation relates to each rather than to the verb,
where it is syntactically placed.
90
Understanding semantics
(26) a.
b.
Each kid will not be brought. (= No kid will be brought.)
Some kids will not be brought. (= Not all kids will be brought.)
Sentences with a definite subject NP do not offer the possibility of two ways of
negation. VP negation is the only negation available and it yields the negation of the
sentence as a whole.
4.5.5 The nominal onion
The discussion of determination shows that an NP may have several layers of
structure; it forms an onion, metaphorically speaking. The core is formed by the
noun; next to the core are adjectives. The next layer consists of quantity specifications
such as numerals, several, many, few, etc. To these, determination is added next; this
layer comprises definite and indefinite articles, demonstratives and possessive
determiners. The outermost layer is constituted by quantifiers such as all, every, each
and both. Most quantifiers do not combine with an article; all is an exception which
shows that quantification supersedes determination. Figure 4.3 displays the nominal
onion with an example and with the elements belonging to the different layers.
Figure 4.3
The nominal onion
all
(
quantifier
all
every
both
the
(
determiner
the
a
sm
this / that
my / etc.
three
(
quantity
blind
mice
( adjective ( noun ) ) )
) )
one, two, …
several
some
much, many
no
…
4.6 GENERIC NPs
4.6.1 Generic sentences with simple indefinites
There is a special use of NPs, where they do not serve concrete reference but the
expression of general conditions. Examples would be the simple indefinite NPs in
sentences like these:
(27) a.
A millipede has a thousand legs.
b.
An icosahedron consists of twenty triangles.
c.
Catholic priests are not allowed to marry.
Meaning and context
d.
Bare mass nouns are indefinite.
e.
Money does not buy happiness.
91
Using such sentences, one does not refer to things in a concrete CoU, but one talks
in general terms about, for example, millipedes, icosahedrons, Catholic priests,
bare mass nouns or money. Because of their ‘general’ character, such sentences, and
the NPs in them, are called generic. Non-generic sentences are called episodic, or
particular. Let us consider the following dialogue between Mary and John about
their neighbour Gerald, a Catholic priest; (28a) is episodic, (28b) generic.
(28) a.
Mary:
‘Just imagine, Gerald is going to get married in October.’
b.
John:
‘Catholic priests aren’t allowed to marry, are they?’
John replies with this generic sentence, because it states a general condition relevant
for Gerald in the given CoU. But using this sentence he is not explicitly referring to
Gerald, as he would if he said, for example, ‘As a Catholic priest, Gerald isn’t allowed
to marry, is he?’
Talking in generic sentences is best understood as a special mode of speech in
which elements of a sentence are used without concrete reference in the CoU. By
doing so, a certain ‘general’ quality of the sentence is accomplished. For generic
elements in a sentence, utterance meaning does not depend on the CoU – unlike the
utterance meaning of referring NPs. There may be something like concrete cases in
the CoU to which a generic statement applies, but these cases are not directly referred
to. The possible cases of application are all treated alike. This is what leads to the
blanket character of generic sentences.
In the semantic literature, generic indefinite NPs are mostly analysed as universal
quantifiers. According to this view, Millipedes have a thousand legs would essentially
mean ›all millipedes have a thousand legs‹. This analysis is not correct. To see this,
consider again the contrast between such a sentence and its negation. The negation
of sentences with generic indefinites is formed by negating the VP of the sentence.
The sentences (29a, b, d) are the negations of the respective sentences in (27), while
the negative sentences (27c, e) are the negations of (29c, e):
(29) a.
A millipede does not have a thousand legs.
b.
An icosahedron does not consist of twenty triangles.
c.
Catholic priests are allowed to marry.
d.
Bare mass nouns are not indefinite.
e.
Money buys happiness.
Again, the sentences and their respective negations form an all-or-nothing contrast.
Consequently, generic indefinites cannot be considered universal quantifiers since
universal quantifiers would yield an all-or-not-all contrast. Thus, the statement in
92
Understanding semantics
Table 4.9, that simple indefinite NPs do not quantify, carries over to simple indefinite
NPs in generic use.
4.6.2 Generic quantification
Quantification is also possible in generic sentences – if one uses explicit quantifiers.
The quantification in generic sentences does not refer to a domain of concrete cases
given in the CoU, but to this kind of case in general. Examples would be sentences
like:
(30) a.
Some think money does not buy happiness.
b.
No millipede really has a thousand legs.
c.
Every Catholic priest can marry.
The respective negations are:
(31) a.
Nobody thinks money does not buy happiness.
b.
Some millipedes really have a thousand legs.
c.
Not every Catholic priest can marry.
In each case, the contrast between the positive and the negative is not all-or-nothing,
but covers all possibilities. Like with episodic quantification, there is the possibility
of internal and external negation.
Indefinite NPs with a quantity specification usually do not yield quantification in
generic use:
(32) a.
Two heads are better than one.
b.
Two’s a company, three’s a crowd.
c.
Too many cooks spoil the broth.
d.
Many children only have one parent.
The first three sentences are not construed as quantifications (over heads, people or
cooks), but as generic statements about cases of two heads vs one head, of two people,
or three, and of too many cooks working on the same broth. Only generic indefinite
NPs with vague quantity specifications such as many, much, more, most, only a few,
etc. may have a quantifying reading. (Note that many in (32d) is quantificational,
while too many in (32c) is not.) The quantificational readings concern the relative
frequency of the cases described, among the totality of cases of the kind. Therefore,
such statements can equivalently be made using adverbs of frequency:
(33) Adverbial generic quantification:
Children often/rarely/frequently/mostly/always/never have one parent.
Meaning and context
93
In principle, the possibility of expressing quantification by means of adverbs instead
of quantificational NPs is also given for episodic quantification. For example, the
quantificational sentence in (15c) can be equivalently phrased with quantifiers in
adverb position:
(34) Adverbial episodic quantification:
The kids are each to get a present.
(= Each kid is to get a present.)
The kids are all to get a present.
(= All the kids are to get a present.)
The adverbial variants of quantification in (33) and (34) display two points
of difference between episodic and generic quantification. First, only episodic
quantification involves reference to the totality of cases, namely the DoQ. Note that
the DoQ is directly referred to with a definite NP (the kids) in (34), while for generic
quantification the kind of cases related to is denoted by a generic plural NP (children)
in (33). Second, the choice of relevant adverbs shows that only generic quantification
is in terms of the frequency of relevant cases among the kind of cases related to.
Stopover
The determination of an NP determines the way in which the NP refers (if
it refers). The central phenomenon is definiteness. NPs can be definite or
indefinite or quantifying. Quantifying NPs represent a semantically complex
mode of expression: episodic NP quantification combines definite reference to
the DoQ with information about the proportion, or number, of cases within it
for which the predication is true. If you like, episodic NP quantification can be
considered a combination of simultaneous definite reference to the DoQ and
indefinite reference to the positive cases within it.
By no means all languages have quantifying determiners like English every.
The majority of languages use adverbial quantifiers for the expression of
quantification.
Generic NPs cannot be straightforwardly aligned with indefinite or definite
reference. All the English examples we discussed above are indefinite NPs,
but the expression of genericity by indefinite NPs is not representative across
languages. Many, if not most, languages would use definite NPs for this purpose.
Chinese, for example, which does not mark definiteness explicitly, uses the
same constructions for generic NPs as for definite NPs. The heterogeneity of
the grammatical determination of generic NPs can be explained by the fact
that generic NPs exhibit traits of both definite and indefinite reference. They
share with definite reference that what they relate to is given independently of
the CoU; with indefinite reference they have in common that they are not about
a particular single case.
94
Understanding semantics
PART 3: PRESUPPOSITIONS
4.7 PRESUPPOSITIONS
We will now turn to the third part of the phenomenon of indexicality –
presuppositions. It concerns indexical relation to the facts that make up the CoU.
The facts that are potentially relevant in a given CoU are much more manifold and
complex than the other components of a CoU (i.e. speaker, addressee(s), time and
place). Correspondingly, the relevant linguistic means are much richer. There is a
connection to the matter of determination: we will once again have to deal with
definite NPs.
4.7.1 An example
By way of illustration, let us consider the following short text to be read coherently:
(35) a.
Tony rose from the breakfast table.
b.
He went into his room and opened the window.
c.
He was sorry that he had not replied to Lucy.
Not only are we able to construe these sentences coherently, we also construct a
possible CoU that this passage would fit into. We will assume that in this context
there are two persons named Tony and Lucy, respectively. Tony has a room; the room
has a window; the window was closed until Tony entered the room to open it; and so
on. How are we led to these assumptions? This will become clearer if we take a closer
look at the wording of the text.
a. Tony rose from the breakfast table.
The sentence contains the name Tony as subject NP. As definite NPs express unique
reference, there must be a uniquely determined something called ‘Tony’ in the CoU.
There are many things which can bear the name Tony – male and female persons,
pets, cuddly toys, boats or songs; however, in this first sentence, it must be a person,
since only a person is capable of ‘rising from the breakfast table’. Later we are led to
assume that Tony is a male person because we will relate the anaphoric masculine
pronoun he to Tony; but from the sentence given we cannot tell the sex of the
bearer of the name. Turning to the VP rose from the breakfast table, we obtain more
information about the CoU. The definite NP the breakfast table informs us that there
is a uniquely determined table where people had breakfast; this, in turn, provides
information about the time of the day. Also, Tony must have been sitting at the table;
otherwise he could not be said to have risen from it.
All this information is not made explicit in the sentence. It does not state: ‘There
is somebody with the name of “Tony”; it is morning; there is a table for breakfast;
Meaning and context
95
Tony has been sitting at that table.’ The statement proper only states that Tony
rose from the breakfast table. The information about the CoU which we can read
off (or reconstruct from) the sentence is conveyed subliminally, as it were, by
certain elements of the sentence. They presuppose conditions that must be given
beforehand. With these implicit presuppositions, the sentence inconspicuously links
to the conditions in the given CoU. If these conditions were not given, uttering the
sentence would not make sense; it would be pointless. These presuppositions are
also technically termed thus:
DEFINITION 6 Presuppositions
Presuppositions of a sentence are those conditions pertaining to the CoU
which it must meet in order for an utterance of the sentence to make sense in
that CoU.
Presuppositions are part of the meaning of a sentence. When a proper name is used
as the subject or object of a sentence, the sentence presupposes that in the given CoU
there is a uniquely determined something which carries this name; to rise, related to
persons, presupposes that the person was not standing before, etc. The following two
sentences contain more cases of presuppositions.
b. He went into his room and opened the window.
The personal pronoun he is definite and therefore presupposes that it refers to a
male person, uniquely determined by this description. Analogously, the possessive
pronoun his presupposes that it refers to a male possessor uniquely determined by
using his. The NP his room is used referentially here and is therefore construed as
definite (recall 4.4.5); thus it presupposes that the description his room uniquely
singles out a particular room as its referent. It is said that ‘he’ went into his room. This
presupposes that ‘he’ was not in this room before. Opened the window contributes
two more presuppositions: the NP the window must uniquely refer to a particular
window, and this window was not open before.
One will construe the text automatically in the way that ‘he’ is Tony, that ‘his’ room
is Tony’s room, and that the window he opened is in that room. This interpretation
suggests itself if we interpret the second sentence in the context of the first. It is,
however, not semantically cogent. For example, if we knew that Tony is a woman,
we would not interpret he as coreferent with Tony; also the owner of the room could
be different from the one who is said to enter it. Similarly, the sentence does not say
that the window is in the room mentioned. We establish these connections within
the sentence and between the sentences by following general strategies of creating
coherence. Such coherence relations do not derive from the presuppositions of
the sentences. Presuppositions are general conditions that arise from the expression
meaning of the sentence. They cannot comprise conditions that are only given in
special CoUs.
96
Understanding semantics
c. He was sorry that he had not replied to Lucy.
The two occurrences of he make the same semantic contribution. Again, they
need not cogently be referring to the same person. The construction be sorry that
presupposes that what is described in the that clause is a fact: one cannot be sorry
about things that did not really happen. Therefore, the construction presupposes
that ‘he’ had not replied to Lucy. This in turn presupposes that there is something
uniquely determined by the name Lucy; this Lucy must have taken some action
‘he’ could have replied to, e.g. asked a question. For the sake of simplicity, we will
assume that this, together with the requirements of the verb reply to, amounts to the
presupposition that Lucy is a person.
4.7.2 Presupposition carriers
13
The expressions and constructions that carry presuppositions, can be grouped. It
would take us too far afield if we were to discuss all known kinds of presupposition
carriers. We will confine ourselves here to definite NPs and certain classes of verbs.
4.7.2.1 Definite NPs
When definite NPs are used for reference, they generally carry the presupposition that
the description of the referent thereby given is unique. That means first of all that a
referent exists in the given CoU (existential presupposition), but in addition also that
the given description is unique (uniqueness presupposition). This holds for definite
descriptions as well as for all other types of definite NPs: proper names, personal
pronouns, NPs with adnominal demonstratives and referential possessive NPs; Table
4.10 lists the different types of definite NPs and the presuppositions they carry.
Analogous presuppositions can be stated for the pronouns you (plural) and they,
as well as for the other possessive pronouns besides my.
Episodic quantifying NPs carry the same presuppositions as a definite NP referring
to the DoQ (cf.4.5.2); thus, each child has the same presupposition as the children.
4.7.2.2 Verbs
The example in (35) contained several verbs that carry presuppositions of different
types. For one, every verb imposes certain conditions on the things and persons it
predicates about. We saw in the examples that Tony must be a person; otherwise
he could not be said to ‘rise from the breakfast table’. At closer inspection, every
verb comes with such conditions for the referents of its subject, object and other
complements. The verb reply is restricted to persons; go is rather unspecific as to the
13 In the literature you will often find the term presupposition trigger for what is here called
presupposition carrier. The notion of a ‘trigger’ of presuppositions derives from the idea that
presuppositions are some kind of inference drawn when certain utterances are produced and
interpreted. This is not the understanding of presuppositions given here. Presuppositions are just
part of the meaning of the respective expressions or constructions, not something to be derived by
inference.
Meaning and context
97
Table 4.10
Definite NPs and their presuppositions
Definite NP
Presupposition
proper name
There is somebody or something which is uniquely determined by carrying this
name.
the dog
There is a dog and it is uniquely determined which dog is referred to.
this apple
There is an apple which is referred to and it is uniquely determined by deixis which
apple is referred to.
my skirt
There is a skirt which belongs to the speaker and it is uniquely determined which of
the speaker’s skirts is referred to.
I
There is somebody who produces this utterance and it is uniquely determined who
it is.
you (singular reading)
There is a person to whom this utterance is addressed and it is uniquely determined
who it is.
he
There is a male person referred to and it is uniquely determined who it is.
we
There are persons referred to, among them the speaker, and it is uniquely
determined who they are.
kind of referent it requires for the subject, but in the construction ‘go to one’s room’,
it is restricted to persons, too. These kinds of presupposition are called selectional
restrictions. They will be discussed in more depth in 5.7.
The verb open is representative for a large number of verbs used to express a
certain state coming about. ‘To open the window’ means to do something which
leads to the result that the window is open. Thus, on the part of the window the verb
denotes a transition from the window not being open to its being open. All such
verbs of change presuppose that the resulting state did not apply before. You cannot
open something which is already open, and so on. Table 4.11 lists examples.
Table 4.11
Verbs and their presuppositions
Verb
Presupposition
she opened the door
before, the door was not open
she left the kitchen
before, she was in the kitchen
she went into her room
before, she was not in her room
she lay down
before, she was not lying down
she became calmer
before, she was less calm
she stopped musing
before, she was musing
she picked up her guitar
before, she was not holding her guitar
she had a glass of water
before, she had not had this water
98
Understanding semantics
Some verbs with a that clause complement presuppose what is said in the that clause.
An example occurred in (35): He was sorry that he had not replied to Lucy. Such verbs
are called factive verbs. They include among many others regret, admit, report,
recognize and know.
(36) Tony regrets/admits/recognizes/knows that Lucy does not want to see him.
presupposes: Lucy does not want to see him.
By presupposing the content of the that clause, constructions with factive verbs also
presuppose whatever the that clause itself presupposes.
Certainly not all verbs with a that clause complement presuppose its content. For
example, say, think, hope, expect, fear, claim, plan, request, etc. do not:
(37) Tony says/believes/expects/fears/claims that Lucy does not want to see him.
does not presuppose: Lucy does not want to see him.
There is a further condition, which we already encountered in connection with
distributive predication in 4.5.3, namely the condition that the parts of the whole
to which a distributive predication is applied form a uniform whole – i.e. that
the predication is either uniformly true or uniformly false for all parts. Since the
character of a predication in most cases depends on the verb, we may mention this
presupposition here. It is called the presupposition of indivisibility and it requires
that the object of a distributive predication is uniform, or homogeneous, with
respect to what the predication says about it. For collective predications or simple
predications about just one case, this condition holds anyway. It can therefore be
considered a general presupposition of any predication whatsoever. The formulation
uses the technical term argument (see chapter 5) for the object of a predication:
DEFINITION 7 Presupposition of Indivisibility
The argument of a predication is indivisible with respect to the predication; a
predication is true or false of its argument as a whole.
As we have seen in the discussion of (22) and (23) above, it is this condition which
is responsible for the fact that a distributive predication and its negation form an
all-or-nothing contrast.
4.7.3 Presuppositions and truth conditions
Presuppositions immediately bear on the issue if a sentence is true or false. As we have
already stated in 1.1.2, this question arises only when a sentence is actually uttered
in a CoU. When uttered, the presuppositions must be given in the CoU. Otherwise
the question as to whether the sentence is true or false cannot be answered. Let us
consider a simple example to illustrate the point:
Meaning and context
(38) a.
99
The dog stopped barking.
The sentence carries the presupposition of existence and uniqueness of the referent
of the dog. If there is no dog in the given context, or if there are several dogs and
the dog referred to is not uniquely determined, it cannot be decided whether the
sentence is true or false. One just would not know which dog is being talked about.
If we assume that the presuppositions for the dog are fulfilled, there remains the
presupposition contributed by stopped barking: the dog has been barking. If the dog
had not been barking, the question as to whether it had stopped doing so or not
would not arise, i.e. whether the sentence is true or false.
The same holds for the negation of the sentence:
(38) b.
The dog did not stop barking.
The negation of the sentence carries the same presuppositions: the sentence refers
to a uniquely determined dog, and this dog has been barking. For if it had not been
barking, there would be no point in saying that it did not stop barking.
The same presuppositions also hold for the corresponding yes–no question:
(38) c.
Did the dog stop barking?
Only if (38a) is true, can the question be answered ‘Yes’; only if (38b) is true, can it
be answered ‘No’.
This shows that for declarative sentences presuppositions are necessary conditions
for the sentence having a truth value at all in a given CoU. We have, therefore, to
modify the definition of truth conditions we gave in 2.2.2 in order to take account of
presuppositions.
DEFINITION 8 Truth conditions (revised)
The truth conditions of a sentence are the general conditions under which it is
true, provided that all presuppositions of the sentence are fulfilled.
The presuppositions are not themselves part of the truth conditions, but exist prior
to them. This is clear from the fact that they also hold when the sentence is false.
In the literature, presuppositions are often treated as some sort of inference,
14
similar to logical entailment (see 5.3.1) or to so-called conversational implicatures.
However, this is a confusion of cause and effect. Presuppositions are conditions
14 Conversational implicatures are a central notion in Grice’s theory of language use (mentioned
in 1.1.2). According to Grice, people engaged in a conversation will normally assume that their
interlocutors adhere to basic social rules of conversation, such as telling the truth, being relevant,
being economical, etc. Conversational implicatures are inferences drawn on the basis of this
assumption. For instance, we will normally believe what a speaker says because we assume he or
she tells the truth.
100
Understanding semantics
on the CoU. If someone utters a sentence in a concrete CoU, his interlocutors will
assume that the utterance is meant to make sense in the given CoU and that the
presuppositions therefore are to be taken as given. Unless there is evidence to the
contrary, the addressee(s) will assume that this is the case. Thus, the presuppositions
of a sentence will normally cause an inference to the effect that they are fulfilled in
the given CoU.
4.7.4 Presupposition tests
These considerations lead to two tests for deciding whether a condition is a
presupposition or not. The negation test can be used to check whether the condition
also holds for the negated (or un-negated sentence). The question test checks if the
same condition holds for the corresponding yes–no question. If the result is positive,
the condition is a presupposition. The three sentences in (38a, b, c) illustrate the two
tests for the sentence the dog stopped barking. Let us return for the last time to the
examples in (35).
(35) a.
sentence:
Tony rose from the breakfast table.
negation:
Tony did not rise from the breakfast table.
question:
Did Tony rise from the breakfast table?
Obviously the negation and the question carry the same presuppositions as we stated
for the original sentence.
In the case of the second sentence – he went to his room and opened the window –
the tests would be applied separately to the two sub-statements it contains: he went
to his room and he opened the window. Again, it turns out that the negations and
corresponding questions carry the same presuppositions as were stated above for
the positive sentence.
For the third sentence, one needs to observe that forming the negation and the
question leaves the that clause untouched:
(35) c.
sentence:
He was sorry that he had not replied to Lucy.
negation:
He was not sorry that he had not replied to Lucy.
question:
Was he sorry that he had not replied to Lucy?
The negation test and the question test can both only be used for declarative
sentences. But this is sufficient for identifying most presupposition carriers.
4.7.5 Presuppositions and the Principle of Consistent Interpretation
The Principle of Consistent Interpretation (3.4.4) requires that an utterance fit its
context. Part of this requirement is that the presuppositions of an utterance be
fulfilled in the given CoU because otherwise the utterance would not make sense. The
Principle has two consequences concerning presuppositions. First, it may happen
Meaning and context
101
that the context as it has been understood up to the time when a sentence is uttered
did not positively support its presuppositions. In this case, very often so-called
accommodation takes place: the context is adapted to the utterance in order to do
justice to the presuppositions. Second, the requirement that certain presuppositions
be fulfilled may trigger meaning shifts.
Accommodation occurs all the time. It is such a common thing to do that we often
exploit the mechanism in conversation. For example, assume Sheila is just about
to leave home for town and Mary asks her, ‘Would you post my letter?’ Sheila will
accept the presupposition that her mother has a letter to be posted, even if she had
not known about that letter; otherwise she would have to assume that her mother’s
request is pointless. Mary would have relied on this inference by Sheila, which saves
her an explanation.
To see the key role of presuppositions in triggering meaning shifts, have a second
look at the examples for metaphor, metonymy and differentiation that we considered
in 3.4.3. Let me give just three examples:
(39) a.
Metonymy:
Moscow declared the Chechen rebels defeated.
b.
Differentiation:
She had forgotten to charge her card.
c.
Metaphor:
His courage evaporated.
The selectional restrictions of the verb declare for its subject referent ask for a human
being. In addition, the whole VP declared the Chechen rebels defeated requires some
authority. This triggers the metonymic shift from the original meaning ›(city of)
Moscow‹ to ›Russian government (residing in Moscow)‹. In the second sentence, the
verb requires a chargeable type of card as the object referent, whence the general
meaning of card will be shifted to a differentiated meaning, say ›SIM card‹ or ›money
card‹. In (39c), too, there is a mismatch between the type of subject referent and the
selectional restrictions of the verb. In this case however, the mismatch is dissolved
by giving the verb a metaphorical reading, approximately ›vanish‹ with selectional
restrictions that fit the given type of referent.
4.8 SUMMARY
We have come a long way in this chapter, starting from deixis and demonstratives,
visiting definiteness, quantification and genericity, to finally arrive at presuppositions.
What unites all these phenomena is that they belong to the interface between
sentence meaning and utterance meaning. The deictic means relate to the immediate
context of utterance for establishing reference out of it. The most important linguistic
means of reference are NPs. NPs are subject to determination; in most cases, they
are either definite or indefinite. Definite NPs are indexical in that they depend on
the given context for determining their reference. Likewise all episodic quantifying
NPs are indexical in referring to a uniquely determined domain of quantification. In
turn, definite NPs are a special case of presupposition carriers. Presuppositions in
102
Understanding semantics
general are indicative of the context of the utterance. By imposing conditions on the
CoU, they indicate what kind of context is required for the sentence to make sense at
all. Presupposition carriers not only serve to establish reference; more generally they
serve to integrate the content of a sentence into the context in which it is used.
EXERCISES
On deixis
1. Consult a Spanish grammar or textbook for the system of personal pronouns. Set
up a table like Table 4.1.
2. What do personal and possessive pronouns have in common, and what is the
difference between them?
3. Use WALS (http://wals.info/) in order to find three languages on different
continents which distinguish inclusive and exclusive WE. Use the Ethnologue
website (http://www.ethnologue.com/) to classify the three languages. For
example, the classification for English is: Indo-European, Germanic, West.
4. Consult Siewierska (2004) to find two languages which do not have 3rd person
pronouns. Use Ethnologue to classify them.
5. Try to find out if there are formal pronouns of address in Finnish and, if so,
determine which of the strategies mentioned in 4.1.2 is used in Finnish. Try to
find out what the rules are for using the formal type of address (interview a native
speaker, if possible).
6. Use the Oxford online dictionary (http://oxforddictionaries.com/) to determine
if, according to the definitions given there, the following nouns are relational or
not: spouse, bed, frog. If more than one definition is given, answer the question for
each of them separately. The answers may differ.
7. Discuss with fellow students the space deictic quality of the verbs bring and fetch.
Give a tentative written description.
8. a. It is very rare to find different feminine and masculine 1st person singular
pronouns, while a difference in gender for 2nd person singular pronouns is
more frequent. What might be a reason for the difference?
b. In languages with 1st person pronouns that differ in social meaning, there
may be 1st person pronouns exclusively used by women. Think of a possible
reason.
On determination
9. Explain the function of the definite article in your own words.
Meaning and context
103
10. Consult WALS to find two languages, each from different continents, which
a. have definite articles but no indefinite articles,
b. vice versa.
Use Ethnologue to classify them.
11. In the following text
a. determine which NPs are definite
b. for the definite descriptions among them, which are pragmatically unique,
and which semantically unique?
The Mole had been working very hard all the morning, spring-cleaning his
little home. First with brooms, then with dusters; then on ladders and steps
and chairs, with a brush and a pail of whitewash; till he had dust in his throat
and eyes, and splashes of whitewash all over his black fur, and an aching back
and weary arms. Spring was moving in the air above and in the earth below
[…] (from Kenneth Grahame, The Wind in the Willows, Amazon Kindle
edition, originally published by Charles Scribner’s Sons, New York, 1913)
12. a. Determine all the indefinite NPs in the previous text. Are there quantifying
NPs among them?
b. Identify all the nouns in the text and determine whether they are count nouns
or mass nouns.
13. Determine which NPs in the following text (a passage from chapter 10.1) are
indefinite. Decide which ones are quantificational and which ones are not.
English has a count noun wood1 and a mass noun wood2. A wood1 consists
of trees, while wood2 is the substance trees largely consist of. Hence, the two
meanings can be clearly distinguished on linguistic grounds. Likewise, the
Japanese meanings of ki1 ›tree‹ and ki2 ›wood2‹ are clearly different. It follows
that neither of the two English expressions matches with either of the two
Japanese words. They only match in one meaning variant, respectively.
14. In the following text (a passage from chapter 3), determine which indefinite NPs
are generic and which are not.
In dealing with different aspects of meaning in the previous two chapters,
expressions were treated as though they had only one meaning (though possibly
composed of different parts). This is, of course, not the case. Many words have more
than one meaning and even complete sentences may allow for several readings.
On presuppositions
15. Consider the sentence Nicola left her bike in front of the supermarket. Which
among the sentences a. to e. state a presupposition of the sentence?
a. There is a uniquely determined person called ‘Nicola’.
b. There is a uniquely determined bike which belongs to a uniquely determined
female person.
104
Understanding semantics
c. The bike belongs to Nicola.
d. The owner of the bike had been riding on it before it was left at the
supermarket.
e. After Nicola left it there, the bike was in front of the supermarket.
16. Try to identify the presuppositions involved in the following three sentences:
(a) The father got back to the hospital after midnight. (b) The wife was sitting
in the chair by the bed. (c) She looked up at the husband and then she looked
back at the child.
(from Raymond Carver, ‘The Bath’, in What We Talk About When We Talk About
Love, p. 41, London: Vintage, 2003)
17. What does the presupposition of indivisibility mean for the following sentence?
The eggs are raw.
FURTHER READING
Deixis
Anderson and Keenan (1985) and Diessel (2012) provide cross-linguistic surveys.
On deixis and anaphora in English see chapter 17 by Lesley Stirling and Rodney
Huddleston in Huddleston and Pullum (2002).
4.1 Siewierska (2004) is a comprehensive survey on the topic of grammatical person:
ch. 3 on person deictic systems, ch. 6 from the perspective of social meaning (‘social
deixis’). Cooke (1968) on terms of address in Thai, Burmese and Vietnamese. In the
World Atlas of Language Structures online (WALS) you will find surveys on systems
of personal pronouns in several hundred languages: ch. 44 (Siewierska) on gender
distinctions; ch. 35 (Daniel) on plurality; ch. 39 (Cysouw) on the inclusive/exclusive
distinction; ch. 45 (Helmbrecht) on politeness distinctions; ch. 43 (Bhat) on 3rd
person pronouns and demonstratives. For the more general aspects: ch. 36 (Daniel,
Moravcsik) on associative plural; ch. 31, 32 (Corbett) on gender systems; ch. 33
(Dryer) and ch. 34 (Haspelmath) on plural marking; on inalienable possession see
chs 58 (Bickel and Nichols) and 59 (Nichols and Bickel).
4.2 Diessel (1999) is a general typological overview on demonstratives; there are
chapters in WALS by the same author on distance contrasts in demonstratives
(ch. 41) and on pronominal and adnominal demonstratives (ch. 42). On place deixis,
see Levinson (1983: 2.2.3).
4.3 Levinson (1983: 2.2.2) on time deixis.
Determination
For determination and NPs in English, in particular definiteness, quantification, the
mass/count distinction, genericity, number and gender see the elaborative account
Meaning and context
105
provided in ch. 5 by John Payne and Rodney Huddleston in Huddleston and Pullum
(2002).
4.4 Löbner (1985) on definiteness. WALS chs 37 and 38 (both by Dryer) on definite
and indefinite articles.
4.5 There is a lot of literature from the field of formal semantics on quantification
(see ch. 13). Most of the literature adopts the approach of Generalized Quantifier
Theory which subsumes indefinite and definite NPs under quantifiers.
4.6 Carlson (2012) on genericity.
Presuppositions
Seuren (1990) on presuppositions, Levinson (1983: 4.2) on presupposition ‘triggers’
(= carriers).
5
?
Predication
We will now have a look at the way in which different kinds of words interact to
form meaningful phrases and sentences, addressing in more detail the mechanism
of composition (1.2). We will see that a phrase or a sentence is by no means a simple
sequence of words, one added to the other like beads on a string. Rather, sentences
have a sophisticated structure in which each part plays its own role and interacts
with the others in its own way. Here, we will focus on one central semantic property
of verbs, nouns and adjectives: the way they provide predications about one or more
of the potential referents of the sentence.
5.1 PREDICATIONS CONTAINED IN A SENTENCE
Let us take a look at the following example, a simple English sentence, and try to
determine the proposition it expresses.
(1)
Johnny sent money to a dubious company.
The first word, Johnny, is a proper name. It forms a complete NP. The VP contains
a finite verb, i.e. a verb which carries tense and, in English, ‘agrees’ (see 5.6.2)
with the subject of the sentence. Here the subject is the NP Johnny and verb tense
is simple past. The bare mass noun money forms another NP, the so-called direct
object of the verb. The last three words, a dubious company, form a third NP, in this
case also containing an adjective between article and noun. This NP is part of a PP
(prepositional phrase) headed by the preposition to. The PP is the so-called indirect
object of the verb. The total VP consists of the finite verb, the direct object NP and the
indirect object PP. In Fig. 5.1 the sentence is analysed into its syntactic components.
Figure 5.1
Grammatical structure of the sentence in (1)
[ [JohnnyNP] [sentV [moneyN] [toP [a dubiousA companyN]] ] ]
NP
NP
NP
Sentence
VP
PP
Predication
107
The three NPs each provide a description of one referent. The subject NP describes
its referent as an entity with the name Johnny, the direct object NP characterizes
its referent as ‘money’ and the indirect object NP provides the information that its
referent is a dubious company. Let us call the three referents rj, rm and rc, respectively.
The verb contributes the information that they participate in an event of sending,
rj being the sender, rc the recipient and rm the object sent. Thus the meaning of
the sentence constitutes complex information about these three referents. If an
expression provides information about a referent, it is said to make a ‘predication’
or to ‘predicate’ about it. In Table 5.1, it is sorted out which words contribute which
predications about which referents. For the time being, we disregard the event
referent of the verb (but see 5.3.2) and we do not treat the proper name Johnny as
contributing a predication. The preposition to is omitted from the list because it does
not contribute a predication. Similarly, articles would not be included.
Table 5.1
Predications contributed by the words in (1)
Word
Predication
sent
rj sent rm to rc
rm was sent to rc by rj
rc was sent rm by rj
money
rm is money
dubious
rc is dubious
company
rc is a company
The simple analysis shows two interesting things. First, while the nouns and the
adjective relate to only one of the referents, namely the referent of the NP they belong
to, the verb relates to all three referents. Second, different words in the sentence
contribute predications about the same referents. Figure 5.2 depicts these relations.
Figure 5.2
Predication structure of sentence (1)
words
NPs
referents
[ JohnnyNP] sentV [ moneyN] toP [ a dubiousA companyN]
rj
rm
rc
108
Understanding semantics
A broken line connects each NP to its referent; arrows point from the single words to
those referents they provide information about. The figure illustrates the central role
of the verb: by predicating about all three referents, the verb ties the whole sentence
together. The three NPs, in contrast, independently supply information about one
referent each.
5.2 PREDICATE TERMS AND ARGUMENT TERMS,
PREDICATES AND ARGUMENTS
Words that contribute a predication are called predicate terms.1 Predicate terms
can be verbs, nouns, adjectives or adverbs (like rapidly). Whether or not proper
names like Johnny constitute predicate terms is a controversial issue. Many formal
semanticists hold that they are just names, i.e. direct expressions for their referents.
Alternatively, proper names can be considered predicate terms that express the
property of ‘being called [Johnny etc.]’. Both positions have their merits. We will
adopt the first position here.
The meanings of predicate terms are concepts that concern one or more entities:
›money‹, ›dubious‹ and ›company‹ concern one entity each, the event concept ›send‹
concerns three. Such concepts are called predicates; the entities they concern are
their arguments. Predicates are ‘applied’ to their arguments: for example, in (1),
›dubious‹ is applied to rc and ›send‹ is applied to rj, rm and rc. Predicates with one
argument are called one-place predicates, predicates with two arguments two-place
predicates and so on. One-place predicates are also called ‘unary’ predicates, twoplace predicates ‘binary’, etc.; therefore the term ‘arity’ is used for the number of
arguments of a verb.
The arguments of a predication are those entities in the world to which the
predication is to be applied. In (1) the arguments of the predicate ›send‹ are
described by the three NPs of the sentence: the subject, the direct object and the
to-PP. These NPs are argument terms belonging to the verb – three constituents of
the sentence which are grammatically related to the verb. In this chapter, the term
complement will be used for all argument terms that form a separate syntactic
2
constituent and have their own referent. Figure 5.3 displays the verb and its three
complements.
1 In the literature, you will often find the term ‘predicate’ used indiscriminately for predicate terms
as well as for their meanings.
2 In syntactic approaches, a distinction is drawn between obligatory and optional argument terms.
Only the former are called complements, the latter ‘adjuncts’. A general definition of the distinction
is notoriously difficult and depends on the theory adopted. We will not tackle the problem here and
use the term complement in a wide sense that also includes optional argument specifications.
Predication
109
Figure 5.3
A predicate term and its complements
predicate term
Johnny
subject
sent
money
to a dubious company
direct object
to-PP
three argument terms (complements)
While predicate terms and argument terms both belong to the level of expressions,
predicates and arguments belong to different levels: predicates are concepts, and
arguments, being referents, are things in the world. Thus, predication always involves
three levels: (i) the level of linguistic expression: predicate terms and argument
terms; (ii) the conceptual level of the meanings of the predicate terms and argument
terms involved; (iii) the referential level of arguments. Figure 5.4 shows the three
levels involved with the predication expressed by a verb; the scheme basically adds
the conceptual level to the diagram in Fig. 5.2. The double horizontal line separates
the level of concepts from the level of the world. The simple vertical lines connect
the expressions in the first line to their meaning in the second; dashed lines connect
these concepts to the referents of the NPs; the thick arrows mark the relation between
the verb predicate and its arguments.
Figure 5.4
Levels involved in a predication by means of a verb
the company
kept
the money
expression level
argument term
NP
predicate term
verb
argument term
NP
conceptual level
argument description
›company‹
predicate
›keep‹
argument description
›money‹
referential level
argument
the company
argument
the money
110
Understanding semantics
Analysing predication by means of NPs that function as argument terms calls for a
different picture (Fig. 5.5). In this case, the argument of the predication expressed by
the noun is not described in a separate argument expression, but is just the referent
of the NP; it is therefore called a referential argument. A predicate term can have
at most one referential argument. Later, we will also consider referential arguments
of verbs (5.3.2).
Figure 5.5
Levels involved in a predication by means of a noun
money
expression level
NP
conceptual level
predicate
›money‹
referential level
argument
(and referent)
the money
When adjectives are combined with a noun, we have yet another situation. The
argument of ›dubious‹ in (1) is the referent of the NP a dubious company. Since
reference is considered a matter of nouns, not of adjectives, the argument of ›dubious‹
is not a referential argument of the adjective, nor is it specified by a complement of
the adjective. Rather, the adjective ‘borrows’ its argument from the noun. In cases like
these, I propose to talk of a parasitic argument.
The analysis of predication demonstrates an important point: a predication only
comes about at the level of utterance meaning, since it is ultimately about referents of
the sentence. Predication only becomes effective in a concrete CoU, and only then it
delivers a truth value by application to the given arguments. This is the reason why the
question of truth of a sentence only arises at utterance level (1.1.2): sentences are true
or false because of the predications they convey about the referents of the sentence.
Since predication is a matter of the utterance level, it is subject to the Principle
of Consistent Interpretation (3.4.4). Elimination and modification of readings can
interfere with the conceptual level of predication. In fact, all the examples of meaning
shifts discussed in 3.4 involve predications and the problem of making predicates
applicable to their arguments. We will return to this topic in the section about
selectional restrictions (5.7.3).
Predication
111
The considerations on the levels involved in predication give rise to another variant
of the semiotic triangle. The configuration in Fig. 5.6 represents the relationships for
a single predication in a given CoU.
Figure 5.6
The semiotic triangle for predication
predicate term
predicate:
a concept
set of
arguments
5.3 VERBS
We will now turn to the three major word classes and the ways in which they function
as predicate terms. Verbs are combined with a separate argument term for each of
their arguments. They also have a referential argument not mentioned so far: the event
to which the verb refers. We will discuss this later in 5.3.2. The referential argument
is not counted when we classify a verb as one-place, two-place, etc. The discussion
will be confined to the most frequent types of verb. If verbs, nouns and adjectives are
categorized as one-, two- or three-place predicate terms, this is based on the number
of arguments in standard constructions. Obviously, it is almost always possible to
add further optional argument specifications to a given predicational construction. It
would make little sense to try to count all possible arguments that might be specified
for a given argument term. The issue of how many arguments a given lexical item has,
however, is not trivial and will be taken up again at the end of this section.
5.3.1 Major types of verb
Intransitive verbs are one-place predicate terms. The only argument is specified
with an NP which, in English, is always the subject of the sentence.
(2)
a.
The cat is sleeping.
b.
The snow has melted.
c.
The door opened.
d.
The author of the love letter did not succeed.
112
Understanding semantics
Transitive verbs are two-place predicate terms with two argument terms, the subject
and the direct object of the sentence.
(3)
a.
The cat is eating the dog’s food.
b.
He wants your help.
c.
The dog cannot open the door.
d.
Thirty-one students filled in the questionnaire.
Note that in (3d) the words in the questionnaire do not form a PP. In is part of the
particle verb fill in, rather than a preposition. You will have noticed that the verb open
appears as an intransitive verb in (2c) and as a transitive verb in (3c).
Ditransitive verbs have three argument terms. For one group, the third argument
term is an ‘indirect object’. In standard English word order, an indirect object is
either placed before the direct object or it is marked with to and placed after it.
(4)
a.
He’ll give my sister the keys.
He’ll give the keys to my sister.
b.
I showed them the photograph. I showed the photograph to them.
Other syntactic categories of complements. It should be mentioned that verb
complements are not confined to NPs. Many verbs have prepositional complements
with a lexically fixed preposition, cf. depend on, refer to, differ from, etc. Other verbs, e.g.
verbs of motion like go or put, can be combined with a wide range of PPs. Some verbs
take that-clauses as complements (know, believe, assume, say), infinitives (try/manage/
begin to) or gerunds (start/stop/keep -ing), to mention only the most common types.
Alternative grammatical means of specifying arguments. There are other ways
of specifying arguments of verbs than combining the verb with a complement. For
example, English imperative sentences usually have no subject. The corresponding
argument is then construed as the addressee(s). The imperative sentence in (5a)
expresses the same proposition as the corresponding declarative sentence (5b). Put
is a three-place verb with a direct and a prepositional object.
(5)
a.
Put the keys on the window-sill.
b.
You put the keys on the window-sill.
In languages such as Spanish or Italian, grammatical person and number are
reflected in verb inflection. The subject
itself can be omitted in suitable CoUs and
3
the resulting sentences are equivalent to sentences with the corresponding personal
pronouns in subject position (recall 4.1.3).
3 Because the pronoun is usually omitted, its use results in some sort of emphasis on it. In this
sense, the pronoun, if present, is not redundant.
Predication
(6)
a.
b.
habl-o
árabe
speak-1S
Arabic
habl-as
árabe
speak-2S
Arabic
≈
≈
yo
hablo árabe
I
speak-1S Arabic
tú
hablas árabe
113
(Spanish)
you speak-2S Arabic
Mechanisms similar to the one described for Spanish can be observed in languages
like Korean or Japanese, which have an elaborate system of social meaning. For
example, certain honorific forms of verb would only be used if the subject argument
is the speaker themselves or, in other cases, an addressee to be treated formally. In
such cases, the subject argument can be inferred from the form of the verb and
the rules of social meaning. Recall, for example, the Japanese sentences (10a-d)
discussed in 2.3.2.
5.3.2 Referential verb arguments
It is widely assumed that verbs also predicate about a referential argument for the
event described, in addition to the arguments hitherto mentioned. For example,
the referential argument of the predicate ›send‹ is what is referred to as an instance
of ‘sending’ when the verb send is used. As a general notion comprising actions,
processes and other occurrences, the term event is commonly used in semantics for
the referents of verbs.
There are good reasons for assuming a referential argument for verbs. First, the
event argument can be considered to serve as the argument of the tense element of
the meaning of the finite verb, which reasonably constitutes a predication of its own
(see 6.4). For example, in (1) the past tense form sent of the verb expresses that the
event of sending happened before the time of utterance. Second, the referential verb
argument can be considered the argument of certain adverbs, e.g. carefully in John
closed the envelope carefully: it is the closing, i.e. the event referred to, which is said to
be ‘careful’. Third, there are many deverbal nouns which denote an event of the kind
the verb denotes, e.g. the simple -ing derivation in (7).
(7)
Johnny’s sending money to the company was a mistake.
If we assume that the verb sent in (1) has an event argument, this argument is the
event of Johnny’s sending money to a dubious company. If (1) and (7) are both
related to the same CoU, the referential argument of the verb in (1) is identical to the
referent of the NP Johnny’s sending money to the company in (7). Thus, the meaning
of the derived noun sending can partly be captured by stating that it is a predicate
with the same referential argument as the verb concept ›send‹ (see also the analysis
of deverbal nouns in 12.2.2).
5.3.3 Deciding on the number of arguments
The question as to how many arguments a predicate term involves is often difficult
to settle. Let us mention just two aspects of the problem. Very many verbs appear in
114
Understanding semantics
more than one construction with a different number or a different quality of argument
terms. One sort of variation can be dealt with in a relatively straightforward way: the
occurrence of the same form in different grammatical categories, e.g. the abovementioned form to open with an intransitive and a transitive use. These must be
considered two different verbs as they belong to different word classes (intransitive
vs transitive verbs) and have different meanings. Intransitive open predicates of its
subject argument a certain change of state. Transitive open predicates of its subject
argument an action which leads to a corresponding change of state of the direct
object argument. Thus intransitive and transitive open express different predications
about their respective subject arguments. Their meanings are clearly different.
The second type of variation is more difficult to handle. A verb like eat can be used
in a wide range of constructions including the types instantiated by the following
examples:
(8)
a.
Fred is eating spaghetti.
b.
Fred is eating spaghetti with a plastic fork.
c.
Fred is eating spaghetti with a plastic fork from a big bowl.
d.
Fred is eating with a plastic fork from a big bowl.
e.
Fred is eating from a big bowl.
f.
Fred is eating.
In all these constructions, unlike in the case of intransitive and transitive open,
the verb eat predicates the same of its subject argument. Obviously, each of the
constructions contains a subject complement; in fact, English does not allow for its
omission. Other languages do, for example Japanese, but if the eater argument is
omitted, it must be clear from the given context who it is who eats. We may therefore
assume that the predicate ›eat‹ necessarily involves an eater argument. In view of
(8f) one might feel that all other complements are optional, since none of them is
grammatically obligatory. But the direct object argument is different. The concept
›eat‹ necessarily also involves a second argument. There can be no eating without
something being eaten. Therefore that argument is understood to be involved in the
situation described, even if it is not specified. Fred is eating is interpreted as Fred is
eating something. This is why we feel that the direct object is omitted in (8d, e, f).
The other arguments that can be added have a different status: Fred is eating is
not necessarily interpreted as Fred is eating with something or Fred is eating from
something. Neither something eaten from nor something eaten with constitutes
a necessary component of an eating event. Accordingly, specifications of such
arguments are not syntactically missing if they are absent. Thus the basic number of
arguments for ›eat‹ is two, and eat is a transitive verb, although its direct object can
be omitted. It must be added that not all transitive verbs allow the omission of the
direct object. Along with many others one such verb is eat’s close relative devour. We
will return to the question of the number of verb arguments in 6.1.
Predication
115
5.4 NOUNS AND ADJECTIVES
5.4.1 Major types of nouns
5.4.1.1 One-place nouns
The majority of nouns constitute one-place predicate terms. Unlike verbs, one-place
nouns are not combined with a separate argument term. They are primarily used as
the head of referring NPs that function, for example, as a verb complement (but see
5.4.3 for the ‘predicative’ use.)
(9)
The dog managed to open the door.
In this sentence, the argument of the one-place noun dog is the referent of the
subject NP the dog, and analogously for the noun door. Thus, both nouns have a
referential argument. In (1) above, the assumed person Johnny is the referent, and
the only argument of the proper name Johnny. Personal pronouns can be considered
as one-place predicate terms, too: I expresses the predication ›is the speaker‹, you
with singular meaning conveys the predication ›is the addressee‹. The predicational
content of 3rd person pronouns is confined to meaning components such as gender
(recall 4.1.1 on the meaning of personal pronouns). The definiteness component of
personal pronouns is not part of their predicational content.
In general, the determination of an NP is independent of its predicational content.
In the case of common nouns, the same predicational content can be combined with
definite, indefinite, quantificational or generic determination. For proper names and
pronouns – which are usually used as full NPs – a particular determination is part of
their lexical meaning. Different cases are displayed in Table 5.2; the pronoun one here
is the impersonal pronoun as in One cannot be too careful in these matters.
Table 5.2
Predication and determination of proper names and pronouns
Predication
Determination
Sheila
has the name ‘Sheila’
definite
she
female
definite
someone
person
indefinite
everything
non-person
quantificational
one
person
generic
5.4.1.2 Relational nouns
In 4.4.2.1, we introduced the distinction between sortal, individual, relational and
functional nouns. Sortal and individual nouns (e.g. priest and pope, respectively) are
one-place predicate terms, while relational and functional nouns (sister and mother,
116
Understanding semantics
respectively) are two-place. The aspect of inherent uniqueness that distinguishes
sortal from individual, and relational from functional nouns, does not matter in this
case. We will therefore use the term ‘relational noun’ in this chapter in the broader
sense that also includes functional nouns.
One group of relational nouns is kinship terms, e.g. uncle and sister in (10).
(10) My uncle is going to marry Molly’s sister.
The two NPs my uncle and Molly’s sister each have a referent: the speaker’s uncle
and Molly’s sister, respectively. These are the referential arguments of the relational
nouns uncle and sister. In addition to the referential argument, each noun has an
argument for the one the referent is a relative of. In the terminology of kinship
relations, this is called the ‘propositus’. In the case of Molly’s sister, the propositus is
Molly. It is specified by the NP Molly’s, which is linked to the noun sister by means
of the possessive ’s. In the case of my uncle, the propositus argument is the speaker,
specified by the possessive pronoun my. Thus both NPs refer to an additional
argument, the speaker and Molly, respectively. We will call the additional arguments
of relational nouns relational arguments.
In addition to the relational nouns uncle and sister, (10) contains the two-place
verb marry. Thus the proposition expressed comprises three two-place predications:
that someone is an uncle of the speaker, that this someone is going to marry someone
else, and that this other person is the sister of Molly. An analysis in the style of Fig.
5.2 may help to illustrate how the predicates are related to their arguments. In Fig. 5.7
the subscript ‘NR’ is short for ‘relational noun’. The four arguments ai, au, am and as are
short for ‘I’, the speaker’s uncle, Molly and her sister, respectively.
Figure 5.7
Predication structure of sentence (1)
words
NPs
[ my uncleNR] is going to marryV [ [Molly’s] sisterNR ]
arguments
ai
au
am
as
The relational argument of two-place nouns is usually specified with a possessive
construction. In English, there are three common alternatives. In the simplest one,
a possessive pronoun precedes the relational noun (my uncle). Alternatively, a full
NP with a possessive ’s can be placed before the noun (the children’s uncle). Finally, a
‘possessor’ can be specified by an of-PP following the noun (an uncle of my mother);
in this case there is also the alternative of the so-called ‘double genitive’ (a friend of my
mother’s), where the possessor NP is in the genitive.They are counted as arguments
Predication
117
proper because the meaning of the respective nouns cannot be defined other than in
terms of two arguments. (Recall the discussion of relational nouns in 4.1.4.)
There are many more relational nouns: words like friend, neighbour, rival, boss;
abstract nouns like name (of), height (of), occupation (of); or linguistic notions
such as meaning (of), pronunciation (of), subject (of), argument (of). In some cases,
relational arguments are specified by non-possessive PPs, cf. ticket to, attack on,
discontent with, equivalence to. Some relational nouns have two additional arguments,
e.g. difference, relationship or distance or nouns derived from transitive verbs, such as
conquest, discovery or invention.
In English, possessive arguments of nouns are not syntactically obligatory. In
many languages, however, possessive constructions with relational nouns differ in
form from those with non-relational nouns. In the case of relational nouns one talks
of inalienable possession, and in the case of non-relational nouns of alienable
possession. (11) is an example from Lakhota, the language of the Sioux. Lakhota
has possessive prefixes, e.g. ni- for 2nd person singular. Possessive prefixes can only
be attached to relational nouns. This happens with the relational noun nági ›spirit‹
in (11a). Ninági means ›yoursg spirit‹ and instantiates the inalienable possessive
construction, ‘possessive prefix + relational noun’. By contrast, wowapi ›book‹ in
(11b) is a sortal noun. It cannot take a possessive prefix; but it can be turned into
a relational noun and then it can. The prefix thá makes a relational noun out of
sortal nouns. tháwowapi is relational and means ›book of‹. Prefixed with ni it yields
›yoursg book‹. nitháwowapi is an example for the alienable possession construction:
‘possessive prefix + thá + nonrelational noun’.
(11) Lakhota (Sioux, Robert D. Van Valin p.c.)
a.
inalienable
ni-nági ki
2S-spirit DEF
‘yoursingular spirit’
b.
alienable
ni-thá-wowapi ki
2S-REL-book
DEF
‘yoursingular book’
5.4.2 Major types of adjectives
5.4.2.1 One-place adjectives
Adjectives in their basic form, the ‘positive’ (recall 1.2.2), are mostly one-place
predicate terms; as we will see, this does not hold for the comparative form.
Grammatically, three ways of using adjectives can be distinguished: the attributive
use, the predicative use and the adverbial use. We will only deal with the attributive
and the predicative use here. In the attributive use, the adjective, within an NP, is4
attached to the head noun, e.g. a dubious company, a red balloon, the stupid driver.
As we saw in 5.2, the argument of the adjective in attributive use is the referential
argument of the noun.
4 In English, attributive adjectives precede the noun. In other languages, the adjective follows the
noun. For example, in Italian ‘stupid driver’ would be conducenteN stupidoA.
118
Understanding semantics
The second way of using adjectives is the predicative use. In the most common
predicative construction in English, the adjective is combined with the so-called
copula verb be, or a similar verb such as become, to form the VP of the sentence. In
such cases, the predication expressed by the adjective relates to the subject referent.
Thus, the subject NP is a complement of the complex expression consisting of the
copula verb plus the predicative adjective. Many languages do not use a copula with
predicative adjectives. They would combine the subject NP 5directly with an adjective
which serves as the grammatical predicate of the sentence.
(12) a.
b.
John is clever.
John umnyj.
‘John is clever’
(Russian)
‘John is clever’
(Japanese)
John clever
c.
John
wa
kashikoi.
John
TOP
clever
5.4.2.2 Two-place adjectives
Some adjectives, e.g. other than, different from, similar to, fond of, satisfied with,
keen on, have a second argument. It is invariably specified by a PP complement, i.e.
by a separate argument term. With respect to their first argument, these adjectives
behave like one-place adjectives. Combined with the PP complement they can only
be used in predicative constructions such as a copula sentence or the relative clause
construction in (13b) (which constitutes a second variant of predicative use). In
attributive use as in (13c), adjectives with complements are ungrammatical.
(13) a.
My uncle is very fond of Molly’s sister.
b.
She wore a sweater similar to yours.
c.
*She wore a similar to yours sweater.
A special case of two-place adjectives are one-place adjectives in their comparative
form. The comparative adds a further argument to the adjective, an entity the first
argument is compared to. Thus big is one-place, and bigger is two-place. The second
argument is specified by a than-PP, i.e. a complement proper. If the comparative is
formed of a two-place adjective, the result is a three-place predicate term (cf. (14c)).
(14) a.
Her hair was oilier than Julio’s.
b.
I hardly can imagine a book more boring than this one.
c.
Actually, my uncle is more fond of Molly than of her sister.
5 Russian has a copula for the past tense, Japanese uses a copula with one type of adjectives but not
with the type represented by kashikoi.
Predication
119
5.4.2.3 Non-predicational adjectives
You should be aware, however, that some adjectives do not serve as predicate terms:
for example, the ‘modal’ adjectives alleged, potential, actual, would-be. They can only
be used attributively: the alleged terrorist, a potential referent, etc. Their predicative
use is impossible because they do not express a predication of their own: it would
not make sense to say of something that it is ‘alleged’ or ‘potential’. Rather, modal
adjectives serve to modify the predication expressed by the noun by relativizing
whether or not it is true. An ‘alleged terrorist’ is someone for whom the predication
‘terrorist’ is not plainly true in the given CoU, but only ‘allegedly’ so, i.e. in some
different context. Similarly, a ‘potential’ referent is something for which the predicate
›referent‹ is only potentially true. A similar group of adjectives relates the predication
to different times or places, e.g. former, future or local, etc. These are instruments
of temporal or spatial deixis (recall 4.2 and 4.3). Yet another different type of
non-predicative adjectives is represented by combinations such as atomic war,
educational disaster, doctoral dissertation. Here, too, the adjective does not provide
a predication of its own (the war is not ‘atomic’, the disaster not ‘educational’, etc.).
Rather the adjective modifies the meaning of the noun by adding a specification to
some conceptual component of its meaning: ›war with atomic weapons‹, ›disaster in
education‹, ›dissertation written to acquire a doctor title‹, and so on. Consequently,
not all adjectives are predicate terms, and not all combinations of an attributive
adjective and a noun express a joint twofold predication.
5.4.3 Nouns in predicative use
Not only adjectives but also NPs, and therefore nouns, can be used predicatively.
Predicative NPs do not refer.
(15) John is a teacher.
In this sentence, only the NP John has a referent. Syntactically the subject NP is a
complement of the copula. Semantically it is passed on, as it were, to the predicate
term a teacher that forms the VP with the copula, and thus serves as an argument
term for teacher. The same mechanism applies for predicative adjectives. The copula
itself is not a predicate term; it is a linguistic means of forming a predicate term out
of an NP or adjective. The argument of the resulting predicate can then be specified
in the subject position. Thus, for example, the construction in (15) can be used to
apply two NP predications to the same argument.
Relational arguments of predicative nouns are specified in the same way as in their
referential use. The predicative use does not affect the syntactic structure of the NP.
(16) a.
b.
This is my uncle.
This is a ticket to Novgorod.
Table 5.3 displays the main types of predicate terms and the way in which their
arguments are specified in the sentence. Predicative uses are not included. Verbs,
120
Understanding semantics
nouns and adjectives differ in the way their first argument is integrated into the
sentence structure, but they are similar in that further arguments take the form of
complements, usually NPs or PPs. For the second argument of relational nouns there
is a special group of (possessive) constructions.
Table 5.3
Types of predicate terms and how their arguments are specified
Type
First
argument
Further arguments
verb intransitive
the bell rang
complement
–
verb transitive
she opened the door
complement
complement
noun 1-place
the postman
referential
–
noun relational
asking her name
referential
possessor
a letter from Johnny
referential
complement
adjective 1-place
a pink envelope
parasitic
–
adjective compar.
thicker than the last one
parasitic
complement
adjective 2-place
full of promises
parasitic
complement
Before the topic of predication is continued, a brief section about predicate logic
is inserted. Ultimately derived from the predicational structure of natural language
(i.e. Ancient Greek), more than two thousand years ago by Aristotle, it is a formal
system widely used in semantics for the analysis and representation of sentence
meaning.
5.5 PREDICATE LOGIC NOTATION
In so-called predicate logic (PL), a simple notation has been developed for
predication. The basic expressions of predicate logic are one-place, two-place, etc.
predicate constants (i.e. predicate terms that do not have the status of variables) on
the one hand, and so-called individual terms on the other. Individual terms serve
as argument terms and are interpreted as referring to ‘individuals’, where these are
whatever may serve as an argument of one of the predicates. There are two sorts of
individual terms, individual constants and individual variables. Roughly speaking,
individual constants can be thought of as proper names and individual variables
as something like 3rd person personal pronouns, i.e. expressions that refer to some
particular individual determined by the CoU. For current purposes, we define a
predicate logic language with the following basic expressions (again, referential verb
arguments are disregarded).
Predication
121
6
one-place predicate constants
cat, money, dubious, company, sleep
two place predicate constants
marry, sister, uncle
three place predicate constants
send_to
individual constants
j [for Johnny], m [for Molly], i [for the speaker]
individual variables
x, y, z
The predicate terms are combined with argument terms in a uniform way. In the
most common notation, predicate terms are followed by the appropriate number of
individual terms enclosed in parentheses and separated by commas. The following
would be simple PL formulae with the meanings indicated.
(17) a.
money(x)
x is money
b.
uncle(j, m)
Johnny is an uncle of Molly
c.
send_to (j, x, y)
Johnny sends x to y
The notation of predicate logic reflects the view that predicates are incomplete
propositions with empty slots for arguments. When combined with the appropriate
number of argument terms, a predicate term yields a formula (a sentence in PL).
Formulae can be combined to form more complex
formulae by using sentential
7
connectives such as the conjunction Ÿ ‘and’. This allows us to analyse the
predicational part of a natural language sentence by ‘translating’ it into a predicate
logic formula. The single predications are connected by truth-conditional conjunction.
In the following examples, referents not specified by a proper name are represented
by variables. Tense, aspect (progressive form of the verb) and articles are neglected.
(18) a.
The cat is sleeping.
8
cat(x) Ÿ sleep(x)
b.
Johnny sent money to a dubious company.
send_to(j, x, y) Ÿ money(x) Ÿ dubious(y) Ÿ company(y)
c.
My uncle is going to marry Molly’s sister.
uncle(x, i) Ÿ sister(y, m) Ÿ marry(x, y)
The method makes transparent which predications a sentence contains, to which
arguments they are applied and how the different predications interact by sharing
arguments. In chapter 13, predicate logic will be treated in more depth and detail
and the method of semantic analysis by translation into a formal language will be of
central concern.
6 It is common practice to use bold type for predicate and individual constants that correspond to
natural language words.
7 More conjunctions will be introduced in 7.4 on sentential logic
8 A formula such as ‘sleep(cat)’ would violate the rules of PL syntax, because the predicate term cat
is not allowed to be used as an individual term in the argument position of sleep.
122
Understanding semantics
5.6 THEMATIC ROLES
5.6.1 The arguments of verbs as thematic roles
The different arguments of a verb predicate are referred to as its roles or
participants. A transitive verb has two roles, for example the eater and the thing
eaten, the opener and the opened, the helper and the helped. Grammar consistently
distinguishes the different roles of a multi-place verb. When the verb eat is used
in its active voice, the eater is always specified by the subject of the verb eat and
the thing eaten by its direct object. The analogue holds for multi-place nouns and
adjectives.
An important question then is whether there is cross-verb consistency in the way
the roles of verbs are marked. Is there something common to the role of the eater
and the role of the helper that is responsible for their appearing in the subject
of the sentence? Is there something in common to all subjects, or to all direct or
indirect objects? Can the roles that are encoded in tens of thousands of verbs be
consistently categorized into a small number of abstract roles? Are these abstract
roles universally applicable to all roles of all verbs in all languages? Semanticists
and syntacticians have tried to answer these questions positively. There is a good
chance of succeeding, but things are not straight and simple. A first look at the
data clearly shows that the subject does not always denote the same role. Consider,
for example, intransitive and transitive open in the following sentences:
(19) a.
The doorO opens.
b.
This keyI opens the doorO.
c.
The childA opened the doorO.
d.
The childA opened the doorO with her own keyI.
While these sentences represent different concepts of opening, it is intuitively
clear that they all fit into one scheme with three roles: (i) an animate agent A
opening something; (ii) an object O that becomes open, (iii) an instrument I
used to open O. In (19a), the subject specifies O, in (b) I and in (c) and (d) A.
Conversely, O is specified by the subject in (a) but by the direct object in (b–d).
The instrument I appears as the subject in (b) and as a prepositional adjunct
in (d). The patterns do, however, exhibit certain regularities: if A is specified, it
appears as the subject. O is always specified in the object position as long as it is
not the only argument term.
Since the first attempts, back in the 1960s, at establishing a set of universal roles,
many theories have been developed in different theoretical frameworks. It is now
common practice to speak of thematic roles ( -roles, theta-roles, with the Greek
letter ‘theta’ for thematic) or semantic roles. Some draw a distinction between
thematic roles and semantic roles, but the difference need not concern us here. The
Predication
123
Table 5.4
Thematic roles
Role
Description
Examples
agent/actor
performs the action expressed by the verb, controls the
event
Johnny wrote a love letter
the cat has eaten the egg
she gave me the keys
you put the keys on the desk
my uncle marries Molly
theme/
patient
undergoes the action/change/event expressed by the verb Johnny wrote a love letter
the cat has eaten the egg
she gave me the keys
you put the keys on the desk
my uncle marries Molly
the door opened
the snow is melting
experiencer
experiences a perception, feeling or other state
I heard him
the outburst surprised her
instrument
an instrument, or a cause, by which the event comes
about
this key opens the door
he opened the door with a key
she was shaking with fear
locative
a location
the keys are on the desk
goal
goal of a movement
put the keys on the desk
path
path of a movement
she rode through the desert
inventory of thematic roles differs from theory to theory, but the roles in Table 5.4
9
are uncontroversial.
General thematic roles are useful in several respects. For example, they help to
account for the meaning relations holding between the three different verbs open
used in (19a), (19b) and (19c, d), respectively. For predicate terms, a description of
their arguments in terms of roles, their so-called argument structure constitutes
an important part of their distinctive properties. Thematic roles also allow a proper
description of phenomena like passive, which change the argument structure of a
verb in a specific way (see 6.1).
5.6.2 Linking
The mechanism by which a language distinguishes the different arguments of
predicate terms is called linking. We will not go into this complex matter here. As far
as English is concerned, some simple linking rules can be stated that were already
indicated in connection with (19): an AGENT role always appears in subject position,
9 We follow the practice of writing thematic roles with small capitals (THEME etc.).
124
Understanding semantics
a THEME can only be the subject if no AGENT is present. These rules hold for active
sentences. In passive sentences, THEMES and other roles can be made the subject of the
sentence. The AGENT complement is deleted, but it can be specified by an additional
by-PP.
(20) a.
b.
active:
The dog (AGENT, subject, obligatory) opened the door
(THEME, direct object, obligatory).
passive:
The door (THEME, subject, obligatory) was opened [by
the dog (AGENT, PP, optional)].
The effective linking of argument terms depends on a proper grammatical distinction
between the subject, the direct objects and other ‘grammatical functions’. In English
sentences, the subject and the direct object differ in three ways, which illustrate three
general linking strategies to be observed in the languages of the world.
∑
Agreement. The finite verb ‘agrees’ with the subject NP in certain
grammatical properties.
In English, the finite verb agrees with the subject NP in grammatical person (1st, 2nd
and 3rd) and number (singular or plural). Agreement shows up in the 3rd person
singular -s of full verbs in the present tense (he/she/it speaks vs I/you/we/they speak_)
and in the forms of the verbs be and have (I am, you are, she is; I was, we were; have vs
has). In other languages, e.g. Georgian and Hungarian, the verb may in addition agree
with the direct object. Yet other languages such as Chinese, Korean and Japanese do
not exhibit agreement between the verb and its complements.
∑
Case. The complements of the verb are distinguished by case.
In English, the subject is in nominative case, the object in objective case. However,
the difference shows up only with some pronouns (I, he, she, we, they vs me, him, her,
us, them and who vs whom). Prepositions, too, can be subsumed under this linking
strategy. As a rule, there is a small set of cases and/or prepositions, e.g. nominative,
objective and to in English, that are used for those complements which specify the
arguments proper. Additional complements that specify original arguments – such
as an instrument of eating or a vehicle of motion – are marked with prepositions or
cases that are semantically more specific. For example, the role of the instrument in
English is marked with by or with (19d). With is not restricted to marking instrument
terms; it is polysemous. The PPs in (21) are specifications of special roles not
contained in Table 5.4:
(21) a.
b.
Johnny ate the salad with disgust.
Johnny ate the salad with his friend.
Predication
∑
125
Word order. Often different positions in the sentence are provided for
different grammatical functions, in particular for subject and direct object.
In English, the subject NP precedes the finite verb, the direct object follows it. Since
English has only rudimentary case marking and agreement, word order is its main
linking strategy. Some languages, such as Chinese, that lack case and agreement
altogether rely completely on word order as grammatical linking mechanism.
Of course, grammar is not the only resource for determining which roles of the
verb are specified by which complements in the sentence. If the grammatical linking
information is insufficient, the assignment of roles is a matter of semantics or context.
For example, the German sentence (22) contains two singular NPs (Kaffee and John)
which could both be any case except genitive. The verb is 3rd person singular, like
both NPs, too. Unlike English, German allows, under certain circumstances, the
inversion of subject and object. Thus the grammatical information is insufficient
for distinguishing between the AGENT complement and the THEME complement. In
normal contexts, the sentence will of course be interpreted as saying that John drank
coffee rather than the other way round.
(22)
Kaffee
trank
John
coffee
drank
John
If the determination of the roles is left to semantics or context, this is not a case of
linking. The notion of linking is restricted to grammatical means of role distinction.
5.7 SELECTIONAL RESTRICTIONS
A predicate cannot be applied to arbitrary arguments, and consequently, a predicate
term cannot be combined with arbitrary complements. In addition to the requirements
of grammar and conditions such as the Principle of Consistent Interpretation (3.4.4),
complements must fulfil certain semantic, or logical, conditions and this leads to
restrictions on possible argument terms. The discussion here will be restricted to
verbs, but analogous considerations apply to adjectives and nouns in predicative use.
5.7.1 Selectional restrictions of verbs
Two somewhat strange examples may illustrate what kinds of conditions are involved.
(23) a.
b.
The cook has murdered an eggplant.
The potatoes are frying the cook.
If taken literally, the two sentences describe impossible situations. The verb murder
requires a living being as its THEME/PATIENT argument, usually a human being. Only
a minority of speakers of English would use the verb murder also for the killing of
126
Understanding semantics
animals, probably none would use it for plants or even fruits. Likewise, the verb fry
requires an AGENT argument capable of acting. It need not be a person – one could
imagine an animal putting, and keeping, something on a hot stone in order to get
it fried. But potatoes cannot fill this role. There are fewer problems with the THEME
argument of (23b). Although highly unlikely, the THEME of a frying event can be a
person: persons are friable. But the THEME role too underlies logical restrictions. For
example, words, numbers, properties or addresses cannot be fried.
The logical conditions on arguments are called selectional restrictions (also
‘selection restrictions’). These were briefly introduced in 4.7.2.2. The term is
motivated by the idea that a predicate term selects, and thereby restricts, the range
of possible arguments. Let us assume, for example, that the verb vaccinate requires a
human being as its AGENT. Then, in appropriate CoUs, the following sentences comply
with the selectional restrictions:
(24) a.
b.
The doctor himself vaccinates John.
The next one vaccinates John.
The choice of the subject term in (24a) guarantees that the selectional restrictions
are fulfilled: doctors are persons. It is, however, not necessary that the noun in the
complement NP entails human being. The selectional restrictions only require that
the referent of the complement is a person. The potential referents of the subject NP
the next one in (24b) are by no means necessarily persons: the NP can refer to almost
anything, because the pronoun one can replace an arbitrary count noun. But if the
subject refers to a person in the given CoU, (24b) is semantically correct.
To work out the selectional restrictions of a particular predicate term can be a very
difficult task. Take, for example, the THEME argument of the transitive verb open. What
kinds of things can be opened? We can open a door, e.g. by sliding it open, and thereby
create an opening in the wall that can be passed through. We can open a room by
opening a door to the room. We can open our mouth. We can open our eyes. Or our
arms. Or a fist. We can open a bottle or a tube by removing or opening its lid. We can
open an envelope by slitting it open, or a letter by unfolding it, or a book. We can open
a bank account or a business. We can open a ceremony. We can open a computer file.
We can open perspectives. These are not only different kinds of objects, but in addition
open in almost each case means something slightly different. If I open a door or a lid,
the argument that is ‘opened’ is moved, or removed, in order to create an opening in the
enclosure of a spatial region (e.g. a room or the interior of a bottle). If I open a bag, or
an envelope, the THEME argument is the enclosure and the result of the act is an aperture
in the enclosure. If I open a room, a trunk, a garden, a shop, a box, or my mouth, I refer
to an object which has an enclosure in which then an aperture is brought about. So,
actually, there are two or three different roles involved in these kinds of opening events:
(i) a spatial region (e.g. a room) which is made accessible; (ii) its enclosure; and (iii),
possibly, a (re)movable part of the enclosure that provides, or blocks, access to the
region. Each of these roles goes with its own selectional restrictions. In the variants
open a fist, open a letter, open the arms, open the wings (of a bird), the THEME arguments
Predication
127
play yet a different role and the verb is used to express a different kind of process
similar to spreading or unfolding. In the variant represented by open a ceremony, the
selectional restrictions of the THEME argument require an event or a procedure. Opening
it, we start the ceremony and thereby ‘enter’ a certain sequence of events. Yet other
selectional restrictions govern the use of open in I opened the style file, you must open a
bank account or she opened a computer business.
It would be wrong to conclude that the selectional restrictions for the THEME
argument of transitive open are so general that they cover almost anything. Rather
the verb is multiply polysemous. In each of its meaning variants, the verb expresses
a different process with respect to its THEME argument and imposes different
selectional restrictions. If we take these restrictions seriously, we will be able to
explain how the meaning variants are related, e.g. the readings of open in open one’s
eyes and open one’s eyelids.
5.7.2 The process of fusion
The combination of a predicate term with a complement results in two sources
of information about the argument. First, the complement provides an explicit
specification of it. Second, the predicate contributes implicitly the selectional
restrictions for the argument. These two pieces of information are conjoined when
the meaning of the sentence is composed. We can think of the process as logical
conjunction (i.e. combination by and) of the two pieces of information. Let us call the
process fusion, adopting a term from Jackendoff (1990). It is illustrated in Fig. 5.8.
Figure 5.8
The mechanism of argument description
predicate term
contributes
sentential context
contributes
SELECTIONAL RESTRICTION
ARGUMENT SPECIFICATION
fusion
TOTAL DESCRIPTION
OF THE ARGUMENT
Consider the following simple example (25):
(25) The dog opened it.
The THEME of the opening is merely specified as ‘it’ – which might refer to anything
inanimate. In addition, the referent of it must fulfil the selectional restriction of
128
Understanding semantics
being something ‘openable’. This restricts the range of possible referents of the
pronoun it considerably. In cases such as these, the selectional restrictions contribute
substantially to the description of the argument.
The same does not hold for the AGENT argument the dog in (25). The specification of
the AGENT as a dog is so specific that the selectional restrictions for the AGENT argument
are thereby automatically fulfilled: dogs in general are potential AGENTS of openings.
Here, the selectional restrictions do not add to the specification of the argument.
A sentence like (26) provides a third possibility.
(26) She will marry next June.
The predicate ›marry‹ requires a person as an AGENT; the pronoun she contributes
that the subject argument is female (not necessarily a person) or one of those rare
nonhuman items in English (such as ships) for which feminine gender is used.
Fusion of the restriction ‘person’ with feminine gender yields ‘female person’ as
specification of the argument.
Finally, the fusion of the two descriptions of an argument may lead to a conflict –
if they are incompatible as in (23a) the cook has murdered an eggplant or (23b) the
potatoes are frying the cook. In such cases, fusion results in a logical contradiction.
To sum up, fusion may have different results:
(i) If the selectional restrictions are not more specific than the argument
specification, the total description is identical with the argument specification.
(ii) If the selectional restrictions are more specific than the argument specification,
the total description is identical with the selectional restrictions.
(iii) If the selectional restrictions and the argument specification neither entail
(cases i and ii) nor contradict each other (case iv), fusion results in a total
description of the argument which is more specific than both sources.
(iv) If the selectional restrictions and the argument specification are incompatible,
the total description is contradictory, i.e. inapplicable to any concrete
situation.
In actual communication, contradictions will not be accepted (except as examples for
contradictions as in the current context). Application of the Principle of Consistent
Interpretation (3.4.4) will either eliminate the corresponding reading or lead to an
adjustment with a meaning shift. The principle applies because a predicate term and
its complements constitute immediate sentential context for each other. For example,
in the case of (27), we would assume that the verb drink imposes the selectional
restriction ‘liquid’ on its direct object argument:
(27) She drank the coffee.
The object NP the coffee, taken for itself, does not necessarily denote a liquid. ‘Coffee’
can also be coffee powder, coffee beans or coffee shrubs. In (27c), the selectional
Predication
129
restriction ‘liquid’ rules out all but the drink variant of the meaning of coffee. The
result is a disambiguation of the noun coffee.
Conversely, we can observe disambiguation of the verb:
(28) a.
b.
She corrected her uncle.
She corrected the error in the style sheet.
In (28a) the specification of the THEME/INSTRUMENT argument requires a meaning
variant of the verb in which it can be applied to persons, ruling out a reading such as
the one required in (28b).
5.7.3 Selectional restrictions and meaning shifts
If fusion leads to contradiction, it may be possible to obtain an admissible reading
by means of a meaning shift. Let us first consider an instance of a metonymical shift
(recall 3.4.3).
(29) Moscow declares the Chechen rebels defeated.
The verb declare requires a human being or organization as its AGENT argument, but
the given specification Moscow is the name of a geographic entity. We will change the
meaning of Moscow by a metonymical shift (location → institution located there)
in order to meet the selectional restrictions of the verb. Note how the selectional
restrictions serve as a guide that indicates the direction of the meaning shift.
In the case of metaphorical shifts, it is often the predicate term whose meaning
is shifted, as can be observed in (30). The literal meaning of the verb evaporate
requires some kind of physical substance. The subject, however, refers to a mental
state. To remedy the conflict, the verb meaning is shifted to the more general meaning
›vanish completely‹ which carries selectional restrictions that allow for this kind of
argument.
(30) His courage evaporated.
The processes of metaphor and metonymy regularly affect selectional restrictions.
If an argument term undergoes a metonymical shift, the resulting referent usually
is of a different logical sort, cf. ‘university’ as a location vs ‘university’ as an
institution vs ‘university’ as the university personnel. Likewise, metaphorical
interpretation of an argument causes a shift into a different conceptual domain,
usually also of a different logical sort, e.g. when ‘money’ is conceived of as a liquid
that may ‘flow’. If a predicate is interpreted metaphorically, as e.g. evaporate in (30),
the selectional restrictions change too, as to match with the sort of objects that
make up the target domain.
130
Understanding semantics
5.7.4 Semantic irregularity
The massive occurrence of meaning shifts in the interpretation of actual sentences
blurs a question that is central to semantic analysis, the question of semantic
irregularity. The notion of selectional restrictions provides us with one clear type
of cases: if a specification of an argument in the sentence is logically incompatible
with the selectional restrictions, then the construction is semantically irregular.
Simple as this seems to be, we have seen that semantic regularity is a question of
the readings assumed for the predicate term and its argument specifications. For
instance, sentence (23a) above, the cook has murdered an eggplant, is semantically
irregular only if we assume the original lexical meanings of murder and eggplant.
It becomes regular, i.e. interpretable, if we allow an appropriate meaning shift of
either the verb or the direct object. (Possible interpretations are left up to your
imagination.) It therefore makes more sense to avoid the simple notion of semantic
acceptability and to replace it by a description of the conditions under which it is
possible to make sense of a complex expression. We may then distinguish between
degrees of acceptability, such as (i) interpretable on the basis of the lexical meanings
of all components, (ii) interpretable by means of common types of meaning shifts,
(iii) interpretable only by means of uncommon types of meaning shifts. Probably, a
fourth category, ‘interpretable by no means at all’, does not exist.
5.8 SUMMARY
This chapter focused on predication, the semantic function of the main word
classes, verbs, nouns and adjectives. Built into a sentence, each of these ‘content
words’ adds a predication to the total proposition, about one or more referents.
The three major word classes differ in how their arguments are specified. Verb
arguments, except for the event argument, are specified by complements, i.e.
separate syntactic constituents with a referent of their own. One-place nouns are
mainly used as referring expressions that predicate about their referents. One-place
adjectives are parasitic for their argument. One of the most important insights
concerning sentence meaning is the fact that the predications contained in a
sentence are interconnected by argument sharing. If you take a look back at the
examples, in particular the analyses in Fig. 5.2 and Fig. 5.7, you will see that the
network of predications includes all referents of the sentences. It is this network
structure that makes a sentence a coherent semantic unit. The verb has the key
role in this structure. It occupies the centre of the network like a spider in its web
holding all the threads. This role of the verb corresponds to the fact that most verbs
are two- or multi-place predicates.
Since the meaning of a sentence is a network of predications about its referents
(including reference time and event referents), the sentence as a whole can be
considered one complex predicate expression about the situation referred to. For
example, sentence (1) renders a complex predication about a situation: an event e
Predication
131
takes place, at some time t; e is an event in the past (cf. past tense); it is an event of
sending that involves three referents rj, rm and rc, with rj the AGENT, rm the THEME and
rc the GOAL of sending; rj is someone called ‘Johnny’, rm is money and rc is a company
and dubious. This is what the sentence, due to the sophisticated grammar of English,
is able to express in not more than seven words. Slightly accommodating Fig. 5.6 for
the application to sentences, we obtain the picture 10in Fig. 5.9 for the sentence as a
complex predication about the situation referred to.
Figure 5.9
The sentence as a complex predication
sentence
the proposition,
a complex predicate
the situation
referred to
The study of predication also sheds light on the mechanism of composition.
First, composition is for the most part a matter of integrating all the predications
contained in the word meanings into an overall structure. Second, we have seen how
predicate terms and complements interact in the process of fusion in providing the
total description of the argument. We will later return to the underlying mechanisms
of composition that bring about the complex network of predications contained in a
sentence (cf. 12.4 and 13).
EXERCISES
1. What is the difference between a predicate and a predicate term, between an
argument and an argument term? Discuss the four notions and explain how they
are connected.
2. Discuss the ways in which arguments are specified for verbs, nouns and
adjectives. For nouns, distinguish between referential and predicative uses. For
adjectives, relate to attributive and predicative uses. For predicative uses, assume
a simple copula construction.
10 Actually, the sentence carries even more meaning than this: the grammatical meanings expressed
by aspect and mood of the verb and the meaning of the declarative sentence type.
132
Understanding semantics
3. Give an analysis in the style of Fig. 5.7 of the following two sentences.
a. Arthur opened the door of the kitchen.
b. My daddy met your mother in the supermarket.
4. Express the predications contained in the two sentences in a predicate logic
formula like those in (18). Use the individual constants a, i and u for Arthur, the
speaker and the addressee, respectively.
5. Determine the thematic roles of the verb arguments in the two sentences.
6. The following examples illustrate three types of predicative NP constructions.
Discuss the ways in which the NP a born loser gets its argument in each type of
construction.
a. Peter is a born loser.
b. She called him a born loser.
c. Peter, a born loser, managed to get fired within three hours.
7. What are the linking rules for the following relational nouns? In c, refer to the
respective arguments of the verb discover.
a. distance
b. difference
c. discovery
8. Try to describe the correlations between the two uses of the verbs provide, wipe
and load in terms of thematic roles and grammatical relations (which role
appears as subject etc.):
a. She provided Sheila with boots.
vs She provided boots for Sheila.
b. He wiped the crumbs off the table. vs He wiped the table.
c. She lost.
vs She lost the match.
9. Try to formulate the selectional restrictions (consult the online Oxford English
Dictionary http://oxforddictionaries.com/ for meaning variants and their
definitions, and discuss all of them):
a. of the verb conquer for its direct object argument
b. of the adjective yellow for its only argument
10. Explain in terms of selectional restrictions how the metonymical readings of the
following sentences come about:
a. The university lies in the eastern part of the town.
b. The university has closed down the faculty of agriculture.
c. The university starts again on April 15.
Predication
133
FURTHER READING
Givón (1993, ch. 3) for an extensive discussion of verb types and thematic roles.
Tallerman (2011, ch. 6) on the basic facts of linking and ch. 7 on passive and causative
constructions. Radford (1988, ch. 4) on complements and adjuncts, ch. 7 on thematic
roles and selectional restrictions. You will find a broader discussion of semantic
roles in Kearns (2011, ch. 10), Riemer (2010, ch. 10) and Van Valin (2005, 2.4). Palmer
(1994) is a monograph about the topic. For a recent survey see Davis (2011).
6
?
Verbs
Nouns and verbs are the two basic word classes of language. Almost every sentence
contains both. Verbs are primarily used for predication, and nouns for reference,
where the referring NPs provide the arguments for the predication expressed with
the verb. Chapter 4 dealt with the major aspects of reference and NP semantics. Now
we will turn to the semantics of verbs.
As we saw in the previous chapter, sentences may contain several nested
predications. One of them is the central predication around which the whole sentence
is organized. Recall the first example from chapter 5, repeated in (1a) below. The
sentence contains five predications. The central predication is the one expressed by
the verb sent, while the other predications, expressed by Johnny, money, dubious and
company serve to specify the arguments of the central predication. It is the central
predication that is subject to negation if the sentence is negated, as it is in (1b). And it
is the central predication to which the temporal information applies that is expressed
by aspect and tense, which are not accidentally realized by the choice and the form of
the verb, in this case past perfective (see 6.3 and 6.4 below).
(1)
a.
Johnny sent money to a dubious company.
b.
Johnny didn’t send money to a dubious company.
There are languages which, in certain cases, can express the central predication
without using a verb. These would be predications by means of adjectives, nouns or
PPs in languages which do not use a copula. For example, in Russian the equivalents
of the English sentences I’m a linguist, I’m in Moscow, I’m stupid would just be
expressed as ‘I – linguist’, ‘I – in Moscow’, ‘I – stupid’. Most sentences, however,
contain a verb for their central predication. In many languages, including English,
there are no sentences without a verb.
In all languages, verbs constitute a large part of the lexicon; English has tens
of thousands of them. Verbs express actions to walk, processes to melt, events to
explode and states to know. The situations expressed are of a temporary nature, they
may apply for long or short periods; they are located and ordered in time. Also the
situations may have an internal temporal structure such as one state changing into
another, an activity coming to an end or a process beginning. Therefore, temporal
aspects of predications with verbs are one central issue of verb semantics. We will
turn to them in the sections 6.2 to 6.5 on situation structure, aspect and tense. The
other central issue is the fact that verbs almost always express a predication about
not only their own referent but also about one or more additional arguments. In
Verbs
135
this respect they can be compared with relational nouns (5.4.1.2). Thus, while most
nouns are one-place predicate terms, most verbs are multi-place. The situation to
which a verb refers does not exist independently, but it manifests itself in certain
temporary conditions of the participants. The only exceptions are zero-place verbs
such as rain or thunder which apply immediately to the situation referred to. Verbs
differ in the number of arguments and the roles these inhabit (5.3). We will turn to
general aspects of argument structure in a minute.
Investigating the argument structure of verbs on the one hand and the temporal
structure of the situations expressed on the other will enable us to distinguish major
sub-classes within the vast number of verbs. In addition, this chapter will deal with
grammatical meaning in more depth than any other chapter of the book. We will
treat three general issues of verb grammar: diatheses (such as passive), aspect and
tense. What we will not deal with is the complex issue of mood and modality, i.e. of
elements like the modal verbs and auxiliaries written in italics in (2):
(2)
a.
She can/may/must/shall go home.
b.
She could/might/should/would go home.
Dealing with general aspects of verb semantics, you will realize a very important
general trait of language: language is not just a matter of depicting reality one-toone, but rather a matter of describing reality selectively and from one particular
perspective out of many possibilities.
6.1 ARGUMENT STRUCTURE, DIATHESES AND ALTERNATIONS
6.1.1 Argument structure
Verbs not only differ in the number of arguments, but also in the constellations
of thematic roles. Action verbs have an agent argument, others do not. Verbs of
movement have locational arguments such as GOAL, SOURCE or PATH, in addition to
the role of the AGENT or THEME moving. Verbs of sensation or perception have an
EXPERIENCER argument. Among the possible arguments of a verb predication, there
are core arguments (one, two or at most three). These are typically syntactically
obligatory and appear as subject, direct object or indirect object. AGENT and THEME/
PATIENT are always core arguments. The other arguments are of a more peripheral
nature; they appear as oblique complements (or adjuncts), for example in the
form of prepositional phrases. They are typically optional and not included in the
lexical verb concepts; such peripheral optional arguments include, for example,
INSTRUMENTS. Syntax provides for a privileged status of one argument of the verb,
featuring it as the subject of the sentence. There is a hierarchy among the arguments
of a verb. If there is an AGENT or EXPERIENCER, i.e. a participant that is typically
human, it will appear as the subject of the sentence. In the absence of such a role,
other arguments may be assigned the subject status. Recall the variants of the verb
open in (19), chapter 5: if there is an AGENT involved, it will appear as the subject; in
136
Understanding semantics
absence of an AGENT, an INSTRUMENT, if involved, will assume subject status; if there
is neither an AGENT nor an INSTRUMENT specified, lowest ranking THEME appears as
subject.
6.1.2 Voice, diatheses and alternations
The core argument structure of a verb can vary and it can be changed by certain
grammatical operations. We encountered verbs with varying argument structure in
chapter 5 (eat, give and open). These variations are technically called alternations.
Alternations come with different syntactic constructions, e.g. with or without a direct
object, but without a change of the form of the verb; we will turn to them in 6.1.3. The
more general term for changes of argument structure is diathesis (plural: diatheses);
it also covers morphologically marked variation of the argument structure, like
the passive in English. We will briefly have a look at four diatheses: passive and
antipassive, causative and anticausative.
6.1.2.1 Passive and antipassive
1
We mentioned the passive in 5.6.1. In English, the common passive is formed by
combining the past participle with the auxiliary be. It is only possible with agentive
verbs with a further argument. One of the further arguments becomes the subject
argument, while the subject of the active sentence is either dropped or turned into a
by-PP. (3) gives the passive versions of some sentences.
(3)
a.
active: The boy (AGENT) watched the man (PATIENT).
passive: The man (PATIENT) was watched [by the boy (AGENT)].
b.
active: She (AGENT) threw the letter (THEME) into the shredder (GOAL).
passive: The letter (THEME) was thrown into the shredder (GOAL) [by her
(AGENT)].
c.
active: She (AGENT) taught Dutch (THEME) to the immigrants (RECIPIENT).
passive: Dutch (THEME) was taught to the immigrants (RECIPIENT) [by her
(AGENT)].
d.
active: She (AGENT) taught the immigrants (RECIPIENT) Dutch (THEME).
passive: The immigrants (RECIPIENT) were taught Dutch (THEME) [by her
(AGENT)].
The three-place verb teach alternates between two constructions that correspond to
different passive constructions. The application of passive has two effects: a former
non-subject argument NP is ‘promoted’ to the privileged function of subject. At the
same time, the former subject is ‘demoted’; it is either dropped altogether or degraded
to an optional oblique argument, peripheral to the predication. Thus, the number of
1 There is also the so-called get-passive as in He got injured in the fight, which will not be discussed
here.
Verbs
137
core arguments is reduced by one. In German, the number of core arguments can be
reduced to zero, if the verb has an AGENT as its only argument. There are sentences like
those in (4); in English the same effect can only be achieved with impersonal active
constructions using they, one, you and other impersonal expressions as subject. This
is a secondary way in English to demote the subject argument.
(4)
a.
Hier wird gearbeitet.
here work.PASS.PRS.3SG
lit. ‘here is worked’ = ‘people/they/you work here’)
(German)
b.
Gestern wurde gestreikt.
yesterday strike.PASS.PAST.3SG
lit. ‘yesterday was striked’ = ‘they went on strike yesterday’
(German)
The antipassive in English consists of demoting the direct object argument by
omitting it. It removes the THEME/PATIENT argument from the set of core arguments.
English does not mark the antipassive by a different form of the verb, but there
are languages that do. Thus, the antipassive in English is a matter of alternation of
certain verbs. We already looked at one example of this in the previous chapter: the
use of eat without a direct object:
(5)
active:
antipassive:
She (AGENT) ate her cereals (THEME).
She (AGENT) ate.
The omitted THEME/PATIENT argument is interpreted as ›something/someone‹
depending on the selectional restrictions of the verb for this argument. The
antipassive alternation with eat is responsible for the inconclusive status of the
object argument (recall the discussion in 5.3.3 on the number of arguments of
the verb). In fact, eat can be used in different voices. In the active voice, AGENT and
THEME are both obligatorily specified, by subject and direct object, respectively;
in the antipassive voice, only the THEME is specified, by the subject; in the passive
voice, the THEME obligatorily appears as subject, while the AGENT may optionally be
specified with a by-phrase. Thus, the number of arguments of a verb is not just a
matter of the verb as it is lexicalized, but a matter of the verb and the voice in which
it is being used.
The antipassive is not possible with all transitive verbs. A minimal pair often
quoted is eat vs devour, which does not allow omission of the object.
6.1.2.2 Causative and anticausative
The causative diathesis has the opposite effect of passive: it adds a core argument.
The causative of drink, for example, means ›let drink‹ or ›have drink‹. It introduces
a new AGENT, the ‘causer’ and turns the original subject argument into a PATIENT, the
‘causee’. The causer appears as the subject of the sentence, the causee as direct or
indirect object. (6a, b) are examples from Japanese. Japanese has a suffix -(s)aseattached to the stem of the verb which turns it into a causative:
138
(6)
Understanding semantics
a.
b.
kodomo ga
biiru
child
NOM
beer
‘The child(ren) drank beer’.
o
ACC
nonda.
drink.PAST
obâsan
ga
kodomo ni
biiru
grandma NOM
child
DAT
beer
‘Grandma had/let the child(ren) drink beer.’
(Japanese)
o
ACC
nom- ase- ta
drink- CAUS PAST
The causative form in Japanese has a weak reading of allowance and a strong reading
of causation. English has no grammatical causative. There are three options of
expressing causation: (i) using, if possible, the same verb in a causative construction:
intransitive open > transitive open; (ii) using a corresponding lexicalized causative
verb: eat > feed; (iii) combining the verb with auxiliary-like have or let: eat > have eat,
let eat. Neither (i) nor (ii) constitute general options for English verbs. By contrast,
the Japanese causative can be applied to almost all verbs. The third type of English
construction, too, is generally available; let carries the weak meaning, have the
strong meaning, but lexically causative verbs like feed ‘make eat’, teach ‘make learn’,
show ‘make see’, kill ‘make die’ and very many others only have the strong reading of
causation. The same holds for lexically causative verbs in Japanese, for example for
oshieru ‘teach’, korosu ‘kill’, miseru ‘show’, etc.
In Japanese, the passive can be applied to causatives. First, the causative introduces
a causer into the argument structure, then the passive demotes or eliminates the
causer. What remains is the element of causation in addition to the event as such
that is expressed by the original verb. Compare (6c) to the simple active sentence
in (6a):
(6)
c.
kodomo ga
biiru o
nom- asechild
NOM
beer ACC drink- CAUS
‘someone let/had the child(ren) drink beer’
rare-
ta.
PASS
PAST
(Japanese)
The anticausative is a diathesis that removes a causer AGENT from the set of core
arguments. We already encountered the example of intransitive open. Another
example would be break, widen or burn:
(7)
a.
She (AGENT) opened the door (THEME).
anticausative:
The door (THEME) opened.
b.
She (AGENT) burnt his letters (THEME).
anticausative:
His letters burnt well.
Unlike with the antipassive and the passive, the argument omitted is not understood
as still being involved. Along with the causer AGENT, the whole meaning component
of causation is eliminated; (7a) does not mean ›someone/they opened the door‹ –
although we might infer causation from our world knowledge. Likewise, (7b) does
not mean that someone burnt the letters. In German and other languages, including
Spanish, the anticausative is expressed by using a reflexive pronoun as object. The
Verbs
139
reflexive pronoun is a semantically empty grammatical element that just fills the
object position required by the syntax for the originally transitive verb.
(8)
Die Tür (THEME) öffnete sich.
lit. ‘the door opened itself ’
(German)
6.1.3 Levin’s classification of verbs
Levin (1993) used alternations of verbs in English for a systematic semantic
classification of more than 3,000 English verbs. According to her, verbs with the
same patterns of syntactic constructions and alternations among them have certain
meaning components in common, and these are responsible for their syntactic
behaviour. In the first half of her book, Levin describes eighty alternations in English,
where an alternation is defined by a pair of constructions applicable to the same
verb, but differing in the choice and linking of the arguments involved. For example:
active–passive, active–antipassive or transitive–anticausative would be alternations.
In the second half, she defines fifty-seven major verb classes (many with several subclasses) in terms of the alternations they do or do not exhibit. We will take a look at a
few sample classes, mainly concerning verbs we have already encountered.
Open. The verb open belongs to a sub-class of class 45, ‘Verbs of change of state’ (pp.
240 ff.). They share the ‘causative/inchoative alternation’ and the ‘instrument subject
alternation’. Due to the causative/inchoative alternation, there is a transitive verb
version with an AGENT subject and THEME object (she opened the door) along with the
corresponding anticausative ‘inchoative’ variant (the door opened). The instrument
subject alternation allows the use of an INSTRUMENT subject (this key opens the door).
There are further, minor, alternations possible for this class. Other verbs belonging to
the larger class are break, bend and bake.
Eat. Eat and drink form a sub-class of class 39, ‘Verbs of ingesting’ (pp. 213 ff.). Most
of them allow an antipassive use, which Levin captures with the ‘unspecified object
alternation’: the THEME object can be omitted (Kate ate) and the THEME argument is
understood as existing, but not specified. In addition, these verbs exhibit the ‘conative
alternation’:
(9)
a.
Sheila ate the sashimi.
b.
Sheila ate [at] the sashimi. (conative)
The conative variant means that the act of eating is related only to parts of THEME
– Sheila may be just nibbling at the sashimi, while the basic construction expresses
that THEME is affected as a whole. Note that this alternation is impossible with verbs
such as open.
Give. Give is representative of a larger sub-class within class 13, ‘Verbs of change of
possession’ (pp. 138 ff.). Most of them exhibit the ‘dative alternation’ between, e.g.
‘give RECIPIENT THEME’ and ‘give THEME to RECIPIENT’. Similar verbs are pass and offer.
140
Understanding semantics
The dative alternation is also possible for send, but this verb belongs to class 11,
‘Verbs of sending and carrying’ (pp. 132 ff.).
Shave. Shave belongs to class 41,‘Verbs of grooming and bodily care’. It has a transitive
AGENT–PATIENT use and an intransitive use (he shaved), where the PATIENT is understood
to be the AGENT him- or herself. Levin calls this the ‘understood reflexive object
alternation’. Like the antipassive, this alternation eliminates the direct object from
the construction, but its semantic effect is different (note that Kate ate does not mean
‘Kate ate herself’). In languages with a more extensive use of reflexive pronouns or
verb forms, the equivalent alternation would make use of a reflexive pronoun object:
(10) a.
b.
transitive: Das Kind (AGENT) wusch
the child
washed
reflexive:
Das Kind (AGENT) wusch
the child
washed
die Puppe (PATIENT). (German)
the doll.
sich.
REFL
This group includes verbs such as wash, dress or comb.
Marry. Marry is a ‘verb of social interaction’ (class 36, pp. 200 ff.). There are two
characteristic alternations; some verbs of social interaction exhibit the first, others
the second:
(11) simple reciprocal alternation
a.
Eunice and Eugene hugged/kissed/married.
b.
Eunice hugged/kissed/married Eugene.
understood reciprocal object alternation
c.
Eunice and Eugene quarrelled/fought/agreed.
d.
Eunice quarrelled/fought/agreed with Eugene.
In (11a and c), the subject is an NP conjunction that refers to two arguments;
alternatively the subject may make use of a collective noun (the team, the pair, the
family). The persons referred to inhabit the same role in the situation; there is no
asymmetry as between the usual AGENT and PATIENT, where the AGENT is in an active
role and the PATIENT in a passive one. The construction in (11b) is asymmetric, and
it does allow for an interpretation where the subject referent plays a more active role.
(11d) is syntactically asymmetric, but the understanding of the two roles is reciprocal.
6.2 SITUATION STRUCTURE
Levin’s classification does not relate to the temporal structure of the situations
expressed by the verbs. The temporal or situation structure of the situations
expressed by verbs is by far not as manifold as Levin’s classes. There are a number
Verbs
141
of terms for what is captured with this kind of classification of verbs: there is the
traditional term ‘aktionsart(en)’ (German, lit. ‘kind(s) of action’), others talk of
‘inherent aspect’ or ‘aspectual class’. We will use the term ‘aspectual class’. All these
terms are motivated by the fact that these classes are relevant for understanding the
interaction of verb meaning and aspect. Some distinguish only three or four, others
more, up to a magnitude of twenty (Russian ‘aktionsarten’). Thus, classifications of
verbs according to situation structure are much more general than Levin’s classes.
The main aspects of situation structure are whether a verb is stative or dynamic,
whether it conceives of the situation as simple or complex, and whether the verb
concept provides a criterion for the completion of the event denoted. I will introduce
five classes in this section, and take a look at their interaction with aspect in 6.3.
When talking of aspectual classes, it is most important to realize that the
classification applies to linguistic expressions and their meaning, NOT to types of
situations in the world, such as events, activities, processes, states, etc. Unfortunately
the literature in this field is often inconsistent in this regard. You will find formulations
like ‘work is an activity’. This is a confusion of semiotic levels; it should be: work is an
activity VERB and its meaning an activity CONCEPT.
6.2.1 Accomplishment terms
Accomplishment expressions conceive of a situation as a process or activity that
leads to a specific result. The process or activity is conceived of as uniform and
as involving some sort of constant change; it ‘culminates’ in the result specified.
Accomplishment verbs, i.e. verbs that by their very lexical meaning express an
accomplishment, are hard to find in English. But accomplishments can be easily
expressed by the combination of certain transitive verbs of action with an object NP:
eat an apple, write a paper, build a house, drive to the station.
Let us take a closer look at eat an apple. The expression depicts a situation denoted
as a particular activity of the AGENT (let us assume for the sake of illustration that the
AGENT is a person) and that they perform their activity on the THEME: the AGENT bites
off pieces of the THEME, chews them and swallows them, making the THEME smaller
and smaller. This ongoing activity can be referred to in the progressive mode: She
is eating an apple. Eventually, the AGENT may have reached the end of their activity
by having eaten as much of the THEME as there is to be eaten (people differ in their
eating habits as to which parts of the apple they actually eat – some peel it, some
leave the core, some eat everything except the stalk – but this does not matter). Thus,
apart from the THEME chosen for the example, the concept ›eat an apple‹ specifies two
ingredients of the situation expressed: (i) the kind of process or activity going on and
(ii) a criterion for its culmination and completion. These ingredients constitute an
accomplishment concept.
In Fig. 6.1, the horizontal arrow represents the time line; there is a certain period,
or stage, of a uniform activity or process in time represented by the grey rectangle
covering a section of the time line; the little triangles indicate the uniform and
culmination-oriented dynamicity of this phase; the bold vertical stroke signifies the
culmination point. The broken-line ellipsis encircles the whole situation. This is what
142
Understanding semantics
Figure 6.1
Structure of an accomplishment concept
an accomplishment term refers to: the dynamic activity or process together with its
culmination.
It is very important to realize that situations in the world are not accomplishments
by themselves, or of any other aspectual class. When you read this text, there is a
situation in which you are looking at your screen or tablet or holding a copy of this
book, looking at it, reading the words, trying to make sense of them. Describing
the situation, you might opt for putting it as an ongoing accomplishment, by saying
(12a), or you might choose to say (12b), putting it as just an ongoing activity without
formulating a description that provides a criterion of completion for the ongoing
situation.
(12) a.
b.
I’m reading the section on accomplishment terms.
I’m reading in the verbs chapter.
Thus the difference between an activity as part of an accomplishment and the mere
activity does not lie in the actual events; it lies in the way one chooses to describe
them, and in the conceptual distinctions which a speaker chooses to apply.
It must be cautioned that the term ‘accomplishment’ is not restricted to human or
animate actions, i.e. to situations with an AGENT argument. There are also concepts
for non-agentive processes with a specified result, but these verbs are rare; evaporate
would be one. As it happens, however, the term ‘accomplishment’ is so solidly
established in the literature that a more neutral term never came up.
6.2.2 Activity and process terms
Activity concepts lack a criterion of culmination. We will use the more general
term ‘process’ in order also to capture expressions that lack an AGENT argument. The
structure of a process concept is depicted in Fig. 6.2.
Figure 6.2
Structure of a process concept
Verbs
143
Typical process terms are intransitive verbs that denote some manner of acting by
the AGENT – eat, work, sleep, sing, read, walk, jog – or a process involving the THEME,
e.g. intransitive verbs such as glide, flow, buzz, vibrate. Due to the process component
of the concept, process terms, too, can be used in the progressive.
Accomplishment and activity predications with the same verb. A very
important point to observe with verbs of activity like eat, write, read, build is that
the specification of the object will determine whether they yield an accomplishment
or an activity predication. If the activity is such that you perform it only once on
a given object (like eating or writing, but not necessarily watching or singing),
specification of a quantity of the object argument (‘quantized object’) will lead to
an accomplishment predication. But bare plurals or bare mass nouns do not fix a
quantity and therefore provide no criterion for the activity being completed (recall
4.4.5). Thus, only sentences like those in (13a) can be taken as accomplishment
predications, while those in (13b) express activities.
(13) a.
b.
quantized THEME
non-quantized
she ate an apple/the apple/three apples/all the apples …
she ate apples/soup
In addition, it must be cautioned that some activity verbs with a quantized object
may be taken in a conative sense like in (9b). If so, they take on an activity meaning,
because the conative voice excludes the culmination of the activity. Thus the
conative voice can be understood as removing the culmination component from an
accomplishment concept.
For a large group of verbs, the accomplishment character with quantized objects
(in non-conative voice) depends to some degree on context. These are verbs for
actions which are connected to specific aims, where the aim may be accomplished
or not, or be accomplished more or less, when the action is performed. Consider the
transitive wash the clothes, or tidy the bedroom or the implicitly reflexive wash. These
activities are aimed at achieving certain results, like the THEME or AGENT’s body being
clean or the THEME being tidy. The intended results are part of the meaning, because
the kind of activity these verbs refer to is largely left open in the verb meaning except
for the criterion that they are apt to lead to the result to be achieved. (Just think of
the completely different activities involved with washing clothes by hand or by using
a washing machine – both are ‘washing clothes’). If someone says ‘I’ve washed/I’ve
washed the clothes/I’ve tidied my room’, these will by default be taken as reports of
accomplishments, as activities which reached their goal. But the same expressions
can also be used for referring to the mere activities which may or may not have
succeeded.
There is a standard diagnostic for distinguishing the accomplishment sense
from the mere activity sense. With accomplishment predications one can add an
in-adverbial for specifying the time it took to complete the situation:
(14) a.
b.
c.
He ate an apple in 30 seconds.
He washed in 3 minutes.
He tidied his desk in half an hour.
144
Understanding semantics
For mere activities, one can add a for-adverbial that specifies the length of time
during which the activity was carried out, leaving open whether the aim was
achieved or not.
(15) a.
b.
He washed for 3 minutes (but it didn’t make much of a difference).
He tidied his desk for half an hour (but it didn’t make much of a
difference).
For accomplishment predications, for-adverbials are impossible – or the predication
must be shifted to a different reading. Consider (16a) and (16b):
(16) a.
Sheila drank a cocktail for two hours.
b.
Sheila drove to the bar for 20 minutes.
(16a) coerces a conative activity reading. (16b) is odd if it is to be taken in the sense
that Sheila engaged for 20 minutes in the driving-to-the-bar activity. The sentence
does have an acceptable reading if the adverbial for 20 minutes is taken to specify
the duration of Sheila’s stay at the bar. But then, the for-adverbial relates not to the
situation referred to, but to the state resulting from it. Thus, we have to be careful
when applying the for-test: the adverbial must be taken to apply to the situation itself.
Conversely, an in-adverbial coerces an accomplishment reading on an activity verb:
(17) She jogged in ten minutes.
This sentence would be understood as expressing that the AGENT did her standard
quantum of jogging in ten minutes; ›do one’s standard quantum of jogging‹ is an
accomplishment concept.
Let me sum up what we have said about accomplishment and process terms. Both
refer to a situation as having a dynamic process component. The process component
is a uniform dynamic situation of some kind that continues in time. In addition,
accomplishment concepts contain a criterion of culmination. We mentioned three
linguistic characteristics which can be used as diagnostics. Both, accomplishment
and activity terms, can be used in the progressive, relating to the ongoing process.
Accomplishment terms can be combined with in-adverbials that specify the amount
of time consumed till the situation is completed; process terms cannot. Process
terms can be combined with for-adverbials that specify the overall duration of the
process; accomplishment terms cannot. Application of the ‘wrong’ adverbials either
does not relate to the situation itself or coerces a meaning shift of the VP.
6.2.3 Simple change and achievement terms
Many verbs express just a simple change: the THEME or AGENT is in a certain condition
before, and in a different condition after, the event. Let us refer to these conditions as
the initial condition and the resultant condition. (18) lists a few examples:
Verbs
145
Predication
Initial condition
Resultant condition
a.
She entered the room.
she is not in the room
she is in the room
b.
She turned the TV on.
the TV is not on
the TV is on
c.
She stopped.
she is moving
she is not moving
d.
She started reading.
she is not reading
she is reading
e.
She left.
she is here
she is not here
f.
She died.
she is alive
she is not alive
g.
The door opened.
the door is not open
the door is open
h.
The letter arrived.
the letter is on its way
to its destination
the letter is at its
destination
i.
She reached the top.
she is climbing to the top
she is at the top
(18)
Simple change predications always presuppose that the initial condition holds prior
to the change (recall 4.7.2.2); otherwise this kind of change would be logically
impossible. The initial condition may be a dynamic process, as in (18c, h, i), or a
state, i.e. a constant condition that does not involve change. The same applies to the
resultant condition: in (18d) it is dynamic, in the other cases a state. In the literature,
often all simple change terms are called achievement terms. I prefer to reserve the
term for predications with a dynamic initial condition, i.e. to terms that denote the
culmination of a process, such as arrive or reach. Usually, one of the two conditions is
defined as the negation of the other. In most cases, it is the resultant condition that is
defined positively; we will then have an ‘ingressive’ or ‘inchoative’ verb which is used
to express the beginning of a state or process. If the initial condition is the one which
is positively defined, as in stop, halt, end, leave, die we have an ‘egressive’ verb that
describes a situation as the ending of a state or process. Simple change verbs can be
schematically illustrated as in Fig. 6.3; the initial condition is marked with horizontal
stripes, the resultant condition with vertical stripes. Figure 6.4 displays the special
case of an achievement concept.
Figure 6.3
Structure of a simple change concept
Accomplishment terms and achievement terms differ in one point only: the
referent of accomplishment terms includes the culmination point plus the whole
146
Understanding semantics
Figure 6.4
Structure of an achievement concept
process leading to it; achievement terms only refer to the culmination point, while
the existence of the preceding process is presupposed. Consider the following pair
of sentences:
(19) a.
b.
Sheila drove to the supermarket.
Sheila arrived at the supermarket.
The first is an accomplishment sentence. It refers to a situation which comprises
both the drive and the arrival. If the sentence is negated (Sheila didn’t drive to the
supermarket), it is denied that Sheila undertook the whole action of driving to the
supermarket. Compared to this, the second sentence only refers to reaching the
endpoint of going to the supermarket. If the sentence is negated (Sheila didn’t arrive
at the supermarket) it is understood, that Sheila was on her way to the supermarket,
but for some reason did not complete going there.
The concept of a change term refers just to the change; the change is the transition
from the initial condition to the resultant condition. A simple change concept does
not contain information about the manner of transition, or its duration. In this sense,
the change is ‘simple’. Simple change verbs are often said to be punctual, but this is
misleading because it evokes the wrong impression that the situation in the world
that is referred to with a simple change verb must not have any temporal extension,
or at least be very short. These concepts are ‘punctual’ in the abstract sense that they
do not contain any information about the transition event itself; to these concepts the
changing event might as well be just a ‘point’ without temporal extension. To illustrate
the point, let me mention some Japanese examples: iku (‘go’), kuru (‘come’) or otiru
(‘fall, drop’) are simple change verbs. They cannot be used in the progressive in order
to express that the AGENT/THEME is on their way between A and B.
Simple change verbs are not used in the progressive, because there is no process
defined which the progressive might state as going on. They do not take foradverbials, or if they do, the for-adverbial relates to the resultant condition (she
opened the window for ten minutes). If the initial condition is not a process, they do
not take in-adverbials either, except with special contextual support. For example, she
opened the window in ten minutes would require a context in which her opening the
window was due for some reason. Simple change verbs are fine with at-adverbials
– provided the time specified by the adverbial is long enough to cover the actual
transition. For example, she changed to Beijing University denotes a change that
actually would take a couple of days, at least. Combining this statement with at 2:43
would make no sense, but at the beginning of the next term would be fine.
Verbs
147
6.2.4 Simple occurrence terms
Verbs like click, hit or knock are similar to simple change verbs in not claiming
temporal extension of the situation referred to. They differ in not expressing a
change between an initial and a resultant condition. The world is the same before
and after a click, a hit or a knock. These simple occurrences are conceived of as just
interruptions. Schematically, they can be represented as in Fig. 6.5. The term ‘simple
occurrence’ is introduced here. In the literature they are sometimes referred to as
‘semelfactives’; other classifications do not distinguish this class.
Figure 6.5
Structure of a simple occurrence concept
Characteristically, simple occurrences can be repeated because the conditions
which make a first occurrence possible will not change with it. If you have just
knocked, you can knock again (actually, one usually knocks several times when one
knocks). This is different with change terms: for example, if you have just opened
a window, you cannot open it again, you would have to close it first. By applying
repetition, simple occurrence predications can be turned into predications about
uniform, dynamic processes and then lend themselves to the progressive and to foradverbials (she knocked for 30 seconds, she is presently knocking at his window). Thus,
the progressive and for-adverbials coerce a repetitive reading of these verbs. In their
single occurrence reading, they are not combinable with the progressive, nor with
for- or in-adverbials. Of course, they are fine with appropriate at-adverbials.
Before we proceed to the aspectual class of states, which is very different from
the classes discussed so far, let me sum up the four dynamic classes and their
respective diagnostics. In Table 6.1, a class receives a negative entry for a given test,
if its application is necessarily connected with a meaning shift such as coercing
simple occurrence predications into process predications or applying the conative
alternation to accomplishment terms. Among simple change verbs, achievement
terms allow for in-adverbials while the others do not.
Table 6.1
Diagnostics for the dynamic aspectual classes
Aspectual class
progressive
in-adverbial
for-adverbial
at-adverbial
accomplishment term
+
+
–
–
process term
+
–
+
–
simple change term
–
–/+
–
+
simple occurrence term
–
–
–
+
148
Understanding semantics
While these are just diagnostics, the four classes can be distinguished semantically
by the following properties of the underlying concepts: (i) presence of a process
component as part of the situation referred to; (ii) specification of a resultant
condition; and (iii) conceptual ‘punctuality’. For accomplishment predications, the
resultant condition is the state reached with the culmination of the situation: the
apple is finished, the paper is written, the house is built, the drive has reached its
destination. The distribution of the three properties is displayed in Table 6.2. They
will play an important role for the interaction of verb meaning and aspect.
Table 6.2
Semantic properties of aspectual classes
Aspectual class
process component
resultant condition
‘punctuality’
accomplishment term
process term
+
+
–
+
–
–
simple change term
–
+
+
simple occurrence term
–
–
+
6.2.5 State terms
State terms denote a condition which is, at least temporarily, constant. States are
predicated of times, as we will see in the next section. States may result from a
change, but they do not themselves involve change. For example, the dynamic verb
memorize may lead to a state of knowing which once established is conceived of as
constant. State predications can be schematically illustrated as in Fig. 6.6. There
are not so many state verbs in English, mainly verbs such as like, love, hate, know,
understand, want or have, belong, contain, etc.; some verbs are state verbs only in
certain readings, e.g. cost, weigh, taste, mean in uses like the ones in (20). Other verbs
are ambiguous between a state reading and an ingressive change reading; these
include understand as well as sit and lie.
Figure 6.6
Structure of a state concept
(20) a.
A flight to Glasgow costs 19 euros.
b.
She weighs 59 kilos.
c.
This cake tastes awfully sweet.
d.
Japanese uni means ‘sea urchin’.
Verbs
149
State predications as such are not used in the progressive. Progressive with a state
expression such as she is being rude/silly coerces a meaning shift from state to activity,
in this case to ‘be acting in a rude/silly manner’. The states expressed by state verbs
may last for a long or short period of time, they may be temporary or permanent,
they may have a beginning and an end or not. For example, the state expressed by be
dead has a beginning, but not an end; it is what is called an irreversible state. With
conditions like these, it depends if a state expression can or cannot be combined with
a for-adverbial or with an at-adverbial. Thus the applicability of temporal adverbials
yields an inconclusive overall picture for state terms.
There is one point in which state predications differ clearly from all the dynamic
aspectual classes considered above: when used in an episodic sentence (i.e. applying
to a particular concrete situation, recall 4.6 on genericity) with the present tense, it
is only state predications that can be used for genuine present time reference, i.e.
reference to the time of utterance. With dynamic predications, present tense use
with the plain (non-progressive) form of the verb either is impossible in episodic
sentences or yields future time reference. Consider the following contrast:
(21) a
State
I know the answer.
b.
State
She hates me.
c.
Accomplishment
I write a paper. (??)
d.
Process
I walk. (??)
e.
Simple change
I switch on the TV. (??)
f.
Simple occurrence
I knock. (??)
Dynamic verbs in their plain (i.e. non-progressive) form can only be used with the
present tense in special cases.
(22) a.
I promise I will wait for you.
b.
Now Khedira passes the ball towards Özil, but …
c.
In 1815, Napoleon loses the battle of Waterloo and is banished to Saint
Helena.
d.
I commute to work by subway.
In (22a), promise is used as a ‘performative verb’, i.e. a verb which indicates the
speech act that is performed by using it: saying ‘I promise …’ constitutes a promise.
(22b) is only possible as a running commentary. (22c) is in the ‘historic present’,
referring in the present tense form to past events and situations and thereby
simulating a witness perspective. (22d) is a habitual statement, i.e. a generic sentence
about a situation that recurs regularly; a single such situation need not be in progress
at the time of utterance. (22a) and (22b) constitute, or at least claim, coincidence
of the event expressed with the utterance itself; historic present refers to the past,
150
Understanding semantics
habitual predications do not refer to single events. We will return to these special
cases in 6.4.2. Apart from these exceptions, for dynamic verbs, the present tense with
reference to the time of utterance is only possible in the progressive – and this is
restricted to accomplishment and process predications:
(23) a.
I am writing a paper.
b.
I am walking.
c.
I am switching on the TV. (??)
d.
I am knocking. (repetitive reading only)
The classification presented here ultimately goes back to the very influential work
on English aspectual classes in Vendler (1957/67); many call aspectual classes
simply ‘Vendler classes’. He introduced the four classes of accomplishment, activity,
achievement and state terms and used the diagnostics of progressive and for-, in- and
at-adverbials. His proposal was later generalized and elaborated, to be extended to
more general and additional types of verbs, as well as to different languages. The
classification presented here can be considered essentially common sense in the field.
6.3 ASPECT
As to the description of aspect and tense, there are a lot of half-truths to be met in
the literature, in particular in descriptive grammars. In the following two sections, I
2
will try my best to give a clear and consistent picture of the matter.
6.3.1 The central distinction: imperfective vs perfective aspect
Talking of ‘aspect’, the fundamental distinction is between imperfective and
perfective aspect. It can be illustrated with a minimal pair such as this:
(24) a.
b.
imperfective:
I was watching TV [when the door bell rang].
perfective:
I watched TV for a while [then went to bed].
The two sentences might be used with reference to the very same instance of the
speaker’s watching the TV. The distinction between perfective and imperfective
aspect is not to be found in the way things are in the world. Rather, it is a matter of
the way in which the speaker chooses to talk about the world.
The imperfective variant is a predication about a given time; in the case of (24a),
the time is determined by the punctual event of the door bell ringing (a simple
occurrence expression) and, less specifically, by the past tense. What the main clause
of the imperfective sentence says is that, at the time considered, the speaker was in
2 The approach taken here owes much to the theory of aspect developed in Galton (1984).
Verbs
151
the state of watching TV. An imperfective predication is about a contextually given
time; it tells what kind of situation is given at that time. It does not tell anything
about the situation before or after the time referred to. Therefore people often say that
imperfective aspect relates to uncompleted situations (in fact the term imperfective
is motivated by this view); but this is not really adequate – since imperfective aspect
leaves open what happens after the time referred to, the situation may be later
completed or not. Very often you will find the definition that with imperfective aspect
the situation is ‘seen from within’. This is not a very clear notion. The crucial point
is that predications in the imperfective aspect are predications about a given time
which describe the situation at that time. If the situation extends beyond the time
predicated about, the time can be considered embedded in the total situation, and
therefore the total situation is in some sense ‘seen from within’.
The time referred to may be a point in time or an extended period. In any event,
the state expressed by the imperfective predication must not change during this time.
This is a consequence of the presupposition of indivisibility (4.7.2.2) which applies
to every predication whatsoever, including the predication about time expressed by
an imperfective predication. In this sense, the time referred to is LIKE a point, i.e. not
divided. It is the same sense of punctuality which we encountered with ‘punctual’
verbs. The time referred to is located by the tense of the verb: in (24a) it is a past
time, i.e. located before the time of utterance. Thus a tensed imperfective sentence
constitutes a twofold predication about a given time: a temporal localization by
means of grammatical tense and a specification of the conditions at that time by the
rest of the predication.
The perfective sentence in (24b) does not predicate about a time, it predicates
about an event – the referential argument of the verb. In this case, the predication
describes what the speaker did: watching TV (and then going to bed). The event is,
in addition, located in time, expressed by the past tense: it occurred at some time
before utterance time. Thus, tensed perfective sentences, too, provide a twofold
predication, but this time the two predications are about an event – by tense about
its temporal location, and by the rest of predication, about the referential argument
of the verb. Due to the presupposition of indivisibility, the two predications must
apply to the event argument as a whole: the event, or situation, referred to must be
of the kind the sentence describes and it must be wholly located in the interval of
time referred to. The perfective aspect is therefore often described as ‘referring to an
event as a whole’. You may also find the more traditional description that perfective
aspect refers to a completED event. This is wrong, for the simple reason that it is
possible to use perfective aspect referring to future events which, of course, would
not be completED yet. Perfective aspect deals with complete events, not necessarily
with completED events.
Figure 6.7 illustrates the difference between imperfective (left) and perfective
(right) aspect. It is a matter of perspective. Imperfective aspect views the situation
from a given time, stating that the world is, at this time, in a certain state. Being
about one time, the predication can only be a state predication because any kind of
change can only be stated with respect to at least two times. Perfective aspect views
the situation from an event in the world, stating that and when it happens. Thus,
152
Understanding semantics
Figure 6.7
Imperfective aspect and perfective aspect
Imperfective aspect
Perfective aspect
event
world
state
time
world
interval
time
time
imperfective aspect ‘looks’ from a time into the world, and perfective aspect ‘looks’
from an event in the world onto the time line. These are two basic ways of correlating
situations in the world with the time line.
The difference, and the interaction, of perfective and imperfective aspect can be
illustrated with narrative texts, such as stories, tales, reports or novels. The following
is the beginning of the translation of Grimm’s fairytale of the frog prince. The tensed
predications are marked for imperfective (ipf) and perfective (pf) aspect. The
perfective predications are put in italics.
(25) One fine evening a young princess put on her bonnet and clogs (pf), and went out
to take a walk by herself in the wood (pf); and when she came to a cool spring of
water (pf), that rose (ipf) in the midst of it, she sat herself down to rest a while
(pf). Now she had a golden ball in her hand (ipf), which was her favourite
plaything (ipf); and she was always tossing it up into the air (ipf), and catching
it (ipf) as it fell (pf). After a time she threw it up so high (pf) that she missed
catching it again (pf) as it fell (pf); and the ball bounded away (pf), and rolled
3
along upon the ground (pf), till at last it fell down into the spring (pf).
The narration starts with a time adverbial One fine evening that defines the temporal
frame for the series of events described in the following. All the events mentioned in
the passage are located within this time interval. It corresponds to the marked time
zone in the diagram for perfective aspect. The first two events are told chronologically
(putting on bonnet and clogs and then going out). The first event is located
indeterminately within that evening, setting a temporal starting point for the story.
The second event is taken as following next; narration time advances a bit. Some time
will have elapsed until the next event mentioned takes place, her coming to the spring
in the wood. The narration here inserts an imperfective predication about the location
of the spring. This is background information predicated about the time reached in
3 Quoted from Grimm’s Fairy Tales, Amazon Kindle edition.
Verbs
153
the narration. The spring, of course, has existed much longer, but this is irrelevant at
this point; what matters is the state of the world in the situation described. Now she
sits down – the next event in the series. By the following two imperfective statements
we learn that at this moment she has a golden ball in her hand and that it is her
favourite plaything. There follow two predications in the progressive; progressives are
imperfective. They express continued uniform action as a temporary state. We will
relate the total predication to some stretch of time, because the repetition of tossing
and catching the ball necessarily consumes time; we will locate this interval after she
has had her rest. The text then describes a series of events in perfective aspect, one
following the other, carrying narration time further to the situation where the frog
will now appear. What we can see looking at this example is the dynamic function of
perfective aspect – it takes the story forwards – and the static function of imperfective
aspect which gives information about a given situation at a given time.
6.3.2 Perfective aspect
Among the aspectual verb classes introduced above, state terms, when used with
the plain form of the verb, yield imperfective predications; dynamic aspectual
types yield perfective aspect with the verb in its plain form and in episodic use.
If the perfective aspect is applied to state predications, they are turned into event
4
predications. One way to turn a state concept into a simple occurrence concept is
by adding an adverbial that provides the state with a temporal delimitation. The
resulting event consists of a state that pertains for the time specified:
(26) I was in the garden for a while and then washed the dishes.
Another common strategy is to use a state term to refer (metonymically) to the
beginning of the state; this turns a state predication into an ingressive simple change
predication:
(27) a.
b.
I’ll be with you in a minute.
She had a ticket within half an hour.
6.3.3 Imperfective aspect
While state predications immediately yield imperfective aspect, event predications
must be turned into states for enabling imperfective aspect. The most prominent
means is applying the progressive. The progressive is a variant of the imperfective
aspect; it expresses the state of a dynamic condition that continues uniformly, such
as some activity being done or some process going on. In English, the progressive
is possible with accomplishment and process predications, as well as with simple
occurrence predications (knock, beat) in a repetitive interpretation. As mentioned
4 In the discussion of aspect, I will use the term event predication for dynamic predications in
general.
154
Understanding semantics
above, certain states (be silly, be rude, be polite) can be used in the progressive, taking
on the meaning of the activity of acting in a certain way.
Another way of using event predications in imperfective aspect is the habitual
mode of speaking. Habitual predications express actions and events that are repeated
regularly; the repetitions form what is called a ‘serial state’:
(28) a.
b.
This shop opens at 9:30 a.m.
She listens to her mp3 player on the subway.
While these are examples where the plain verb is used, there are also special forms
in English for the expression of habitual predications, but these are restricted to the
past tense:
(29) a.
b.
This shop used to open at 9:30 a.m.
She would listen to her mp3 player on the subway.
6.3.4 Perfect aspect
In addition to perfective and imperfective aspect, we will discuss two more aspects:
perfect and progressive. The perfect aspect, too, yields a state predication about a
given time. The state is a state that results from a previous event. Consider a simple
example:
(30) I have written my term paper.
The sentence predicates about the present time that it is located in the state resulting
from a past event of the type ‘speaker write a term paper’. While reference is
primarily to the time predicated about, the event from which the current state results
is also referred to as having occurred before the time referred to. The perfect aspect
can be schematically depicted as in Fig. 6.8.
Figure 6.8
Perfect aspect
event
world
time
time
Verbs
155
In English, the perfect aspect is expressed by using the auxiliary have with the
past participle. It lends naturally to all classes of predications that provide a resultant
condition: accomplishment and change predications (recall Table 6.2). The perfect
is possible for non-resultative verbs, too. In this case, context must provide a sense
in which the situation at the given time results from the former event. For example,
consult in (31) is a simple occurrence verb that does not come with a specified
resultant condition. The sentence would be read in the sense that the speaker’s
consulting the dentist has led to some result, like now knowing better what to do
with their aching molar.
(31) I’ve consulted my dentist.
One variant of perfect, the experiential perfect takes any occurrence of an event as
resulting in a change. With this variant one would express that one has had a certain
kind of experience:
(32) Have you ever had renal colic?
English has a special use of the perfect, called the perfect of persistent situation.
It is used to express that a certain state has prevailed up to the time referred to. In
German, for example, this would be expressed in plain present tense.
(33) a.
b.
We’ve lived here for six years now.
Wir wohnen [present tense] hier jetzt seit sechs Jahren.
lit.: ‘We live here now since six years.’
(German)
It is very important to realize that the relation between the former event and the
subsequent state is not just temporal precedence, but a causal relation: the former
event is the cause of the subsequent state. It is this causal relation which connects the
time referred to to the former event. The causal connection is indicated by the curved
arrow from the event to the state in Fig. 6.8. Often this connection is described in
terms of ‘relevance’: the former event is relevant to the subsequent situation.
6.3.5 Prospective aspect
Prospective aspect is the temporal mirror of perfect aspect. It refers to a state that
leads up to a future event:
(34) a.
I’m leaving!
b.
It’s going to rain.
c.
The train is about to leave.
Be going to and be about to are usually cited as ways of expressing prospective
aspect: the event expressed is scheduled or on its way at the time referred to. The
156
Understanding semantics
be V-ing form normally used for the progressive, however, can also be considered
an expression of prospective aspect if it is combined with a punctual verb. (With
non-punctual verbs, the form inevitably yields the progressive reading.) This is the
case in (34a), taken as an announcement of an imminent action; and this is the way
in which we are arriving or he is dying would be interpreted. These constructions do
not express that a situation as described by the verb is going on; rather they describe
a situation that is causally linked to a future event of the kind – the situation is such
that this kind of event is likely to arise from it. Figure 6.9 illustrates the prospective
aspect.
Figure 6.9
Prospective aspect
event
world
time
time
While the prospective be V-ing form is restricted to punctual verbs, to be going to
V is possible with all aspectual classes, including state terms. The construction to be
about to V is restricted to prospective states of imminent events and hence restricted
to event predications.
Concluding the section on aspect, we find that the perfective aspect constitutes a
predication about an event, while the other three aspects – imperfective, perfect
and prospective – are stative. They all express a state predication about a time: the
given time is related to either the active phase of the event itself or to a state the
event led to or to a state that is likely to lead to such an event. I will call the event
referred to in the perfective aspect and the time referred to in the stative aspects as
the tense argument of the predication. In the aspect diagrams it is represented by
the symbol 䡵.
As we will see in 6.5, and have partly already demonstrated in this section,
aspect is not always marked explicitly. For example, I went to work on the subway
can be interpreted as an episodic perfective predication about a particular event
or as a habitual imperfective statement. Many languages do not mark perfective
vs imperfective aspect; the Romance languages mark it only in the past tense and
standard German does not mark it at all. Nevertheless, if we interpret a given
sentence with its central predication, we have to make up our mind whether we are to
take it as stating that an event of a certain kind occurred, and if so when (perfective),
or whether we are to take it as a description of a certain state that existed at a given
Verbs
157
time (imperfective). Since these are predications about quite different things, there is
no way of leaving this issue open. Therefore, whether marked for aspect or not, every
sentence will receive some aspectual interpretation or other.
We can sum up the considerations on aspect in the following definition:
DEFINITION 1 Aspect
The aspect of a verbal predication concerns the way in which the situation
expressed and the time referred to are related to each other. Aspect determines
whether the predication is about an event in time (perfective aspect) or about
a time related to a situation (imperfective, perfect and prospective aspect).
6.4 TENSE
Aspect concerns the way in which a given time is aligned with a state (imperfective,
perfect, progressive) or how an event is aligned with time (perfective), but aspect
does not locate the situation in time. This is the function of tense.
DEFINITION 2 Tense
Tense locates the situation expressed in time.
Thus tense and aspect are independent, but – as we will see immediately – they
interact. We will confine the discussion in this book to ‘absolute’ tenses, i.e. tenses
which relate to the time of utterance: past, present and future tense. So-called relative
tenses primarily relate to a ‘reference’ time, which itself may be related to the time
of utterance. The combination of relative tense with absolute tense yields temporal
relations like past-in-the-past (pluperfect), past-in-the-future (‘future past’) or
future-in-the-past.
6.4.1 Interaction of tense and aspect
As a start, let us have a look at the three standard absolute tenses – present, past and
future tense. In principle, they can be combined with all aspects:
(35)
perfective
imperfective
perfect
prospective
Past tense
I promised
I was reading
I had finished
I was leaving
Present tense
I promise
I’m reading
I have finished
I am leaving
Future tense
I will promise
I will be reading
I will have finished
I will be leaving
While this all looks very smooth, there is a problem with the second case in (35),
the combination of perfective aspect with present tense, the ‘present perfective’ for
158
Understanding semantics
short. As we saw in connection with the examples in (21) and (22) above, dynamic
predications can be used with present time reference only in some very special cases.
Recalling the remarks in 4.3 on temporal deixis, the reference of present, past and
future tense is determined by the temporal relation to the time of utterance, where
the time of utterance is the time which the event of uttering the sentence takes.
Figure 6.10 illustrates the definition of ‘past’, ‘present’ and ‘future’ TIME, to which past,
present and future TENSE refer, respectively. The utterance U is an event in the world.
It defines a time on the time axis, the ‘time of utterance’, abbreviated TU. This is the
present time; past is prior to TU, and future subsequent to it.
Figure 6.10
Past, present and future time
U
world
time
past
future
TU
present
Now, what is it that tense relates to TU? Grammatical tense is applied to a
predication, the main predication of the sentence. The predication is interpreted
as carrying a certain aspect, and it is the tense argument of the predication that is
related by tense to TU. Let us consider the past tense examples in (35). I promised is
a past perfective; it predicates about an event as a whole; the past tense of the verb
locates this event in the past. Consequently, the event is completed by the time of
utterance. This is small wonder with a punctual verb like ‘promise’, but it also holds
for accomplishment and process verbs denoting events of arbitrary duration: she
built a house means that the house is finished.
The past imperfective I was reading predicates about a contextually given time;
this time lies completely within an ongoing process of the speaker reading; past
tense places this TIME in the past – not the state in its total temporal extension. The
situation of reading may be continued after the time referred to; it may even extend
into the future. Thus, with imperfective predications, it is just one time in the past
that is located within the situation.
The past perfect I’ve finished is similar to the past imperfective; it differs in that the
state predicated of the time is defined as resulting from a previous event of finishing.
Again, this state can extend into the future.
The past prospective I was leaving or I was about to leave predicates of the past
time referred to that a later event of leaving was being envisaged or prepared at that
time. The event may later actually take place or it may fail to come about; it may take
Verbs
159
place in the past or in the future (cf. for instance I was about to leave the day after
tomorrow). The three cases of past perfective, past imperfective and past perfect are
illustrated in Fig. 6.11.
Figure 6.11
Past perfective, imperfective and perfect (right)
event
U
U
state
U
TU
TU
TU
time
past perfective
I jogged
event
past imperfective
I was jogging
time
past perfect
I had jogged
The diagrams are obtained by adding the U–TU component of Fig. 6.10 to the basic
pictures of perfective, imperfective and perfect aspect. For past tense, U–TU is added
to the right of the tense argument 䡵. Placing U–TU to the left of 䡵, yields the pictures
for future tense. The situation with the present tense is more complex; we will turn
to it in a minute.
6.4.2 Past, present, non-past and future tense
Past tense. There is little to say about the past tense beyond what was stated in
the previous section. The past tense places the tense argument before utterance
time. In actual discourse, the tense argument will be temporally located with much
more accuracy – recall the interpretation of the narrative passage in (25). This is
a consequence of interpretation in context: we will inevitably try to connect the
event or time referred to to other events and times in order to come up with a
coherent interpretation of the given sentence in its context. Except by tense, the
tense argument may be more precisely located by temporal adverbials such as the
at-adverbials discussed above or expressions like in a minute, soon, yesterday, this
morning, etc. This is, of course, also possible for present and future time reference.
In languages without grammatical tense, this is the only means of explicit temporal
location of the tense argument.
Present tense. Applying the present tense to the stative aspects can be done simply
by identifying the tense argument with TU. The result is a predication saying that the
state described applies at TU.
In order to align a perfective statement with U–TU one would temporally have to
match two events: the tense argument event and the utterance event. For ‘normal’ events
160
Understanding semantics
this is impossible. Consider ‘AGENT-eat-an-apple’ as a tenseless perfective predication.
Application of the present tense would mean to identify the event of the AGENT eating
the apple – which of course would take some time – with the event of uttering, for
example, ‘I eat the apple’. This is impossible, because the two events do not coincide
temporally. There are basically two scenarios in which coincidence can be considered
as given; both were mentioned in connection with the examples in (22). The first case is
performative verbs. They refer to an event which comes about by making an utterance
which describes this event. Thus, in the case of performative verbs used in this way (in a
‘performative utterance’), the event of utterance indeed coincides with the event referred
to of performing the speech act expressed. The second case is running commentaries,
like a live soccer game commentary on radio or TV. In this style of report, simultaneity
of commentary and events described is claimed and tried to be achieved – and accepted
by the hearer as a manner of speaking, although, strictly speaking, the events reported
usually minimally precede the corresponding verbal reports.
The other two cases are different. Historic present is an established non-literal
as-though use of present tense. The effect of quasi-witnessing the situations
described comes about because the present tense in its literal use relates to really
present situations. It would be wrong to conclude from the existence of historic
present tense that relation to a past time is just another meaning of the present tense,
or that the present tense relates indiscriminately to past and present time. If this
were the case, the historic present would not have its witnessing effect. It is crucial to
observe that this is a SHIFTED meaning of the present tense.
Habitual predications with event verbs in their plain form (I go to the university by
subway) involve an aspectual shift from singular events to a serial state. As we saw
above, they are a variant of imperfective predications, and hence not relevant here.
There are state predications such as ducks are birds and two plus two equals four
which do not refer to temporary conditions. In many descriptions of the present
tense you will find examples such as these cited as evidence for a ‘timeless’ or
‘eternal’ variant of the present tense. Ascribing cases like these to the semantics of
(an alleged) variant of present tense is inadequate. In such cases, too, the meaning
of the present tense is the same as with all stative predications. Present tense places
the time of utterance within the state described. Strictly speaking, ducks are birds is
a predication only about the time of utterance. It lies, however, in the nature of the
predications expressed by such sentences that they apply forever if they ever apply.
Therefore, it FOLLOWS from the fact that if such a predication is true of any arbitrary
time when it is uttered then it is always true. There is no need to claim a special
meaning variant of the present tense for these cases.
Non-past tense. Many languages have a non-past tense (usually called ‘present
tense’). Non-past includes relation to the present and to the future. Therefore, with
the non-past tense no problem arises with perfective aspect. A non-past perfective
will be taken as referring to a future event, while a non-past imperfective will be
taken to relate to the time of utterance, unless the context indicates reference to
a future time. In German, the so-called present tense is in fact a non-past tense.
Consider the following sentences:
Verbs
(36) a. Es regnet.
lit. ‘It rains’ = ‘It is raining.’
161
(German)
b. Morgen regnet es.
(German)
lit. ‘tomorrow rains it’ = ‘It will be raining tomorrow’ or ‘it will rain tomorrow’
c. Ich fahre nach Brüssel.
lit. ‘I go to Brussels’ = ‘I’m going to Brussels.’
(German)
They are all in the plain ‘present’ tense form. (36a) is taken in an imperfective
progressive reading (note that there is no progressive form in standard German). In
(36b), the verb has the same ‘present’ tense form; the sentence can be used either as
a future imperfective or a future perfective; future time reference is only indicated
by the adverb meaning ›tomorrow‹. (36b) is taken as referring to a future event; it is
a non-past perfective. English is held by some scholars to have no future tense, but
only a non-past tense. While this view is controversial (and will not be adopted here),
English uses ‘present’ tense forms for future time reference in certain subordinate
clause, for example in before-clauses:
(37) Before I leave, let’s fix a date for our next meeting.
Future tense. Future tense relates the tense argument to the future. For perfective
aspect, the event referred to is completely located in the future; for the stative
aspects, the time referred to lies in the future, but this does not mean that the
state that applies at this time is confined to the future. There is no contradiction
in saying: ‘I’m sick today and will be sick tomorrow’, meaning that the state of
being sick will not be interrupted between today and tomorrow. It is often argued
for English, German and other languages that what is traditionally called ‘future
tense’, like they will work, sie werden arbeiten (German), is not a real tense, but the
expression of some kind of modality, like probability, conjecture, etc. One argument
against considering these forms as a real future tense is the fact that they also have
non-future readings. Both, they will work and sie werden arbeiten can also be used
with present time reference for uttering a conjecture, as in (38a) and its German
equivalent (38b).
(38) a.
b.
What do you think they are doing? – I think they will be working.
Was, glaubst du, machen sie gerade? – Ich denke, sie werden arbeiten.
This, however, is just a matter of polysemy. The forms have two different functions;
one is the modality of conjecture with present or non-past tense, and the other is
future tense. Another argument against the future tense analysis of such forms is that
future time reference in both languages can also be expressed with non-past ‘present’
tense; in fact, in German using the non-past forms is the normal way of referring to
future time. What this shows is that the so-called present tense is in fact a non-past
tense. Existence of a non-past tense does not preclude existence of an additional
162
Understanding semantics
future tense proper. Tense systems are not necessarily such that the different tenses
available mutually exclude each other. To sum up, English and German can well be
considered to have a future tense.
6.4.3 The verbal onion
Similar to the noun phrase, the verb forms an ‘onion’ with several structural layers.
The innermost is the verb stem; it is followed by one or more diatheses. These two
layers determine the argument structure. They are followed by the layer of aspect; it
determines the tense argument. It is followed by tense, which predicates about the
tense argument. The general structure is given in Fig. 6.12. This hierarchy is directly
reflected in the order of affixes in Japanese verbs as shown in (39a). Two more layers
can be added – which we do not deal with: mood/modality (39b) and sentence
type (39c):
(39) a. mat- aserarete i
stem diath 1 diath 2 aspect
wait- CAUS
PASS
CONTINUATIVE
‘[someone] is having [her] wait’
ru
tense
(Japanese)
PRESENT
b. mat- aserarete i
ru
darō
stem diath 1 diath 2 aspect
tense
modality
wait- CAUS
PASS
CONTINUATIVE
PRESENT TENTATIVE
‘I think [someone] is having [her] wait’
c. mat- aserarete i
ru
darō
ka
stem diath 1 diath 2 aspect
tense
modality sentence type
wait- CAUS
PASS
CONTINUATIVE
PRESENT TENTATIVE QUESTION
‘Will [someone] be having [her] wait ?’
Figure 6.12
The verbal onion
(((
( verb stem) diatheses )
passive
causative
etc.
aspect
perfective
imperfective
perfect
progressive
)
tense
)
past
present
future
The tense argument is the central object of anchoring when it comes to establishing
reference for a sentence. If the situational element is an event, the event is anchored
along with its participant arguments – not only in time, but also within the facts
applying at that time. Similarly, if the tense argument is a time, the predication with
its arguments is located at this time and within the given circumstances.
Verbs
163
6.5 SELECTED TENSE AND ASPECT SYSTEMS
We will conclude this chapter by having a brief look at four tense and aspect systems;
the description will be confined to the major tenses and aspects introduced above.
English. The English system has two aspectual distinctions. The first distinction is
expressed by the plain form vs the form be + gerund. The latter expresses progressive
in most cases (and in some cases prospective). For the majority of dynamic verbs,
this form can be considered as marking imperfective aspect. The imperfective
aspect is not formally marked with state verbs. The perfective aspect goes unmarked
throughout. The second aspectual distinction is between perfect and non-perfect.
The perfect aspect is expressed by have + past participle; non-perfect is not marked.
These two distinctions can each be combined with three tenses, past, non-past and
future tense. Future tense is (almost) obligatory in main clauses with future time
reference.
German. The German verbal system is formally similar to the English one, except
5
that it lacks a distinction corresponding to the English progressive form. German
has one aspectual distinction, perfect vs non-perfect. Perfect is expressed by the
auxiliary haben ‘have’ or sein ‘be’ + past participle. The non-perfect (plain) forms
can be used indiscriminately for imperfective as well as perfective predications.
As in English, there are three tenses, past, non-past, and future tense, which can
be combined with the non-perfect and the perfect aspect. Complicating the matter,
the present perfect forms have a second function of simple past; thus the so-called
present perfect (‘Perfekt’) in German is either a non-past perfect or a past nonperfect. In colloquial German, the present perfect forms are replacing the genuine
past forms more and more as the standard means of relating to the past. The nonpast forms are the unmarked way of relating to the future, but genuine future tense
is also possible.
Russian. Russian distinguishes between imperfective and perfective verbs. Almost
all verb stems are imperfective; by using a rich set of prefixes, they are turned into
perfective verbs. Prefixation also affects meaning; for some verbs a certain prefix
yields a neutral aspectual counterpart with the same meaning except for inherent
aspect, but for most verbs there is no neutral counterpart. For example, imperfective
čitat’ (read) has the perfective derivations pročitat’ (read, recite), počitat’ (read a
while), načitat’ (read a lot) and dočitat’ (finish reading). Thus, the distinction between
the imperfective and perfective aspect is inherent to the verb meaning and lexicalized
rather than a matter of the grammatical form. In addition, there is ‘secondary
imperfectivization’: adding a suffix to a perfective verb yields an imperfective
predication. This instrument is a means of grammatical aspect, like the progressive
and perfect forms in English. Russian does not have a distinct perfect form. There is a
5 This applies to standard written German. Colloquial German and most, if not all, German dialects
have an equivalent of the English progressive form: am + infinitive + sein (inflected): cf. ich bin am
lesen ‘I am reading.’
164
Understanding semantics
past tense that can be applied to all verbs. The formal present tense, however, functions
as a present tense with imperfective verbs and as a future tense with perfective verbs.
There is no present tense with perfective verbs; all the exceptional cases of present
perfective mentioned above – performative verbs, running commentaries, historic
present and habituals – are expressed with imperfective verbs. Thus, the formal
present tense is a real future tense with perfective verbs, not just a non-past tense.
Present tense with imperfective verbs is restricted to present time reference. For
future time reference with imperfective verbs, there is a special future tense which is
morphologically different from the future tense of perfective verbs.
(40) (Russian)
Past tense (fem. sing.)
Imperfective verb
6
Present tense (1st sing.)
Future tense (1st sing.)
Perfective verb
ita-la
pro ita-la
ita-ju
---
budu itat’
pro ita-ju
Japanese. Japanese has two tenses: non-past tense marked with the ending -(r)u and
past tense marked with the ending -ta. It has a major aspectual distinction between
perfective (plain form of the verb) and ‘continuative’, marked with a suffix -te i(gerund + be) before the tense ending. For taberu ‘eat’ the four forms are:
(41)
non-past tense
past tense
Plain form
tabe-ru
tabe-ta
Continuative
tabe-te i-ru
tabe-te i-ta
Similar to English, the plain forms yield the imperfective aspect with state verbs.
State verbs do not take the continuative form; there are only very few, including i‘to be’ which appears in the continuative form. With dynamic verbs, the plain form
expresses perfective aspect. The continuative form yields a progressive reading with
accomplishment and process terms and with simple occurrence verbs in repetitive
use; with accomplishment and simple change terms, the form yields a resultative
perfect. Thus, for accomplishment verbs such as kiru ‘put on’, the continuative form
has two readings – kimono o [ACC] ki-te i-ru can mean ‘[she] is putting on a kimono’
and ‘she is wearing a kimono.’ In addition, the continuative form can express general
perfect for all dynamic verbs.
6.6 CONCLUDING REMARK
Taking a closer look at the general semantics of verbs, we are now able to recognize
a very important trait of human language – the way in which a situation in the
6 Historically deriving from adjectival past participles, Russian past forms are inflected for number
and gender, but not for person.
Verbs
165
world can be put into words to describe it is by no means determined by the
way the situation ‘is’. First, of course, it is up to the speaker which aspects of the
situation they want to talk about. But even if the speaker chooses to talk about a
very particular situation, say a ‘read’ situation by a particular person of a particular
book, there are many different ways to do so. The speaker can choose to mention
the AGENT or not and to mention the THEME or not; if the speaker mentions both,
they can manipulate the structure of the sentence, choosing the active voice for
highlighting the AGENT or the passive voice for promoting the THEME. The speaker
can choose between various aspectual variants of ‘reading’, describing the situation
as an activity or an accomplishment. The reader can choose to use the sentence as
a description of the situation at the given time (imperfective), or as an event that
happened (perfective), or as a situation resulting from a reading event (perfect), or
leading to it (prospective). Finally, the speaker can place the situation in time relative
to when they utter the sentence. All these choices can be applicable to the very same
‘objective’, ‘real’ situation. Thus, the semantic distinctions we encountered in this
chapter are not distinctions that apply to the situations in the world; rather they
are conceptual distinctions among the ways in which we may cast reality when we
describe it by means of words and grammar.
EXERCISES
1. Describe in your own words the effect on the argument structure of the diatheses
a. passive b. antipassive
c. causative
d. anticausative
2. For each of the following alternations, find five more verbs that exhibit it. Do not
consult Levin (1993).
a. the causative/inchoative alternation
b. the dative alternation
c. the simple reciprocal alternation
3. Define in your own words the five aspectual classes introduced in 6.2:
accomplishment, process, simple change, simple occurrence and state terms.
4. Use the diagnostics mentioned in 6.2 (progressive, for-, in-, at-adverbials, present
time reference) to determine the aspectual classes of the following predications;
consider the whole phrases. Set up a table.
a. say
b. dress
c. exceed
d. hammer
e. melt (intransitive)
f. snow
g. write e-mails
h. grow
166
Understanding semantics
5. Describe in your own words the distinction between the imperfective and
perfective aspects.
6. What is common to the
a. perfective and perfect aspects
b. imperfective and perfect aspects
7. What is the aspect of the following sentences:
a. She baked cookies.
b. They tasted awfully sweet.
c. I tasted them.
d. The dog was sick on the floor.
e. I’ll never have any of those again.
8. Put the predication ‘Sheila-mow-her-lawn’ into the following tense-aspect
combinations:
a. present progressive
b. past habitual
c. past perfect
d. future perfective
e. present prospective
9. Discuss the problem of the present perfective.
10. There are four different uses of would + verb in English. Try to describe them in
terms of aspect, tense and mood/modality.
FURTHER READING
Argument structure. On passive in English: Huddleston & Pullum (2002: ch. 16.10
by Ward, Birner & Huddleston); on passive and other diatheses in English Givón
(1993: ch. 8). Levin (1993) on her classification. On causative and passive in Japanese,
Shibatani (1990: ch. 11.4.1–11.4.2).
Aspect and tense. Comrie (1976) and Galton (1984) on aspect and the interaction of
aspect and tense, Comrie (1985) on aspect. On English tense and aspect, Huddleston
& Pullum (2002: ch. 3.3–3.5 by Huddleston), Givón (1993: ch. 4.1–4.3). In WALS
(Dryer and Haspelmath, eds, 2011, http://wals.info), you will find four chapters by
Östen Dahl and Viveka Velupillai on ‘Perfective/imperfective aspect’ (ch. 65), ‘The
past tense’ (ch. 66), ‘The future tense’ (ch. 67) and ‘The perfect’ (ch. 68).
7
Meaning and logic
The logical approach to meaning is a first step into the investigation of meaning
relations. Taking up the notions of truth and reference from chapter 2, we will
consider sentences from the perspective of their truth conditions. The logical view
allows the introduction of basic concepts such as logical consequence (entailment),
logical equivalence and incompatibility. In 7.5 these notions are applied to words.
7.1 LOGICAL BASICS
7.1.1 Donald Duck and Aristotle
Let us start with the provocative (and highly important) question: ‘Is Donald Duck a
duck?’ Suppose you are one of those who answer spontaneously: ‘Why, of course!’ In
that case you would subscribe to the truth of (1):
(1)
Donald Duck is a duck.
Well, ducks are birds, and birds are animals. Would you also say that (2) is true?
(2)
Donald Duck is a bird.
And how about (3)?
(3)
Donald Duck is an animal.
It would not be surprising if you were less sure about the truth of (2) and would not
subscribe to the truth of (3). But, if (1) is true, (2) is true, and if (2) is true, so is (3).
Hence, if (3) is false, there must be something wrong: (2) must be false as well and,
consequently, (1) cannot be true either. That is logic: if Donald is a duck, then he is a
bird. If he is not a bird, he cannot be a duck.
Well, then, let us take a second look at the original question. Why is it that we are
inclined to say that Donald is a duck? Well, it is a fact that his name is Donald ‘Duck’
and that Donald looks like a duck, at least roughly, i.e. if we ignore his having arms
with hands instead of a duck’s wings. But names are just names, and beyond his
looking like a duck there is little to be said in defence of Donald’s duckness. Does he
quack rather than talk? Does he swim or fly like a duck? Would we expect him to dive
168
Understanding semantics
for food as ducks do? No. As far as we know, Donald Duck behaves, feels, and thinks
in every respect like a human being. So, let us try this:
(4)
Donald Duck is a human being.
Have you ever seen a human being with a duck’s body, with feathers, a beak and duck
feet? Is he not much too short for an adult man? Could he run around without his
pants all the time if he were not a duck? If we are serious about the initial question,
we have to admit that (4) is not true either:
(5)
Donald Duck is neither a duck nor a human being.
But if we decide to take this stand, we are throwing out the baby with the bath
water. According to (5), Donald could be anything except a duck or a human. This
is certainly not what we want to say. If anything, Donald is a duck or a human.
Somehow, he’s both at the same time:
(6)
Donald Duck is both a duck and a human being.
He is a duck that behaves like a human, and he is a human being in a duck’s guise. If
we take (5) and (6) together, we get (7) and (8):
(7)
Donald Duck is a duck and he isn’t.
(8)
Donald Duck is a human being and he isn’t.
What does logic say about this? That is very clear: (6) contradicts (5), and (7)
and (8), as they stand, are each self-contradictory. This cannot be: (6), (7) and (8)
cannot be true. Therefore, Donald Duck cannot exist, or (6), (7) and (8) are false.
The consequence is acceptable, in a sense. In a world where ducks are ducks and
cannot be human, and vice versa (e.g. in what we consider the real world), we would
not accept that something or someone like Donald Duck really exists. We would not
accept that (6), (7) and (8) are true of anything that really exists.
The underlying principle goes back as far as Aristotle. In his work Metaphysics, he
formulated the following fundamental law of logic, and not only of logic, but of truth
in general (Aristotle, Metaphysics, 1005b, p. 262):
DEFINITION 1 Law of Contradiction
The same attribute cannot at the same time both belong and not belong to the
same subject in the same respect.
Meaning and logic
169
Aristotle assumes that, basically, every sentence says that some attribute – the
predicate of the sentence – belongs to some subject. What the principle says is
simply this: a sentence, in a certain reading, cannot be true and false at the same
time. Our reasoning about Donald Duck with the outcome of (5), (6), (7) and (8)
violates this law. If (5) is true, (6) must be false, and vice versa. So, if (5) and (6) are
both true, they must also be both false. (7) says that (1) is both true and false, and
so does (8) for (4).
But, seriously, is not there something to be said in favour of the truth of (5), (6),
(7) and (8)? Yes, there is. And if we take a closer look at Aristotle’s law, we realize how
our findings about Donald Duck can be reconciled with logic: we have to relate the
categorization of poor Donald to different ‘respects’. The apparent contradictions can
be resolved if we replace (5), (6), (7) and (8) by the following:
(5’)
Donald Duck is neither a duck nor a human being in all respects.
(6’)
Donald Duck is a duck in certain respects and a human being in others.
(7’)
Donald Duck is a duck in certain respects, but he isn’t in others.
(8’)
Donald Duck is a human being in certain respects, but he isn’t in others.
or more explicitly:
(5”) Donald Duck doesn’t behave like a duck and doesn’t look like a human being.
(6”) Donald Duck looks like a duck but behaves like a human being.
(7”) Donald Duck looks like a duck but doesn’t behave like one.
(8”) Donald Duck behaves like a human being but doesn’t look like one.
These sentences are no longer contradictory. They are, however, still not compatible
with our experience of the real world. Hence, if we accept the truth of these sentences,
we have to assume a different world for Donald Duck to exist in – and that, of course,
is what we do.
As you might have noted, interpreting the original sentences (5) to (8) in the more
explicit way given in (5’) to (8’) or (5”) to (8”) is an instance of what we introduced
in 3.4.3 as ‘differentiation’. The expressions be a duck and be a human being are
interpreted in a more specific reading than their literal meaning. The process is
triggered by the need to make things fit into their context, i.e. by applying the
Principle of Consistent Interpretation.
170
Understanding semantics
7.1.2 The Principle of Polarity
The basic notion of all logical considerations is truth. As was stated in 1.1.2 above,
truth is not a property of sentences as such, although we mostly talk that way.1 The
question of the truth or falsity of a sentence arises only if the sentence is related to
a certain CoU. Since the CoU may vary, sentences are true in some CoUs and false in
others. Truth and falsity underlie the following fundamental principle:
DEFINITION 2 Principle of Polarity
In a given CoU, with a given reading, a declarative sentence is either true or
false.
This principle too goes back to Aristotle. It entails (for the notion of logical entailment
see the next subsection) the Law of Contradiction since the formulation ‘either true
or false’ is taken in the exclusive meaning of either–or: ‘either true or false, but not
both’. The ‘but not both’ part is the Law of Contradiction. The Principle of Polarity
adds to the Law of Contradiction the condition known as the Law of the Excluded
Middle (Latin Tertium non datur, which means ‘there is no third [possibility]’): there
are only these two possibilities, truth or falsity, and no others, i.e. no in-between, no
both-true-and-false, no neither-true-nor-false. Later in 11.5 we will see how this
principle is reflected in the structure and use of natural language.
The principle as it stands disregards presuppositions. As we saw in 4.7, a sentence
usually carries presuppositions, and if the context fails to fulfil these conditions,
the sentence will indeed be neither true nor false. For the time being, we will just
assume that the sentences we are dealing with are only used in CoUs where their
presuppositions are fulfilled. This is the usual way in standard logic to circumvent
the problems caused by presuppositions. Actually, logic becomes much more
complicated if presuppositions are properly taken into account. We will come back to
this issue at the end of this chapter.
In order to have a neutral term for being true or false, one speaks of the truth value
of a sentence, a notion introduced by Frege. A sentence has the truth value TRUE if it is
true; it has the truth value FALSE if it is false. In general, the truth value of a sentence
becomes an issue only when it is used in a certain CoU. The truth conditions (2.2.2)
of the sentence define the circumstances under which a sentence is true, i.e. in which
CoUs. For example, (9) is true if, in the given CoU, the cat is in the garden, and (10) is
true if there’s milk on the floor:
(9)
The cat is in the garden.
(10) There’s milk on the floor.
1 When we talk of sentences in this chapter, it is tacitly understood that we talk of declarative
sentences. The question of truth or falsity does not immediately apply to interrogative sentences,
imperative sentences or other non-declarative types (recall 2.2.3 on sentence types).
Meaning and logic
171
In a given CoU, referring to a certain cat, a certain garden and a certain floor (of a
certain room), the two sentences may both be true, or both be false, or one may be
true, and the other one false. Their truth conditions are independent of each other.
7.1.3 Negation
Due to the Principle of Polarity, any declarative sentence, for example (11a), when
uttered in a particular CoU, is either true or false. The same holds for its negation, in
this case (11b):
(11) a.
b.
John knows the solution.
John does not know the solution.
By asserting (11a), I not only express that I think that John knows the solution, but
I deny that John does not know the solution. Conversely, if I utter (11b), I deny that
(11a) is true. Any statement that one can express in a natural language is, in this
sense, polarized – it is the result of a decision between just two opposite possibilities:
yes or no, true or false. Polarization pervades language totally and the Principle of
Polarity captures this trait of human verbal communication. As we shall see in 7.7,
this is not relativized by the existence of presuppositions.
It is no surprise then that all languages have systematic means of expressing
the polar contrary of a sentence: negation. Negation reverses the truth value of
a sentence; it makes a true sentence false and a false sentence true. We discussed
negation in 4.5.3 and gave a simple method of determining the negation of a
declarative sentence: transform the sentence into a yes–no question, answer it with
‘No’ and form the sentence that expresses exactly what the negative answer says. In
English, negation is usually achieved by negating the finite verb of the sentence, using
the auxiliary do if the verb is not an auxiliary itself.
(12) a.
b.
c.
John knows the solution.
John will die.
John is clever.
negation:
negation:
negation:
John doesn’t know the solution.
John will not die.
John isn’t clever.
In other cases, negation is formed by negating other parts of the sentence, for
example quantifiers such as all, every, some, always and the like, or by replacing them
with appropriate negative expressions (for quantificational NPs, recall the discussion
in 4.5.4):
(13) a.
b.
c.
d.
e.
Mary is already here.
Everybody knows her.
She’s always late.
She sometimes apologizes.
Only John knows the reason.
negation:
negation:
negation:
negation:
negation:
Mary is not here yet.
Not everybody knows her.
She’s not always late.
She never apologizes.
Not only John knows the reason.
172
Understanding semantics
For the present purposes, we need not be concerned with the exact rules and
subtleties of negation in English. Let us simply define the negation of a sentence
as follows
DEFINITION 3 Negation
If A is a declarative sentence that is not negated itself, its negation is a sentence
that
(i) is true whenever A is false and false whenever A is true and
(ii) is formed out of A by a standard grammatical procedure such as
– adding not to the verb itself, if it is an auxiliary (e.g. was not, cannot)
– adding the auxiliary do to the verb phrase and not to the auxiliary
(e.g. did not know)
– adding not to a quantifier expression (e.g. not every)
– substituting a positive expression by its negative counterpart (e.g. some
by no).
There are only a handful of expressions that require negation by substitution:
quantifier expressions containing some (some – no, somewhere – nowhere, etc.) and a
couple of particles such as already and still (negation: not yet and no more). Negation
by substitution is only relevant if regular syntactic negation with not is impossible.
Table 7.1 displays some examples of exceptional negations. Usually, a positive
sentence has exactly one negation, and this is what will be assumed throughout this
chapter. For the sake of convenience not-A will be used as an abbreviation of the
negation of A.
Table 7.1
Types of exceptional negations in English
Sentence
Negation
He’s always late.
He’s not always late.
Everybody knows that.
Not everybody knows that.
He sometimes apologizes.
He never apologizes.
He was already there.
He was not yet there.
You must insist.
You need not insist.
Only Claude knows why.
Not only Claude knows why.
7.2 LOGICAL PROPERTIES OF SENTENCES
Given the notion of truth conditions, a couple of basic logical properties of sentences
can be defined. A ‘normal’ sentence will be sometimes true and sometimes false.
Meaning and logic
173
This property is called contingency. There are two kinds of sentences that are not
contingent: sentences that are always true and those that are always false. 23
DEFINITION 4 Logical properties of sentences
A sentence (in a given reading) is contingent iff 2 it is true in some CoUs and
false in others.
A sentence (in a given reading) is logically true iff it is true in all CoUs.
A sentence (in a given reading) is logically false iff it is false in all CoUs.3
The three logical notions can be alternatively defined by means of tables displaying
which truth values are possible or impossible in the respective cases (cf. Table 7.2).
As is common practice in logic, 1 is used for TRUE and 0 for FALSE. Note that the
‘impossible’ entries in the tables for logical truth and logical falsity imply that the
other truth value is possible since a sentence cannot have no truth value at all. As this
follows automatically, there is no need to write ‘possible’ into the free slot.
Table 7.2
Logical properties of sentences
A is contingent
A
A is logically true
A
1
possible
1
0
possible
0
A is logically false
A
1
impossible
impossible
0
The sentences in (14) are logically true:
(14) a.
Either Donald Duck is a duck or he is not a duck.
b.
Every duck is a duck.
c.
Ducks are birds.
d.
Two times seven equals fourteen.
(14a) is true in every CoU due to the Principle of Polarity. We might replace Donald
Duck by any other subject and is a duck by any other predicate. It is the sentence
pattern ‘either x is p or x is not p’ which makes the sentence true, independent of any
contextual conditions. Likewise, (14b) is invariably true due to the structure of the
sentence, and (14c) is true because the words duck and bird mean what they mean,
i.e. due to the semantic facts of English. (14d) is a mathematical truth. Two, seven
and fourteen when used as NPs like here always refer to the same abstract objects,
2 Iff is commonly used as an abbreviation of if and only if; it connects two conditions that are
equivalent. If and only if constructions are the proper form of precise definitions.
3 In other terminologies, logically true sentences are called tautologies (tautological) and logically
false sentences, contradictions (contradictory). Informally, logically false sentences were referred to
as ‘self-contradictory’ above.
174
Understanding semantics
the numbers 2, 7 and 14; their referents cannot vary with the choice of a CoU, nor
can the outcome of multiplying 2 with 7. In philosophical terminology, sentences
such as (14c) and (14d) are called analytic(ally true), while the notion of logical truth
(and falsity) is reserved for cases such as (14a) and (14b) which owe their truth (or
falsity) to specific rules of logic. From a linguistic point of view, there is no essential
difference between the type of truth represented by the four sentences: all four of
them are true due to their structure and the meanings of the words they contain.
The following sentences are logically false:
(15) a.
Donald Duck is a duck and Donald Duck is not a duck.
b.
Donald Duck is neither a duck nor is he not a duck.
c.
Ducks are plants.
d.
Two times seven is twenty-seven.
(15a) violates the Law of Contradiction, (15b) the Law of the Excluded Middle. (15c)
violates the semantic rules of English and (15d) the rules of mathematics.
The examples show that even logical truth and falsity rest on some basic
assumptions:
∑ the Principle of Polarity
∑ the semantics of the language
These assumptions are absolutely indispensable. The Principle of Polarity is at the
very heart of the notions of truth and falsity. The semantic rules of language are
necessary to make it possible to deal with questions of truth at all. If the sentences
and the words they consist of did not have their proper meanings, there would be no
point in asking any logical or semantic questions.
The logical properties of a sentence are connected to the information it is able to
convey. Contingent sentences may be true or false. Thus, when they are actually used
for assertions, they convey information about the situation referred to: it must be
such that the situation expressed does in fact pertain. If John tells Mary there is beer
in the fridge, she learns something about the situation referred to (provided John
is not lying and Mary believes him). If John uttered a logically true sentence like
Ducks are birds, Mary would not learn anything about the given situation. She would,
at best, learn something about English. Similarly, she would ask herself what John
wants to tell her by saying Either Donald Duck is a duck or he is not a duck. Taken as a
MESSAGE, as information about the world, it would be uninformative. Only contingent
sentences can convey information about the world.
If logically true or logically false sentences are actually used to say something
about the situation in the given CoU, application of the Principle of Consistent
Interpretation (3.4.3) leads to a reinterpretation. Logically false sentences are freed
from their inner contradictions and can then be interpreted as contingent statements.
For example, the compositional meanings of the logically false sentences in (6), (7)
and (8) can be shifted, by means of differentiation, to yield the contingent readings
in (6”), (7”) and (8”). Logically true sentences, too, can be shifted to contingent
Meaning and logic
175
readings. For instance, the logically true sentence Ducks are birds can be used to
point out the fact that ducks are able to fly; this might be relevant in some contexts,
e.g. when the question is discussed whether or not Donald Duck is a duck. From this
interpretation of the sentence, the contingent implication would derive that Donald
could fly (if he were a duck).
7.3 LOGICAL RELATIONS BETWEEN SENTENCES
7.3.1 Logical entailment
Recall our discussion above: necessarily, (16B) is true, if (16A) is true. This is an
instance of a logical relation between sentences called ‘logical entailment’.
(16) A
Donald Duck is a duck.
B
Donald Duck is a bird.
The relation of logical entailment is defined by one crucial condition: it must be
impossible that B is false if A is true. (This is the case with A and B in (16): it is
impossible that Donald is not a bird (= B false), if he is a duck (= A true).)
DEFINITION 5
A
B
A logically entails B /
B logically follows from A /
AfiB
1
1
0
1
0 impossible
1
iff : necessarily, if A is true, B is true.
0
0
Two arbitrary sentences A and B may be independently true or false. This yields four
possible combinations of truth values. Logical entailment rules out one of these four
combinations, A-true-B-false. If A entails B, the truth values of A and B depend on
each other in a particular way: B cannot be false if A is true, and A cannot be true if B
is false. Thus, logical entailment results in a certain link between the truth conditions
of the two sentences. If A entails B, then A is called the premise of the entailment,
and B the conclusion.
The definition of entailment does not specify whether A and B are actually true. A
and B might both be true or both be false, or A might be false and B true. Entailment
means: if A is true, then B is true. But it does not stipulate whether A is true and
therefore one does not know if B is true, either.
If A entails B, it may or may not be the case that B also entails A. If it does, we
talk of mutual entailment. Mutual entailment means that not only A-true-B-false
is impossible, but also A-false-B-true, i.e. B-true-A-false. This results in a second
‘impossible’ entry, in the third row. Alternatively – and this is the normal case – B
176
Understanding semantics
does not entail A. In this case, A unilaterally entails B; the combination A-falseB-true is therefore possible. Thus, entailment is either unilateral or mutual because
A-false-B-true is either possible or impossible, respectively. Table 7.3 shows the
general condition for entailment along with the conditions for the two subcases of
unilateral and mutual entailment.
Table 7.3
Logical entailment and special cases
General entailment
A B
1
1
1
0
0
0
Unilateral entailment
A B
Mutual entailment
A B
1
1
1
1
1
0
impossible
1
0
impossible
1
0
1
possible
0
1
impossible
0
0
0
0
0
impossible
(16) is a unilateral entailment because Donald need not be a duck if he is a bird.
Likewise, the following entailments are unilateral:
fi B
(17) A
It’s raining heavily.
(18) A
Ann is a sister of my mother. fi B
(19) A
Today is Monday.
fi B
It’s raining.
Ann is my aunt.
Today isn’t Wednesday.
If it is raining, it need not be raining heavily; Ann could be my aunt because she is
my father’s sister; it may not be either Wednesday or Monday. (20) is an example of
mutual entailment.
(20) A
Today is Monday. fi
‹ Tomorrow is Tuesday.
Mutual entailment will be called ‘logical equivalence’ in the next subsection. Since
there is a term of its own for this special case of entailment, the term ‘entailment’ is
normally taken as denoting the unilateral case.
There is one way of reversing an entailment: if A entails B, then, if B is false, A is
necessarily false; if B could be false and A not, the entailment would not hold. Thus,
if A entails B, then not-B entails not-A. We can swap the two sides of an entailment if
we negate both sides. Therefore, the converse holds, too: if not-B entails not-A, then
A entails B.
(21) Entailment reversal
A fi B iff not-B fi not-A.
Table 7.4 shows how the truth values of not-A and not-B co-vary with those of A
and B.
Meaning and logic
177
Table 7.4
Entailment reversal
A entails B = not-B entails not-A
A
B
1
1
1
0
not-B not-A
impossible
0
0
1
0
0
1
0
1
0
0
1
1
Let us now take a look at a few examples which are not instances of logical entailment,
although in each case sentence B would, under normal circumstances, be inferred
from A. What matters, however, is whether the consequence is really necessary or
whether it is based on additional assumptions.
(22) A
Mary is John’s mother.
fi
/ B
Mary is the wife of John’s father.
(23) A
John said he is tired.
fi
/ B
John is tired.
(24) A
The beer is in the fridge. fi
/ B
The beer is cool.
There are no logical reasons for drawing these conclusions. It is logically possible that
parents are not married, that John was lying, and so on. In most cases we draw our
conclusions on the basis of our world knowledge, i.e. of what we consider as given,
as normal, plausible or probable. The notion of logical entailment does not capture
all these regularities and connections. It merely captures the really ‘hard’ cases of an
if-then relation, those based on the Principle of Polarity and the semantic facts alone.
Other, more mundane inferences like those in (22) to (24) could be turned into valid
entailments too – if all the tacit premises were added.4 For instance, in the case of
(24) we would have to add that the fridge was on and functioning normally, that the
beer was in it long enough, that it was not in a vacuum flask, etc. If you come to think
of it, it is like a never-ending story of additional premises. In everyday life, when we
actually say things like ‘If the beer is in the fridge, then it is cool’ we rely on a rich
body of shared knowledge and assumptions which need not be made explicit but
are nevertheless necessary for drawing the conclusions we draw. Thus, an ordinary
if-statement involves a huge complex of implicit premises. Logical entailment strips
an inference of all implicit assumptions by making explicit what the premise is (or
the premises are) and relating only to these.
What does logical entailment mean for the meanings of A and B? If A and B are
contingent and A unilaterally entails B, both sentences contain information about
the same issue, but the information given by A is more specific than the information
given by B. The truth conditions that B imposes on the situation are such that they
4 The notion of entailment can be straightforwardly generalized to include more than one premise.
Also, more than one premise can be made into one by connecting all with and into one sentence.
178
Understanding semantics
are always fulfilled if A is true. Therefore, the truth conditions of B must be part of
the truth conditions of A. In general, if no further logical relation holds between A
and B, A will impose additional conditions on the situation referred to. In this sense,
A contains more information; it is more informative and more specific, than B. The
situation expressed by A is a special case of the situation expressed by B. As we shall
see in 7.3.5, this does not hold if A and/or B are not contingent.
One further property should be noted here: logical entailment is what is called
a transitive relation. The general property of transitivity5 is defined as follows: a
relation R is transitive iff ‘x is in relation R to y’ and ‘y is in relation R to z’ entails ‘x is
in relation R to z’. Applied to entailment, this means the following:
(25) Transitivity of the entailment relation
If A fi B and B fi C, then A fi C.
For example, Donald is a duck fi Donald is a bird; Donald is a bird fi Donald is an
animal; hence Donald is a duck fi Donald is an animal. The property of transitivity
immediately follows from the way entailment is defined. Suppose A fi B and B fi C;
then if A is true, necessarily B is true; if B is true, necessarily C is true, hence: if A is
true, necessarily C is true, i.e. A fi C.
7.3.2 Logical equivalence
DEFINITION 6
A and B are logically equivalent,
A¤B
iff:
necessarily, A and B have equal truth values.
A
B
1
1
0
1
0 impossible
1 impossible
0
0
Equivalence means having identical truth conditions. Like entailment, equivalence
is a transitive relation, but unlike entailment it is a symmetric relation. Since
the combinations A-true-B-false and A-false-B-true are both ruled out, the table
combines the conditions for A fi B and B fi A: equivalence is mutual entailment.
Thus, if A and B are contingent, A must contain all the information B contains and
B must contain all the information A contains. In other words, the sentences must
contain the same information. Let us consider a few examples:
(26) A
He is the father of my mother. ¤ B
(27) A
Today is Monday.
¤ B
Yesterday was Sunday.
(28) A
The bottle is half empty.
¤ B
The bottle is half full.
(29) A
Everyone will lose.
¤ B
No-one will win.
He is my maternal grandfather.
5 There is no connection between the notion of a ‘transitive relation’ and the syntactic notion of a
‘transitive verb’.
Meaning and logic
179
The equivalence in (26) is due to the definition of maternal grandfather. For (27), we
have to assume that every Monday is necessarily preceded by a Sunday, an assumption
that may be taken for granted for the point to be made here. The equivalence in (29)
holds if we assume a reading of lose and win in which lose means ›not win‹. Given
these assumptions, all four cases rest merely on the semantic facts of English.
7.3.3 Logical contrariety
DEFINITION 7
A is logically contrary to B /
A logically excludes B / B is incompatible with A
iff:
necessarily, A and B are not both true.
A
B
1
1
0
1 impossible
0
1
0
0
The defining condition is that A and B cannot both be true. Obviously, this is a
symmetric relation. If A and B are contraries, then A entails not-B and B entails
not-A: the ‘impossible’ entry in the first row yields both. Other common terms for
contrariety are logical exclusion and incompatibility. Examples are:
(30) A
It’s cold.
B
It’s hot.
(31) A
Today is Monday.
B
Today is Tuesday.
(32) A
Ann is younger than Mary.
B
Ann is older than Mary.
If A and B are incompatible, it may or may not be possible that both are false. Usually,
they may both be false. In (30) to (32) this is the case: it may be neither cold nor
hot, neither Monday nor Tuesday, and Ann and Mary may be the same age. The two
contraries do not form an exhaustive alternative – there are more possibilities. If,
however, two contraries cannot both be false, they form an either-or-alternative. This
special case of contrariety is the last of the four logical relations introduced here.
7.3.4 Logical contradiction
DEFINITION 8
A and B are logical contradictories
iff:
necessarily, A and B are not both true.
A
B
1
1
0
0
1 impossible
0
1
0 impossible
The definition of contradiction adds to the definition of contrariety the condition
that A and B cannot both be false. If A and B are contradictories, then in every CoU
180
Understanding semantics
either A is true and B is false or B is true and A is false. The classical case of logical
contradiction is formed by a sentence and its negation (33); (34) and (35) show,
however, that there are other cases as well:
(33) A
It’s late.
B
(34) A
Today is Saturday or Sunday. B
Today is Monday, Tuesday, Wednesday,
Thursday or Friday.
(35) A
Everyone will win.
Someone will lose.
B
It’s not late.
Although the B sentences in (34) and (35) are not the negations of the A sentences,
they are nevertheless logically equivalent to the respective negations, today is neither
Saturday nor Sunday and not everyone will win. A sentence and its negation are by
definition always logically contradictory.
Logical contradiction of A and B is logical equivalence of A and not-B, or of B and
not-A. Hence, contradiction too can be defined in terms of entailment.
(36) The logical relations in terms of entailment
A and B are contraries
iff A entails not-B
A and B are equivalent
iff A entails B
and
A and B are contradictories iff A entails not-B and
B entails A
not-A entails B
Table 7.5 displays the crucial conditions for the four logical relations we introduced.
Each ‘impossible’ entry corresponds to an entailment relation. The table also shows
that entailment subsumes equivalence, and contradiction is a special case of
contrariety.
Table 7.5
Logical relations
A
B
1
1
1
0
0
1
0
0
A¤B
A¤B
impossible
impossible
contraries
contradiction
impossible
impossible
impossible
impossible
7.3.5 Logical relations involving logically true or false sentences
According to the definition, there is one simple condition for A logically entailing B:
A-true-B-false is impossible. This condition is automatically given if either A-true or
B-false is impossible. Consequently, A generally entails B if A is logically false or if
B is logically true. More specifically: if A is logically false, entailment holds for any
Meaning and logic
181
conclusion whatsoever, and if the conclusion is logically true, it logically follows from
any premise whatsoever. For logically false A, we even have A entails not-A, and for
logically true B, not-B entails B. This is, however, no contradiction. In both cases, the
premise is never true and hence the entailment is never carried out. The two cases,
A logically false and B logically true, each combined with an arbitrary partner B and
A, respectively, yield the picture in Table 7.6. In mathematical jargon, these cases are
called ‘pathological’.
Table 7.6
Pathological cases of entailment
A is logically false
B is logically true
A B
A B
1
1
impossible
1
1
1
0
impossible
1
0
0
1
0
1
0
0
0
0
impossible
impossible
With logically false premises or logically true conclusions, we get entailments such
as the following. In (37a), A is logically false, in (37b) B is logically true. Mary is tired
is contingent. Putting (37a) and (37b) together, we obtain (37c) by transitivity of
entailment:
(37) a.
A Ducks are dogs. fi B Mary is tired.
fi B Ducks are birds.
b.
A Mary is tired.
c.
A Ducks are dogs. fi B Ducks are birds.
The point with these examples is that they are cases which one would never actually
state as entailments. If the premise cannot be true at all, there is no point in stating
that if it were true then something else would follow.6 Likewise, if the conclusion is
true anyway, it makes no sense to point out that it is true assuming that some premise
holds. This is another difference between natural language ‘if ’ and logical entailment.
The use of if–then statements in actual communication is subject to pragmatic
conditions: they normally make no sense if the if-clause could not be true or the thenclause not false. Logical entailment, however, is a logical relation that simply holds
or doesn’t. If it holds, it does so regardless of its pragmatic relevance. The notion
that these cases are ‘pathological’ reflects the fact that stating them violates the usual
pragmatic conditions.
When you learnt about logical entailment in the preceding sections, you will
have thought about possible relevant applications of the notion. In doing so you
6 We do, however, exploit this case rhetorically when we say things like: ‘I’ll eat my hat if [A]’ in order
to express that A is impossible.
182
Understanding semantics
naturally did not think of the pathological cases, but of instances where premise
and conclusion are contingent. This led you to forming the intuition that there is
a connection in meaning between premise and conclusion of a valid entailment.
And if you really captured the notion, your understanding will have formed that the
connection is a purely semantic one, not involving any further implicit premises.
However, the examples in (37a) and (37b) show that one may have logical
entailment between sentences even if they lack any semantic connection. The same
holds for the logical relations of equivalence, contrariety and contradiction. One
can easily find these relations in semantically unrelated sentences, if one or both
are not contingent. For example, all logically true sentences are equivalent, and so
are all logically false sentences. All logically false sentences are incompatible and
contradictory with all logically false sentences and vice versa. Therefore – and this is
very important to realize – logical relations are not meaning relations. They may hold
good independently of the meaning of the sentences involved, and even despite the
meaning relations between them, as in the case of (37c).
The pathological cases yield not only unexpected logical relationships, but also
unexpected combinations of relationships. For example, if A is logically false and
B an arbitrary sentence, then A entails B and A and B are contraries – due to the
‘impossible’ entries both in row 2 and row 1, respectively.
As we will see in a minute, all this does not mean that our intuitions about a
connection between logical relations and meaning have to be thrown overboard
altogether. If we consider only contingent sentences, a connection does exist. But for
the moment, it is important to realize that logical relations do not warrant a meaning
relation, let alone a particular meaning relation.
7.3.6 Logical relations between contingent sentences
Let us assume that A and B are both contingent. This has far-reaching consequences
for the significance of the logical relations. First of all, we may fill in the entry ‘possible’
into many cells of the defining tables. For example, the definition of entailment rules
out A-true-B-false. The assumption that A is contingent allows us to fill in ‘possible’
in row 1 because otherwise A would turn out to be logically false. Likewise, we can
make the same entry in row 4: A-false-B-false must be possible because otherwise B
would be logically true. Table 7.7 displays the resulting picture for the four relations.
You can easily figure out the other entries yourself.
Table 7.7
Logical relations between contingent sentences
A
B
1
1
1
0
0
1
0
0
AfiB
A¤B
contraries
contradict
possible
possible
impossible
impossible
impossible
impossible
possible
possible
impossible
possible
possible
possible
possible
impossible
Meaning and logic
183
The restriction to contingent sentences renders the four relations much more
specific. Compared to the original definitions in Table 7.5, the relations here all carry
two ‘possible’ entries in addition to the defining ‘impossible’ entries. For example,
the general definition of equivalence leaves open whether the cases A-true-B-true
and A-false-B-false are possible or not. Fixing these issues makes equivalence
between contingent sentences a more specific relation than equivalence in general.
As a consequence, the more specific relations can no longer co-occur more or less
freely. This is so because ‘possible’ and ‘impossible’ entries in the same row are
incompatible, while ‘impossible’ and no entry are compatible. Therefore entailment
in general is compatible with contrariety. But for contingent sentences, entailment
and contrariety have incompatible entries in row 1 and hence exclude each other.
More generally, entailment and equivalence between contingent sentences are
incompatible with contrariety and contradiction.
Still, some cells remain open. The open cell in row 3 for entailment makes
equivalence compatible with entailment, and the open cell in row 4 makes contrariety
compatible with contradiction. We have already made these two observations above.
They reflect that equivalence is a special case of entailment, and contradiction of
contrariety.
Within the domain of contingent sentences a further logical relation can
be introduced, the relation of non-relatedness, as it were. It is called logical
independence and holds between two sentences iff all four truth value combinations
are possible. This amounts to A entailing neither B nor not-B and B entailing neither
A nor not-A, or in other words: the truth value of A does not determine the truth
value of B or vice versa.
When one tries to find examples of pairs of contingent sentences that are related by
one of the logical relations (except independence), one will realize that, now indeed,
this is only possible if the sentences bear some meaning connection. For example, if
two sentences have the same truth conditions and are hence logically equivalent, then
they must have closely related meanings, because it is the meanings that determine
the truth conditions. It cannot be formally proved that a logical relation between
contingent sentences is always due to some meaning connection. But the assumption
is one of the most important working hypotheses for semantics. It can be formulated
as follows:
Working hypothesis
If two contingent sentences exhibit the relation of logical entailment,
equivalence, contrariety or contradiction, this is due to a particular way in
which their meanings are related.
The restriction to contingent sentences does not impose any serious limitation on
the field of semantic research. Therefore, logical relationships are very valuable
instruments for the investigation of meaning relations not only of sentences but also
of words (to which we will turn in 7.5).
184
Understanding semantics
7.4 SENTENTIAL LOGIC
7.4.1 Logical connectives
Sentential logic7 (SL, for short) is a simple formal system with rules for combining
sentences. The sentences are usually simply represented by variables. They are
combined by means of certain basic connectives such as and and or. Needless
to say, the Principle of Polarity is assumed to hold: every simple or complex SL
sentence is either true or false. The only connectives to be considered are those
whose meaning can be exhaustively described in terms of the truth values of the
sentences they are applied to. (This rules out connectives such as because, when,
but, nevertheless, therefore, etc.) We will only introduce three such connectives: Ÿ for
›and‹ (conjunction), ⁄ for ›and/or‹ (disjunction) and Æ ›if … then‹ (subjunction);
in addition we introduce negation, written with the ‘negator’ ÿ.
DEFINITION 9
Negation in sentential logic
If A is an SL sentence, then ÿA is, too.
ÿA is true if A is false, and false if A is true.
Conjunction in sentential logic
If A and B are SL sentences, then (AŸB) is, too.
(AŸB) is true if A and B are both true, and false if either or both are false.
Disjunction in sentential logic
If A and B are SL sentences, then (A⁄B) is, too.
(A⁄B) is true if A and/or B are true, and false if A and B both are false.
Subjunction in sentential logic
If A and B are SL sentences, then (A→B) is, too.
(AÆB) is true if A is false and/or B is true, and false if A is true and B is false.
The subjunction symbol is usually read ‘if … then’, but its real meaning in SL is only
loosely related to natural language ›if‹. We will come back to this issue in a minute.
The truth conditions of complex SL sentences can be given in truth tables (Table
7.8). The right column shows the resulting truth value of the connection, given the
truth values of A and B in the first columns. Thus, for example, if A is false and B is
true (line 3 in the tables), then (AŸB) is false, (A⁄B) is true and so is (A→B).
7 Sentential logic is also called propositional logic and statement logic. We prefer the term sentential
logic, because the units of the system are sentences, rather than statements or propositions. It is
sentences which are connected by connectives, and it is sentences for which the logical notions are
defined.
Meaning and logic
185
Table 7.8
Truth tables for logical connectives
Negation
Conjunction
Disjunction
Subjunction
A
ÿA
A
B
(A Ÿ B)
A
B
(A ⁄ B)
A
B
(A → B)
1
0
1
1
1
1
1
1
1
1
1
0
1
1
0
0
1
0
1
1
0
0
0
1
0
0
1
1
0
1
1
0
0
0
0
0
0
0
0
1
With these four definitions, we can form complex expressions such as:
(38) a.
ÿÿA
b.
(A Ÿ ÿA)
c.
(A ⁄ ÿA)
d.
ÿ(A Ÿ ÿB)
e.
(A ⁄ ÿB)
f.
(A → (A ⁄ B)) etc.
It follows directly from the definition that certain complex sentences are logically false
or logically true due to their form. For example, all sentences of the form (A Ÿ ÿA)
are logically false: A and ÿA necessarily have opposite truth values; therefore they
can never be both true, and so (A Ÿ ÿA) is never true. This fact implements the Law
of Contradiction. Among the logically false sentences in (15), (15a) has this form.
The other four cases in (15) call for different explanations. Similarly, the Principle of
Polarity is reflected in the fact that (A ⁄ ÿA) is logically true: since A has to be either
true or false, either A is true or ÿA; it follows that (A ⁄ ÿA) is always true (cf. lines 2
and 3 in the truth table for disjunction).
7.4.2 Logical connectives and their natural language counterparts
It must be cautioned that the logical connectives do not necessarily coincide with
the natural language correspondents that are used for reading the formulae. Usually,
natural language negation does not negate the sentence as a whole, but only a certain
part of it. To see this, look at the grammatical side of negation (see Definition 3
in 7.1.3): it is only the finite verb or a quantifier that is negated, while parts of the
sentence remain unaffected. As we saw in 4.7, all presuppositions of a sentence are
preserved under negation. For example, negation of the simple sentence Mary is
sleeping (which would be Mary is not sleeping) does not affect the subject Mary:
negation does not deny that it is Mary whom the sentence is about.
The logical connectives Ÿ and ⁄ can only be used to connect sentences, but their
natural language counterparts and and or can also connect single words or phrases
such as adjectives, verbs, VPs and NPs.
186
Understanding semantics
(39) a.
Warsaw is [the capital of Poland and Poland’s biggest city].
b.
[John and Sheila] will marry.
c.
Mary [fell and broke her collar bone].
In (39a, b), and connects two NPs, in (39c), it is two VPs. Sometimes a sentence with
and connecting less than whole sentences can be transformed into an equivalent
conjunction of two sentences. This is the case with (39a), but look what happens if
we apply this procedure to the other two sentences:
(39) a’.
Warsaw is the capital of Poland and Warsaw is Poland’s biggest city.
b’.
John will marry and Sheila will marry.
c’.
Mary fell and Mary broke her collar bone.
The most plausible reading of (39b) is that John and Sheila marry each other
(‘reciprocal’ reading), but (39b’) would rather be used if each of them marries
independently. (39b) may also be used in the sense of (39b’), but (39b’) cannot have
the reciprocal reading. In (39c) we would naturally assume that there is a causal
connection between Mary falling and breaking her collar bone: she broke her collar
bone when she fell. By contrast, (39c’) sounds like relating to two independent events.
Disjunction by natural language or is mostly used in the sense of ›either … or‹
and not in the sense of logical ›and/or‹. Also, there are constructions such as the one
in (40), where or is rather equivalent to Ÿ . Imagine Sheila is asked what kind of ice
cream she would like, and she answers (40a). This is not the same as (40b), which
sounds as though she could not make up her mind. Rather (40a) amounts to (40c).
(40) a.
Pecan or chocolate is fine.
b.
Pecan is fine or chocolate is fine.
c.
Pecan and chocolate are fine.
In the case of subjunction, the reading of → as ‘if … then’ is actually strongly
misleading. Natural language if A then B is only used if there is a connection between
A and B. For example, if one says (41), it will be interpreted as saying that there is
some regularity to A and B being both true; for example, Paul regularly drinks on
weekends and will have a hangover on Mondays. At least we will assume that it being
Monday and Paul having a headache is something of a general observation – for
whatever reason.
(41) If it is Monday, Paul has a headache.
Subjunction, by contrast, does not express any regularity that would connect B to
A. It just says that the truth value constellation of A and B is such that we do not
have A-true-B-false in the given CoU. Thus, subjunction is not really the logical
Meaning and logic
187
counterpart of natural language if. I would rather recommend reading (A→B)
simply as ‘A arrow B’.
There is a straightforward connection between subjunction and logical entailment:
if (A→B) is true in any CoU whatever, then A entails B. This is so because if (A→B)
is necessarily true, we never have A-true-B-false, whence A entails B. Conversely, if
A entails B, then the subjunction is true in every CoU. Thus, subjunction is a much
weaker statement than entailment because it does not claim anything beyond the
particular CoU given. As we have seen in 7.3.1, even entailment is usually a weaker
relation between two propositions than is expressed by natural language if.
7.5 LOGICAL RELATIONS BETWEEN WORDS
The logical relations between sentences can easily be applied to establish
corresponding relations between lexemes and other expressions below sentence level.
To be precise, this is possible for all predicate expressions (chapter 5). To establish
logical relations between two expressions, we insert them into an appropriate test
sentence with variables representing their arguments and check the resulting logical
relations. In order to be able to represent the arguments with variables, the test
sentences for nouns and adjectives must be predicative constructions. Appropriate
test sentences are illustrated in Table 7.9.
Table 7.9
Test sentences
Test word
Test sentence
count noun
car
x is a car
mass noun
mud
x is mud
adjective
dirty
x is dirty
intransitive verb
smell
x smells
transitive verb
sell
x sells y
7.5.1 Logical equivalence
Let us first consider the case of equivalence. Examples are hard to find, but here are
two:
(42) A
x is a female adult ¤ B x is a woman
(43) A
x costs a lot
¤ B x is expensive
What follows from these equivalences for the meaning of the expressions? (42)
means that whatever can be called a woman can be called a female adult and vice
188
Understanding semantics
versa. More technically: the potential referents of woman and female adult are the
same, i.e. the expressions have the same denotation. Similarly, due to (43) cost a lot
and be expensive are true of the subject referent under the same conditions. Rather
than introducing a new term, we will extend the notion of logical equivalence to
words and complex expressions such as female adult and cost a lot.8
DEFINITION 10 Logical equivalence of predicate terms
Two predicate terms (in given readings) are logically equivalent iff they
necessarily yield the same truth value for the same arguments.
In terms of the denotations of A and B, A and B are logically equivalent iff their
denotations are identical (Fig. 7.4).
7.5.2 Entailment and logical subordination
Let us assume the test sentences for two expressions result in entailment:
(44) a.
b.
A x is a duck
fi B x is a bird
A x shrinks y fi B x changes y
According to (44a), whatever can be called a duck can be called a bird. Put more
technically, the denotation of duck, the more specific term, is included in the
denotation of the more general term bird. Due to the second entailment, the
denotation of shrink is part of the denotation of change. Every act of shrinking
something is an act of changing it.
The resulting relation between a general term and a specific term will be called
logical subordination (subordination for short):
DEFINITION 11 Logical subordination
A predicate term A is a logical subordinate of B (and B a logical superordinate
of A) iff B is true of some arguments whenever A is true of them.
In terms of the denotations of A and B, A is a subordinate of an expression B, iff the
denotation of A is included in the denotation of B (Fig. 7.4). If A is a subordinate
of B, B is called a superordinate of A. Alternatively, the term ‘entailment’ can also be
applied to verbs, adjectives and nouns. In set-theoretical terms, A is a subordinate of
8 Some authors consider logical equivalence a variant of synonymy, for example Cruse (1986: 88),
who uses the term cognitive synonymy for the same relation. Synonymy and equivalence must,
however, be distinguished, a point we will come back to in 7.6.1.
Meaning and logic
189
B iff the denotation of A is a subset of the denotation of B. In the cognitive terms to
be introduced in chapter 11, the denotation of a subordinate term is a subcategory of
the denotation of its superordinate terms.
Figure 7.1
Denotations of duck and bird
denotation of the more general term bird
denotation of the more
specific term duck
7.5.3 Logical incompatibility
Usually, a superordinate expression does not have just one subordinate, but a set
of co-subordinates. For example, all the other terms for types of bird, such as owl,
pigeon, penguin, sparrow, swan, are co-subordinates of duck. In addition, they are
mutually exclusive: x is a swan logically excludes x is an owl, and so on for all other
pairs. Two terms A and B will be called logically incompatible iff their denotations
have no elements in common, or equivalently:
DEFINITION 12 Logical incompatibility
Two predicate terms A and B are logically incompatible if they cannot both be
true of the same arguments.
The denotation of swan could be represented by the hatched circle in Fig. 7.1: an area
included within the area for the denotation of bird and not overlapping with the area
symbolizing the denotation of duck.
The representation of lexemes in hierarchy trees such as in Fig. 7.2, a hierarchy for
linguistic terms, is based on the two relations of subordination and incompatibility.
Since trees are used for representing different relations and structures (e.g. syntactic
trees for the syntactic structure), it is important to realize what the arrangement in
a given kind of tree signifies. In trees that depict lexical hierarchies, the vertical lines
express logical subordination. Co-subordinates are arranged at the same level and
understood to be mutually incompatible.
190
Understanding semantics
Figure 7.2
Hierarchy of word classes
word
noun
count
noun
verb
mass
noun
transitive
verb
adjective
intransitive
verb
The tree in Fig. 7.2 is incomplete in several respects. Further subordinates of word
could be added, e.g. article or preposition. We could also expand the tree by adding
subordinates of adjective. Further subdivisions would be possible below the lowest
level, distinguishing subclasses of count nouns, intransitive verbs, etc. In contrast, the
small tree in Fig. 7.3 is, in a sense, complete. In English, there are only two specific
terms for siblings: no further co-subordinates can be added to sister and brother. Also
there are no English words for subordinates of brother and sister. (Other languages,
e.g. Hungarian, Turkish and Japanese, have different terms for elder and younger
sisters and brothers.)
Figure 7.3
Hierarchy of sibling terms
sibling
sister
brother
7.5.4 Logical complementarity
The subordinates in Fig. 7.3 are not only incompatible but form an exhaustive
alternative, a strict either-or constellation. The corresponding test sentences x is a
sister and x is a brother are logical contradictories – provided we presuppose that x is
a sibling. This meaning relation is called logical complementarity: two terms A and B
are logically complementary iff their denotations have no elements in common and
together exhaust the set of possible cases. Equivalently:
DEFINITION 13 Logical complementarity
Two predicate terms A and B are logically complementary if they necessarily
have opposite truth values for all arguments.
Meaning and logic
191
The notion of complementarity is always relative to a given domain of relevant cases.
Absolute complementarity does not occur in natural languages. Take any ordinary
noun – for example, banana; try to imagine an absolute complementary, say, nonbanana. The denotation of non-banana would have to exclude bananas, but include
everything else that could be denoted by any noun whatsoever plus all those things
for which we do not have any expressions at all. A word with such a meaning is hard
to imagine. Good examples for complementaries are member–non-member (domain:
persons), girl–boy (domain: children), child–adult (domain: persons), indoors–
outdoors (domain: locations). A survey of the logical relations at word and sentence
level is given in Table 7.10.
Table 7.10
Logical relations between words
Word relation
Corresponding
sentence relation
Example
equivalence
equivalence
woman
–
subordination
entailment
bird
– duck
incompatibility
contrariety
duck
–
complementarity
contradiction
member –
female adult
swan
non-member
Logical relations between content words result in certain relations between their
denotations. They can be depicted with the Venn diagrams known from elementary
set theory in Figure 7.4.
Figure 7.4
Denotation relationships
equivaleent
sub
bordinate
incompatib
ble
compleementary
We emphasized above that the logical relations between sentences do not constitute
meaning relations. The same holds for the logical relations between words.
7.6 LOGIC AND MEANING
It will now be demonstrated why logical relations must not be confused with
meaning relations such as synonymy (two expressions have the same meaning) and
hyponymy (the meaning of an expression contains the meaning of a superordinate
expression). These and other meaning relations will be introduced in chapter 8.
192
Understanding semantics
7.6.1 The semantic status of logical equivalence
It is tempting to assume that logical equivalence means identity of meaning, and in
fact this is often done in the literature.9 However, if meanings are considered concepts,
this turns out to be wrong. All logical notions are based on truth conditions and
denotations. As we saw in 2.2.2, truth conditions and denotations are determined by the
meaning of an expression, but its meaning is not exhausted, or fully captured, by them.
Apart from that, logical properties and relations only concern descriptive meaning;
other dimensions of meaning such as social meaning, expressive meaning or the
meaning of sentence type are disregarded and therefore irrelevant for logical notions.
7.6.1.1 Truth conditions and non-descriptive meaning
As far as the restriction of logical notions to descriptive meaning is concerned, let
us recall the criteria for correct use with respect to descriptive, social and expressive
meaning that were stated in Table 2.5. If, for example, one describes the truth
conditions of the sentence this is a barbecue, one has to describe the denotation of
the word barbecue. This, in turn, indirectly says something about the descriptive
meaning of barbecue, because the descriptive meaning determines the denotation. In
this sense, truth conditions relate to descriptive meaning. But they have nothing to
do with social meaning and expressive meaning. For example, the German sentence
and the English sentence in (45) differ in the meanings of the pronouns Sie and you
(cf. 2.3.1; we take the rest of the sentences to be equivalent, in particular the verbs
verhaften and arrest).
(45) a.
b.
Ich werde Sie verhaften.
(German)
I will arrest you.
The German pronoun of address Sie has the same descriptive meaning, but in
addition a social meaning indicating a formal relationship between speaker and
addressee(s). The difference, however, has no bearing on the truth conditions. If
the speaker of the German sentence used the informal pronoun of address instead,
the resulting sentence would have exactly the same truth conditions, although it
might be socially inappropriate. Similarly, expressions with the same descriptive but
different expressive meanings do not differ in truth conditions. Opting, for example,
for (46b) rather than (46a) is not a matter of objectively given facts but of subjective
preference.
(46) a.
b.
John didn’t take his car away.
John didn’t take his fucking car away.
Consequently, it may happen that words and sentences are logically equivalent, but
differ in non-descriptive meanings. We will now see, that logical equivalence does not
even mean equal descriptive meaning.
9 For example in Lyons (1995: 63).
Meaning and logic
193
7.6.1.2 Logical equivalence and descriptive meaning
As we saw in 7.2, all logically true sentences have identical truth conditions. Hence
they are all logically equivalent. Clearly, logically true sentences may differ in
descriptive meaning (cf. the examples in (14) and (15)). The same, of course, holds
for logically false sentences. Thus non-contingent sentences provide a particularly
drastic class of examples of logically equivalent sentences with different meanings.
But even for contingent sentences, equivalence does not mean that they have the
same descriptive meaning. To see the point, consider once more sentences (27) to
(29), here repeated for convenience:
(27) A
Today is Monday.
¤ B
Yesterday was Sunday.
(28) A
The bottle is half empty. ¤ B
The bottle is half full.
(29) A
Everyone will lose.
¤ B
No-one will win.
In all three cases, A and B intuitively do not have the same meaning, but somehow
they amount to the same. They express the same condition in different ways. It is
part of the meaning of (27A) that the sentence refers to the day that includes the
time of utterance, and part of the meaning of (27B) that it refers to the immediately
preceding day. (28B) highlights what is in the bottle, and (28A) what is not. (29A)
is about losing, (29B) about winning. What determines the situation expressed by
a sentence, i.e. its proposition, are the elements of the situation and how they are
interlinked. The situation expressed by (27A) contains the day of the utterance as an
element and specifies it as a Monday. The situation expressed by (27B) is parallel, but
different. More than simply defining truth conditions, a natural language sentence
represents a certain way of describing a situation which then results in certain truth
conditions. Whenever we speak, we make a choice among different ways of expressing
ourselves, different ways of putting things; we are not just encoding the facts we want
to communicate. There is usually more than one way to depict certain facts.
Although less obvious, the analogue holds for words. For example, in German the
big toe is called either großer Zeh (›big toe‹) or dicker Zeh (›thick toe‹) or, by some
people, großer Onkel (›big uncle‹). The Croatian term is nožni palac ›foot thumb‹.
These would all be terms with different descriptive meanings because they describe
what they denote in different ways. More examples of logically equivalent expressions
with different descriptive meanings can be easily found if one compares terms from
different languages which have the same denotation. English has the peculiar term
fountain pen for what in German is called Füllfederhalter (›fill feather holder‹, i.e. a
‘feather holder’ that can be filled) or just Füller (›filler‹, in the meaning of ›something
that one fills‹); in Japanese, the same item is called mannenhitsu, literally ›tenthousand-years brush‹. For a bra, German has the term Büstenhalter, ›bust holder‹;
the French equivalent soutien-gorge literally means ›throat(!) support‹, while Spanish
women wear a ›holder‹ (sujetador) or ›support‹ (sostén) and speakers of the New
Guinea creole language Tok Pisin put on a ›prison of the breasts‹ (kalabus bilong susu).
194
Understanding semantics
Of course, the meaning of mannenhitsu is not literally ›ten-thousand-years
brush‹ and kalabus bilong susu does not literally mean ›prison of the breasts‹.
But a comparison of the examples shows that the respective concepts are based
on different properties of the referents. Different languages may adopt different
naming strategies for the same categories of things. An interesting field is terms for
technical items. The English term power button (e.g. of an amplifier) rests on the
concept ›button‹, which is borrowed from the domain of clothing, and connects it
in an unspecific way with the concept ›power‹, again a metaphor. The equivalent
French term is interrupteur d’alimentation, literally ›interrupter of supply‹; the object
is primarily named after its function of interrupting the current – a somewhat
arbitrary choice, since the power button can also be used for switching the device
on; the second part d’alimentation specifies what is interrupted, namely the
‘alimentation’, a metaphorical term for power supply, originally meaning ›feeding‹,
›nourishing‹. German has yet a different solution for naming that part: Netzschalter,
›net switch-er‹, where Netz is the mains.
Thus, let us fix the following important point:
Logical equivalence is not a sufficient criterion for having the same meaning,
or even the same descriptive meaning.
The converse is, of course, true: since the descriptive meaning determines truth
conditions and denotations, two expressions with the same descriptive meaning
necessarily are logically equivalent. Employing the semiotic triangle in a somewhat
distorted form, Fig. 7.5 displays a configuration of two equivalent expressions with
different descriptive meanings.
Figure 7.5
Logically equivalent expressions with different descriptive meanings
expression1
descriptive
meaning1
expression2
denotation/
truth conditions
descriptive
meaning2
Logically equivalent expressions have the same logical properties: they share the
same denotations or truth conditions. Logically equivalent expressions entail the
same expressions and they are contrary or complementary to the same expressions.
In terms of reference and truth conditions, the use of logically equivalent expressions
Meaning and logic
195
amounts to the same, but the objects and situations referred to may be being described
in conceptually different ways, as is illustrated by the examples in (27) to (29).
7.6.2 The semantic status of logical entailment
The fact that the denotation of a logical subordinate is a subset of the denotation of
the superordinate may be due to a reverse relation between the descriptive meanings
of the two terms. Let us roughly consider the descriptive meaning as a set of
conditions that a potential referent must fulfil. Then a ‘duck’ must fulfil all conditions
a ‘bird’ must fulfil plus those particular conditions that distinguish ducks from
other sorts of birds. The descriptive meaning of duck in this sense contains [all the
conditions which make up] the descriptive meaning of the superordinate term bird.
This is what we intuitively mean when we say that the term duck is ‘more specific’
than the term bird. Generally, AfiB may be due to the fact that the meaning of A fully
contains the meaning of B. This is the case, for example, for the sentences in (47): A
has the same meaning as B except for the addition that the beer is cool.
(47) A
B
There’s cool beer in the fridge.
There’s beer in the fridge.
However, since entailment is only a matter of truth conditions, there need not be such
a close connection between sentences or other expressions related by entailment.
(48) is a simple example of two sentences A and B where A entails B, but the meaning
of B is not contained in the meaning of A:
(48) A
B
Today is Sunday.
Tomorrow is not Friday.
The analogue holds for logical subordination. Consider, for example, the expressions
son of x’s mother-in-law and x’s husband. One’s husband is necessarily a son of one’s
mother-in-law. (The converse is not true.) Hence, x’s husband is a subordinate of son
of x’s mother-in-law. But while the meaning of x’s husband is something like ›man x
is married to‹, the meaning of son of x’s mother-in-law is ›male child of the mother of
the person x is married to‹. The latter contains ›mother-in-law‹ as an element of the
resulting concept, but this element is not part of the concept ›husband‹. Therefore,
the meaning of the superordinate is not part of the meaning of the subordinate.
Some authors use the term hyponymy for logical subordination.10 In this volume,
hyponymy will be reserved for the meaning relation that holds between A and B if
(i) B is a logical subordinate of A plus (ii) the meaning of A is fully contained in
the meaning of B. The notion will be formally introduced in 8.2.1. The point to be
10 E.g. Lyons (1977), Cruse (1986).
196
Understanding semantics
stated here, in analogy to the relationship between logical equivalence and identity
of meaning, is this:
Logical entailment and subordination do not necessarily mean that the
descriptive meaning of one expression is included in the descriptive meaning
of the other.
Analogues hold for the other relations of logical incompatibility and complementarity.
For example, as we could see in connection with (34) and (35), A and B can be logical
contradictories without one being the negation of the other.
7.6.3 Logic and semantics
The discussion has shown that logical properties and relations are only indirectly
related to meaning. Rather, they concern denotations and truth conditions, i.e.
aspects of linguistic expressions which are determined by meaning, more precisely
by descriptive meaning. A logical approach to meaning is therefore limited as follows:
∑ It does not capture those parts of meaning that do not contribute to the
determination of truth and reference. Expressions with the same descriptive
meaning but different social or expressive meanings cannot be distinguished by
logical methods.
∑ It does not capture descriptive meaning itself, but only effects of it.
∑ It does not capture differences between the descriptive meanings of expressions
with identical truth conditions or denotations. In particular, it fails to yield any
insights into the meanings of non-contingent sentences.
If the limitations of the logical approach, and the nature of the results it is able to
produce, are carefully kept in mind, it is, however, a powerful instrument for the
semanticist. Due to the fact that truth conditions and reference are determined by
descriptive meaning, the logical approach can produce the following kinds of results:
∑ If two expressions are not logically equivalent, their meanings are different.
∑ If an expression A does not entail an expression B, the meaning of B is not part
of the meaning of A.
∑ If one of the logical relations holds between two expressions, their descriptive
meanings must be closely related.
The logical approach therefore provides us with simple instruments to test expressions
for their meaning relations with others. This makes logical relations very important
data for semantic analysis. If two expressions are equivalent, or if one entails, or
excludes, the other, this is a fact which semantic analysis has to account for.
Meaning and logic
197
7.7 CLASSICAL LOGIC AND PRESUPPOSITIONS
7.7.1 Presuppositions and the Principle of Polarity
Sentential logic is based on the assumption that language is ruled by the Principle
of Polarity: there are just two possibilities for any sentence in any context – truth
and falsity. Adopting the Principle of Polarity, we can define logical properties,
relations and basic connectives in terms of truth and falsity. All these notions are
straightforward. The result is a clear and rigid – and very simple – basic system for
thinking about the meaning of natural language sentences. However, the Principle of
Polarity is not as innocent as it might appear. Every natural language sentence comes
with presuppositions. Thus, for every natural language sentence, the case that it is
neither true nor false may in fact occur.
This does not mean that the Principle of Polarity is simply wrong: presupposition
failure is not a third possibility on a par with a sentence being true or being false.
Rather it leads to a truth-value gap, and there being a truth-value gap does not
amount to a third truth value. In order to take care of presuppositions, the Principle
of Polarity has to be remedied as follows:
DEFINITION 14 Principle of Polarity (revised)
In a given CoU where its presuppositions are fulfilled, with a given reading, a
declarative sentence is either true or false.
Thus, the Principle of Polarity does hold – but it only holds given that the
presuppositions are fulfilled. Still, ‘there is no third’, in the sense of ‘no third on a par
with truth and falsity’. This is in accordance with the revision of the definition of
truth conditions which we undertook in 4.7.3: presuppositions are not part of the
conditions for being true or false, but preconditions for having a truth value.
In accordance with the Principle of Polarity, the main predication of a sentence is
always presented as a true-or-false matter. Usually, the main predication is expressed
by the VP. Every verb carries presuppositions, namely the selectional restrictions on its
arguments. In addition, change-of-state verbs carry presuppositions concerning the
initial states preceding the event denoted. On top of these verb-specific presuppositions,
there is the general Presupposition of Indivisibility carried by any predication
whatsoever (4.7.2). In fact, this presupposition makes sure that the outcome of a
predication cannot be ‘split’ into truth and falsity. Subject and object NPs, too, often
carry presuppositions, if they are definite or quantificational. This is how every
sentence comes with presuppositions. These presuppositions are not just a nuisance,
but rather a very forceful device for communicating the type of context which serves
as the background for a true-or-false alternative. Compared to the whole world, such
contexts are extremely narrow spaces of possibilities and it is only this heavy cutting
down on possibilities and focusing on very particular aspects of the world that makes
the true-or-false mode of verbal communication possible and meaningful.
198
Understanding semantics
Let me illustrate this with a very simple example. Assume you say ‘I’m tired.’ By
using the 1st person singular pronoun as the subject of the sentence, you cut down
the zillions of possible referents in the universe, and among them the billions of
people on earth, to just one case – yourself. The rest of the universe fades into
oblivion. Then, by using the present tense, you select a very limited time slice of your
existence. The Presupposition of Indivisibility restricts it to a time span where you
are either tired or not tired throughout, thus not more than a couple of hours. Finally,
the predication be tired selects one very specific aspect of your physical condition out
of the unlimited number of aspects concerning your person at that time. This is how
a normal everyday sentence zooms in on some very particular aspect of our world.
The focusing is mainly achieved by presuppositions which drastically narrow down
the context of any utterance to a very specific true-or-false affair. In addition to the
effect of presuppositions, any actual utterance is further massively focused by the
mutual assumptions of background knowledge.
Figure 7.6 is an attempt to illustrate the difference between two basic strategies:
disregarding presuppositions, as in traditional logic, or including them into one’s
semantic considerations. If presuppositions are disregarded, utterances are treated
as conveying an unconditioned polar alternative, represented by the schema V. With
presuppositions, an utterance presents certain things as given and uncontroversial
– the presuppositions – and on this ground presents a polar alternative; this is
symbolized by the Y schema. The Y schema holds independently of the particular
CoU because it is implemented by the mere semantic content of the utterance. When
a sentence is uttered in a concrete CoU, its presuppositions are embedded into the
vast complex of background assumptions with which the interlocutors operate. This
is the third, Y-in-the-swamp picture. Traditional logic treats sentences according
to the V schema, neglecting the actual basis for polarization. A logic adequate for
capturing the logical structure of natural language would actually have to cope with
the Y picture of sentences.
Figure 7.6
Binary, presuppositional and contextual polarity
S
not-S
S
not-S
S
not-S
PRESUPP
PRESUPP
BACKGROUND
Meaning and logic
199
7.7.2 Radical negation
In 4.7 and throughout this chapter, including the definition of negation, it was
assumed that (normal, regular) negation does not affect the presuppositions of a
sentence, and this will be assumed in the rest of the book. There are, however, cases
where one wants to deny a presupposition of an utterance by someone else and uses
negation for that purpose. Imagine the following dialogue:
(49) A:
B:
The Turks are celebrating their victory.
Oh, no. The Turks are NOT celebrating their victory – they didn’t win, you
know.
By repeating A’s statement and negating it, B first complies with A’s perspective which
apparently presupposes that the Turks have won. Knowing better, B contradicts, by
negating the sentence. However, this is not a standard negation. Had B simply replied
(50), with normal intonation, B would have committed to the same presuppositions
as A did and denied that the Turks are celebrating.
(50) The Turks are not celebrating their victory.
Instead, B puts a heavy accent on the negation and adds a rectification of A’s
presupposed assumption that the Turks won. This reply amounts to a refutation of
one of the presuppositions of A’s utterance. It may involve refusing that the Turks are
celebrating at all, but it need not; it may also be used to convey the notion that the
Turks are actually celebrating something else.
This kind of negation is called radical negation. Radical negation always comes
with an extra emphasis on the negation word. It can be used to deny that one or the
other presupposition is fulfilled. In order to know which aspects of the sentence are
subject to negation, an explicit rectification is necessary, like the second sentence in
(49B). Without these heavy cues, negation will be taken not to affect presuppositions.
Figure 7.7 illustrates the polarity schema for radical negation. The positive alternative
is formed by the original predication plus those presuppositions claimed to fail; the
Figure 7.7
Radical negation polarity
S
not-S
PRES
200
Understanding semantics
negative alternative negates this package. However, the radical negation also rests
on some presuppositions. Trying to negate absolutely all presuppositions would not
result in a meaningful utterance. Radical negation needs a rectification in order to be
interpretable, and any rectification sentence inevitably carries some presuppositions.
7.7.3 The projection problem of presuppositions
Integrating presuppositions complicates logic considerably. One fundamental
problem is the determination of the presuppositions of complex sentences. The
problem is known as the projection problem of presuppositions. If A carries the
presuppositions P and B carries the presuppositions Q, what are the presuppositions
of (AŸB), (A⁄B) or (A→B) – let alone all the sorts of combinations and embeddings
of sentences that are available in natural language? If there is presupposition failure
for either A and B, will the presuppositions fail for the whole? Let us consider two
examples concerning conjunction:
(51) Greg started drinking and Sue stopped smoking.
If the conjunction is interpreted as the logical Ÿ, the two conjuncts Greg started
drinking and Sue stopped smoking are taken as two independent sentences with
the same CoU as background. (51) can only be true if for each conjunct its
presuppositions are fulfilled in the given common CoU; (51) would be false if at least
one of the conjuncts is false, which again requires its presuppositions to be fulfilled.
Hence, the presuppositions of both conjuncts are presuppositions of the whole. Now
consider (52):
(52) Helen married Greg and got divorced only three years later.
The sentence is uttered incrementally: the first conjunct first and then the second
conjunct. The first conjunct carries the presupposition that before the time
considered Helen had not been married. Uttering Helen married Greg changes the
CoU: it now contains the fact that Helen married Greg. The second conjunct shifts
the time three years ahead and it finds its presupposition – that Helen had been
married – in the context provided by the first conjunct. What this shows is that we
cannot always assume that the presuppositions of a conjunction as a whole are just
the presuppositions of the conjuncts taken together. Sentence (52) as a whole does
not presuppose that Helen had been married to Greg.
Similar problems arise with logical disjunction and subjunction and with natural
language or and if. The problem of presupposition projection is very complex and
there is a lot of literature about it. For the purpose of this book, this much may suffice
on this topic.
Meaning and logic
201
EXERCISES
1. For the following sentences S, which one is the proper negation, A or B? Check the
truth conditions of A and B: which one is necessarily true if the positive sentence
is false?
a. S It sometimes rains here.
A It never rains here.
B It doesn’t always rain here.
b. S She is cold and sick.
A She is neither cold nor sick.
B She either isn’t cold or isn’t sick.
c. S Everybody helped me.
A Nobody helped me.
B Not everybody helped me.
d. S She’s already here.
A She’s not yet here.
B She isn’t here anymore.
2. Which of the following statements are true, which ones are false?
a. If A is contingent, then not-A is, too.
b. If A is not logically true, then not-A is.
c. If not-A entails B, then not-B also entails A.
d. Not-A never entails A.
e. A cannot entail both B and not-B.
3. Apply the formal definition of entailment to show that contradiction entails
logical contrariety.
4. Check the following pairs of sentences A and B: which truth-value combinations
are possible? Do they represent cases of logical entailment, equivalence,
contrariety, contradiction or none of these relations? In case of entailment, check
both directions.
a. A The door is open.
B The door has been opened.
b. A The room is not large.
B The room is small.
c. A She never empties the ashtray. B She often empties the ashtray.
d. A She’s sometimes late.
B Sometimes she is not late.
e. A By midnight, half of the
B At midnight, only half of the
guests had left.
guests were left.
5. What is the logical relationship between A and B, in terms of entailment(s), if
they cannot both be false?
6. Show that two logically false sentences are both equivalent and contraries.
7. Which logical relation applies to the following pairs of words?
a. book, textbook
b. textbook, dictionary
c. free, occupied
d. young, old
e. receive, send
f. left, right
8. Which dimensions of meaning are not captured by the logical method and why
are they not?
202
Understanding semantics
9. Discuss the limits of the logical method for the investigation of descriptive
meaning.
10. Discuss the ways in which the logical method is useful for the investigation of
word meaning.
FURTHER READING
Cruse (1986, ch. 4) for logical relations between words. Partee et al. (1993, ch.6) for
definitions of the logical properties and relations with respect to logic.
8
Meaning relations
It is very difficult to describe the meanings of lexemes or sentences explicitly. In
fact, semanticists do not even agree as to what kind of entities meanings are. The
conceptualist point of view taken in this book – which regards meanings as concepts
– is the prevalent view, but by no means uncontroversial. It is, however, relatively easy
to establish meaning RELATIONS between sentences or expressions. One influential
traditional position, the structuralist position, holds that the description of meaning
is exhausted by the description of meaning relations (chapter 9). We do not share this
view. This notwithstanding, meaning relations are most important semantic data.
The logical relations that were introduced in the last chapter are defined in
terms of relations between the denotations of expressions, or truth conditions of
sentences; for example, two content words are logically equivalent iff they have
identical denotations. Consequently they are relations concerning the extra-linguistic
correlates of expressions (i.e. entities at the right bottom corner of the semiotic
triangle). Meaning relations, by contrast, are defined in terms of conceptual relations,
between meanings themselves (i.e. entities at the left bottom corner of the semiotic
triangle). For example, two expressions are synonymous iff they have the same
meaning. While logical relations may be indicative of meaning relations of some
kind, it is important to realize that logical relations and meaning relations are defined
in a fundamentally different way.
8.1 SYNONYMY
The meaning relation simplest to define is (real, not approximate) synonymy:
DEFINITION 1 Synonymy
Two expressions are synonymous iff they have the same meaning.
Synonymy in the strict sense, also called total synonymy, includes all meaning
variants for two, possibly polysemous lexemes and it includes all meaning dimensions,
including descriptive, social and expressive meaning. Two polysemous expressions
are synonymous iff they have the same set of meaning variants; and expressions with
204
Understanding semantics
descriptive meaning and other meaning dimensions are synonymous iff they have
the same descriptive as well as the same non-descriptive meaning. Examples are
very hard to find; good candidates are pairs of expressions with the same meaning
but different regional preference, such as the two German expressions for Saturday
Samstag and Sonnabend; the latter is preferred in Northern Germany, but both belong
to common usage. Trivial cases of synonyms are pairs of abbreviations and their full
forms, e.g. the English words perambulator and its shortened form pram, compact
disc and CD or the linguistic terms NP and noun phrase. These cases are trivial in that
the two forms can be considered alternative variants of the same lexemes.
While total synonymy is rare or trivial, there are many cases of partial synonymy.
Two lexemes may have one or more meaning variants in common. Examples would
be English hide and conceal, which are very similar in meaning but, at a closer look,
1
not identical. A large group of cases is constituted by pairs such as mail and e-mail,
disc and compact disc: a compound lexeme is synonymous with its second part (its
so-called ‘head’, see 8.2.1) when taken in a specific reading. The specific reading is
a differentiation (recall 3.4.3.1) of the more general reading in which, for example,
mail may also refer to a letter on paper or a parcel or to the institution of mail. Other
instances of partial synonymy are English spectacles and glasses which may both
denote the same sort of objects that persons wear on their noses to look through;
but glasses may also just be the plural of glass in one of its other meaning variants.
In everyday use, the term ‘synonymy’ is applied more broadly to pairs of
expressions which can mean the same in certain contexts. This non-technical notion
of synonymy is not based on the real meanings of the expressions, but just on their
use, i.e. on their denotations. In our terminology, such ‘synonyms’ are equivalents
in certain contexts rather than having a similar conceptual meaning. If candidates
for synonymy are tested in various contexts, they will most probably turn out to be
matching in some, but not in others. For example, it is easy to perceive of contexts
where the word glasses can be exchanged for spectacles; however, of course, there are
also cases where this is impossible (e.g. she drank two glasses … but not: … two
spectacles of orange juice).
Having identical meanings, synonymous expressions are logically equivalent, since
meaning determines denotation and truth conditions. The converse is not true, as
we saw in 7.6.1: logically equivalent expressions may have different meanings. A
remarkable group of such cases is constituted by pairs consisting of a new expression
introduced for the sake of political correctness and the expression it is meant to
replace. Consider, for example, the expression blind and its replacement visually
impaired, introduced in order to avoid discrimination. The new ‘politically correct’
expression is (meant to be) denotationally equivalent, but it is chosen because it does
not possess the same meaning. Language provides many different ways of denoting
the same thing, and the alternative expressions may well differ in meaning.
1 See Cruse (1986: §12.1) for a discussion of this and similar cases.
Meaning relations
205
8.2 HYPONYMY
8.2.1 Hyponymy
The meaning relation of hyponymy can be defined as follows: 2
DEFINITION 2 Hyponymy
An expression A is a hyponym of an expression B – and B a hyperonym of
A – iff
(i) the meaning of B is a proper part of the meaning of A and
2
(ii) A is a logical subordinate of B, i.e. A entails B.
The meaning of A must contain further specifications in addition to the meaning
of B, rendering the meaning of A, the hyponym, more specific than the meaning of
B. Hyponymy is a relation between lexemes – or more complex expressions – that
results from a relation between their MEANINGS and gives rise to a relation between
their DENOTATIONS (cf. condition (ii)). These relations are shown in Fig. 8.1 which
integrates the semiotic triangles for a hyponym and its hyperonym.
Figure 8.1
Hyponymy
duck
hyponym
meaning of hyponym
meaning of
hyperonym
bird
hyperonym
denotation
of hyperonym
denotation
of hyponym
In the semantic literature, hyponymy is often not distinguished from logical
subordination; condition (ii), alone, is regarded sufficient. This is the result of a
tendency to restrict semantics to the investigation of descriptive meaning and this, in
2 Recall 7.5.2 for the definition of subordination. Strictly speaking, the definition has to be related
to single meaning variants of expressions. For example, the word compound is a hyponym of
expression (in one of its readings) if it is taken in its linguistic sense, but a hyponym of substance
(again, only in one of its meanings), if taken in the sense of chemical compound. Thus, the
omnipresence of polysemy complicates the matter of meaning relations considerably. For the sake
of simplicity, polysemy will be ignored throughout this chapter unless indicated otherwise.
206
Understanding semantics
turn, to the logical aspects of meaning. In this book, however, the distinction between
logical and conceptual notions is considered a point central for the understanding of
semantics and linguistic meaning.
Examples of hyponymy are countless. For hyponymy among lexemes, consider
pairs of subordinates and superordinates in hierarchies like those in Figs 7.2 and
7.3. Here we are entitled to assume the crucial meaning relation (i) in addition to the
logical relation (ii) because, for example, it is part of the meaning of verb that verbs
are words; thus the meaning of word must be contained in the meaning of verb. Other
examples would be beer (hyponym) and beverage (hyperonym), or car and vehicle,
respectively.
A very common type of complex expressions consists of one expression, the ‘head’,
combined with another expression, called the ‘modifier’. Modification occurs in
syntax, e.g. adding an adjective to a noun, and in word formation. The modifier adds
a specification to the meaning of the head. Examples would be compounds such as
book + shop (recall 1.2.5) or syntactic combinations such as red cap. If such complex
expressions are interpreted compositionally, they are generally hyponymous to their
mere heads. Condition (i) is fulfilled since the meaning of the modifier adds some
specification to the meaning of the head. As to condition (ii), usually the denotation
of the whole will be part of the denotation of the head: a book shop is a shop; a red
cap is a cap. We called compounds of this kind ‘regular compounds’ in 1.2.5; they will
be treated in more detail in the next subsection.
In connection with synonymy, cases were mentioned where simple lexemes are, in
a special reading, synonymous with more specific compounds (e.g. disc and compact
disc). Since the special reading is a conceptual differentiation of the more general
reading, it adds content to the latter. In general, if the meaning of an expression A
is a differentiation of the meaning of B, then A is a hyponym of B. A similar case
concerns certain alternations: one use of a verb may be hyponymous with another:
if I eat an orange, I eat, if I open the door, the door opens; thus, transitive verbs like
open are hyponymous with their intransitive antipassive variant and with verbs of
change like open, the causative variant is hyponymous with the inchoative variant
(recall 6.1.2, 6.1.3).
As was pointed out in 7.6.2, not all subordinates are hyponyms; two expressions
may fulfil condition (ii) without fulfilling (i), for example mother-in law’s son and
husband. There are also pairs of expressions which fulfil condition (i), but not (ii). In
barfly the head is to be taken metaphorically, whence a barfly is not a fly. Similarly,
Blue Helmet metonymically refers to certain persons wearing blue helmets; thus a
Blue Helmet is not a helmet. In yet a different type of compounds, represented e.g. by
blue-green, the meaning of the second part (or the first part, likewise) is clearly part
of the meaning of the whole, but the whole is not subordinate to one of its parts: bluegreen is not a special shade of green, but a colour between blue and green, which is
3
neither blue nor green. Apart from such compounds, there are also combinations of
3 In a sense, compounds such as blue-green or Blue Helmet are regular, too. They are just less
frequent, and their formation and interpretation obey different rules than the type represented by
bookshop.
Meaning relations
207
adjectives and nouns that are not hyponymous with the noun itself: an ‘alleged thief ’
4
is not necessarily a thief, a ‘former butcher’ usually not a ‘butcher’.
8.2.2 Regular compounds
A regular compound has two parts, the modifier and the head; the modifier adds
a specification to the meaning of the head noun. For example, in apple juice, the
modifier apple adds the specification ›produced from apples‹ to the concept ›juice‹.
Similarly, park bench means ›bench in a park‹, headache pill ›pill for curing a
headache‹, company flat ›flat belonging to a company‹, flower shop ›shop selling
flowers‹, and so on. The interpretation of regular compounds ensures that the
compound is a hyponym of its head.
The meaning of the modifier is also bound into the meaning of the compound, but
not in the same way as the meaning of the head. Therefore, the resulting meaning
relation between the compound and the modifier is not hyponymy (apple juice is not
an apple, a flower shop not a flower, etc.). Rather, the meaning relation depends on the
way in which the meaning of the modifier is built into the meaning of the whole. For
example, the meaning relation between flower and flower shop is as follows: referents
of the first are sold in referents of the second. Special though this meaning relation
is, it provides a basis for a general pattern of composition (cf. analogous compounds
such as stationery shop, ticket counter, hot dog stand). The other examples mentioned,
park bench, company flat and apple juice represent different patterns. More will be
said about the relations between the modifier concept and the head concept in 12.3.4.
Figure 8.2 illustrates the meaning relations within regular compounds in general,
using apple juice as an example.
Figure 8.2
Meaning relations between a regular compound and its components
apple
juice
MODIFIER
HEAD
(various)
hyponymy
apple juice
COMPOUND
4 Recall 5.4.2 on non-predicational adjectives
208
Understanding semantics
8.3 OPPOSITIONS
8.3.1 Examples
If people are asked to specify the opposites of words like old, buy and aunt, they will
answer spontaneously. The answers will vary, however, because words may have
different kinds of opposites. The intuitive notion of opposites covers a variety of
meaning relations. In this section, we will take a closer look at them. The technical
term for such relations is oppositions.
This is what people answer if they are asked to give the opposite of old, buy and
aunt:
(1)
old:
(i) new
(ii) young
buy:
(i) sell
(ii) steal, borrow, rent
aunt:
(i) uncle
(ii) nephew, niece
The two opposites of old reveal that old is polysemous. Old as an opposite of young
relates to age, primarily of living creatures. The opposites old and young can also
be used for inanimate entities that are metaphorically conceived as subject to
developments such as growth, ripening and ageing (cf. a young nation, an old wine, a
young language). In such cases, young and old denote early and late stages of internal
development. However, old can also be used for cars, books, buildings, words. Old
then means that the respective object has been in use for a long time (probably
resulting in wear and tear, in becoming out of date, etc.); new is the opposite of this
meaning variant. As to their relation to old, both opposites are of the same type:
old denotes one extreme on a scale, while young and new denote the other extreme.
The difference between new and young and the two corresponding meanings of old
concerns the underlying scale.
What counts as the opposite of buy is less clear. If we take a look at the second set
of verbs (steal, borrow, rent), we realize that all these verbs denote alternative ways of
acquiring things. Borrow and rent contrast with buy in that the transfer of possession
is temporary. This meaning element is absent in the cases of buy and steal. Buy and
rent involve a compensation with money, steal and borrow do not. The three opposites
under (ii) are of the same kind: they are alternatives in certain respects, but they do
not denote an opposite EXTREME on a scale.
The opposite sell, however, is completely different. There are two ways of viewing
the relation between buy and sell. First, any event of buying is at the same time an
event of selling: if x buys something from y, then y sells that thing to x. To conceive
of an event as an instance of selling rather than buying is just a converse way of
conceptualizing this transfer: an object z changes from the possession of y into the
possession of x, while x pays y an equivalent amount of money in exchange for z. Buy
describes the event as an action on the part of the new possessor x, sell describes it as
an action on the part of the original possessor y. Buy and sell express the same with
reversed roles. Second, we can compare x buying something with x selling it. Then,
Meaning relations
209
buy and sell are opposed in the sense that buy means ‘moving into’ the possession
of the object and sell ‘moving out’ of it. Viewed in this way, the two verbs express
opposite directions of transfer.
Whenever we think of oppositions in the intuitive sense, we try to find something
common to the meanings of the two expressions and a crucial point in which their
meanings differ. In the case of the opposition aunt/uncle, the point in common is
the kind of kinship relation: being a sibling of a parent, the spouse of a sibling of
a parent or something similar, a relative one generation back but not in direct line.
What differs is the sex of the relative. If we consider the pairs aunt/nephew and aunt/
niece, we encounter the same kind of relation as between buy and sell under the first,
reversed-roles, perspective: if x is an aunt of y, then y is a nephew or niece of x. The
kinship link between x and y is the same, but aunt refers to one side of the link and
nephew/niece to the other. The relation is not strictly reversed, since the terms aunt,
nephew and niece also contain a specification of sex. We cannot say that if x is an aunt
of y then y is a nephew of x, because y could also be a niece of x. Nor can we say that
if x is a nephew of y then y is an aunt of x, because y could also be an uncle.
8.3.2 Antonyms
DEFINITION 3 Antonymy
Two expressions are antonyms iff they express two opposite extremes out of a
range of possibilities.
The prototypical examples of antonymous expressions are pairs of adjectives such
as old/young, old/new, big/small, thick/thin, good/bad, light/dark, difficult/easy.
Their meanings can be illustrated by means of a scale of age, size, diameter, quality,
brightness, difficulty, etc. which is open on both sides. One might object that scales
such as the scale of size are not open at the lower end, i.e. that zero size delimits the
scale at its bottom. However, in our imagination, things can be smaller and smaller;
for every small thing there can be something smaller. There is no smallest size and
hence no lower bound of the scale of size. The scale of size asymptotically approaches
zero size, but does not include it. (Note that we cannot say something is ‘small’ if it
has no size at all.) One of the adjectives denotes the upper section, the other one
the lower section of the scale. There may be a neutral middle interval where neither
expression applies. Every value on the scale that lies above the neutral interval counts
as, for example, ‘big’; every value below the neutral interval counts as the opposite,
e.g. ‘small’ (cf. Fig. 8.3).
Antonyms are logically incompatible, but not complementary (7.5). If something
is big, then it is not small, and vice versa. But if something is not big, it need not
be small, and vice versa – it can be in the neutral range of the scale, i.e. neither big
nor small. Words for the neutral case are rare and rather artificial recent inventions
(middle-aged, medium-sized). Normally, the neutral case can only be expressed by
210
Understanding semantics
Figure 8.3
Antonymy of big and small
← small
SMALLER
big →
neutral
BIGGER
such cumbersome expressions as neither difficult nor easy. Antonymous adjectives
of the type illustrated in Fig. 8.3 are called scalar adjectives. They are ‘gradable’ and
therefore allow for the full range of adjectival forms and constructions: comparative
(bigger than), superlative (biggest), equative (as big as), or modifications such as very
big, big enough and too big.
In very many cases, the antonym of an adjective is formed by prefixing un- or its
Latin-origin equivalent in-/im-/ir-/il-: pleasant/unpleasant, likely/unlikely, adequate/
inadequate, probable/improbable, rational/irrational, logical/illogical. In each case,
the unprefixed adjective is felt to be positive, not in the sense of an expressive
meaning component, but as intuitively denoting the section at the upper end of the
scale, while the prefixed opposite is the negative counterpart at the lower end of the
scale. We shall see below that pairs of this form (X vs un-X, etc.) are not necessarily
antonyms. The pattern also occurs with pairs of expression that form different types
of opposition.
Antonymy is not restricted to adjectives. There are antonymous pairs of nouns
such as war/peace, love/hate and some antonymous pairs of verbs: love/hate or
encourage/discourage. The pair all/no(thing) is antonymous, as are pairs of adverbs
such as always/never, often/seldom, everywhere/nowhere.
8.3.3 Directional opposites
Pairs such as in front of/behind, left/right, above/below have much in common with
antonyms. For each such pair there is a point of reference from which one looks in
opposite directions on a certain axis. Imagine yourself standing in a normal upright
position, your head not turned or bent. Then the direction in which you look is what
is denoted by in front of (you) and behind (you) denotes the opposite direction. The
two directions are distinguished in several ways: the body has a front, which includes
the face, the breast, etc., and it has a back; when one walks in the usual way, one walks
into the direction that is in front of one. Thus, the body defines an axis in space, the
front–back axis, or primary horizontal axis. Another axis defined by the body is
the vertical head–feet axis. The extension of the direction from the body centre to
the head defines the direction above; the opposite direction is denoted by below.
Finally, the body with its two eyes, two ears, two arms, two legs and its general lateral
symmetry defines another horizontal axis, the left–right axis. The three axes and
how they relate to the corresponding expressions are depicted in Fig. 8.4.
Meaning relations
211
Figure 8.4
Directions
above
behind
left of
right of
in front of
below
DEFINITION 4 Directional opposition
Two expressions are directional opposites iff they express opposite cases with
respect to a common axis.
The type of opposition represented by in front of/behind is called directional
opposition. Directional opposites are related to opposite directions on a common
axis. Further examples that involve the vertical axis are: top/bottom, high/low, up/
down, upstairs/downstairs, uphill/downhill, rise/fall, ascend/descend and many more.
Examples related to the primary horizontal axis are forwards/backwards, advance/
retreat.
A similar axis is the time axis. We talk about things happening ‘before’ vs ‘after’ a
certain time, or ‘later’ vs ‘earlier’. Thus, pairs such as before/after, past/future, since/
until, yesterday/tomorrow, last/next, precede/follow are directional opposites related
to the time axis. A further case is provided by past tense and future tense (recall
6.4). Also related to time are pairs of accomplishment verbs or verbs of change with
opposite resultant conditions (recall 6.2), such as tie/untie, pack/unpack, wrap/
unwrap, dress/undress, put on/put off, get on/get off, arrive/depart, switch on/switch
off, embark/disembark, charge/discharge, enter/leave, begin/stop, start/finish, fall
asleep/wake up, appear/disappear, open/close and many more. One member denotes
the coming about, or bringing about, of a certain state, while the other member
denotes the reversed process or action. Among the examples in (1), the pair buy/sell
in its second interpretation is of this type.
212
Understanding semantics
8.3.4 Complementaries
The type of opposition represented by aunt/uncle, buy/rent or buy/steal is known as
complementary opposition.
DEFINITION 5 Complementary opposition
Two expressions are complementaries iff they express an either-or alternative
in some domain.
Complementary opposites are logically complementary (7.5.4): the negation of one
term is equivalent to the other term, e.g. not even (of numbers) means the same as
odd. Each expression denotes one out of the only two possibilities in some domain
of cases. In this domain, complementary opposites represent an either-or alternative.
Complementary adjectives are not gradable; they do not permit the comparative,
5
superlative, equative or modification with very, etc. Examples of complementaries are
married/unmarried, female/male (as a specification of sex), even/odd (of numbers),
possible/impossible, free/occupied. As some of the examples show, prefixation with
un- or in- is also used for the formation of complementary opposites.
Complementarity more typically occurs with nouns, e.g. pairs of terms for persons
of opposite sex, or pairs such as member/non-member, official/non-official. The
meanings of two complementaries are identical except for one crucial feature in
which they differ (cf. the examples aunt/uncle or buy/rent). This feature has to be a
binary, yes-or-no, feature that leaves just two possibilities, e.g. female or male sex.
8.3.5 Heteronyms
DEFINITION 6 Heteronymy
Expressions are heteronyms iff they denote alternatives in some domain of
more than two possibilities.
Heteronymy is a matter of more than two expressions. A typical example is the set of
terms for the days of the week, the set of basic colour terms (for more on colour terms,
see 10.4) or terms for kinds of animals, plants, vehicles, etc. A set of heteronymous
terms jointly covers a wider range of possibilities. Logically, heteronyms are mutual
contraries: if x is Sunday, it cannot be Monday, Tuesday, etc.; if x is a dog, it cannot be
5 To be sure, adjectives such as married actually can be used in the manner of a grading adjective.
Someone might choose to say, e.g., Mary is more married than John is. In such a case, however,
the adjective married is coerced by the comparative form into a gradable property that is
metonymically connected to marriage, for example ‘feeling married’ or ‘acting like being married’.
This contextual meaning shift is triggered by the comparative.
Meaning relations
213
a duck, a cow or any other species. But unlike antonymy, heteronymy is not related to
scales; heteronyms are not opposite extremes, but just members of a set of different
expressions which have a common hyperonym.
There are large fields of heteronymous terms, such as the sets of co-hyponyms
in larger taxonomies (see 8.4.3): terms for plants, flowers, animals, birds, breeds of
dogs, kinds of clothing, drinking vessels, food, vehicles, musical instruments, etc.
Sets of more than two co-hyponyms occurring at the same level in a lexicalized
hierarchy always form a set of heteronyms. Apart from nouns, there are many fields
of heteronymous verbs, such as the different verbs of motion (walk, run, fly, swim),
and verbs denoting human activities such as eat, work, sleep, dance, etc. Further
examples of heteronymy are colour terms or number words.
8.3.6 Converses
The pairs buy/sell and aunt/niece-nephew represent what are called converses. This
kind of relation is restricted to predicate expressions with two or more arguments.
Such terms express a relation in the widest sense, e.g. an action involving two
participants, a comparison, or a kinship relation. Converses are defined as follows:
DEFINITION 7 Converses
Two expressions are converses of each other iff they express the same relation
between two entities, but with reversed roles.
Examples are above/below, before/after, borrow/lend, parent/child, as well as some of
the technical terms introduced here: entail/follow from, hyponym/hyperonym. These
pairs of expression give rise to the following logical equivalences:
(2)
above/below :
x is above y
y is below x
before/after :
x is before y
y is after x
borrow/lend :
x borrows z from y
y lends z to x
parent/child :
x is a parent of y
y is a child of x
entail/follow from :
x entails y
y follows from x
hyponym/hyperonym :
x is a hyponym of y
y is a hyperonym of x
Converses differ from other types of opposites in that they do not correspond to a
uniform logical relation. The pairs above/below, before/after, borrow/lend, hyponym/
hyperonym are contraries (for example, x is above y logically excludes x is below y).
Parent and child, too, are contraries. Entail and follow from are logically independent
because if A entails B it may or may not be the case that A also follows from B (i.e.
B also entails A, recall 7.3.1). Some expressions could even be considered their own
214
Understanding semantics
converses, namely all terms which express symmetric relations (different from, sibling
of, married to) or events that involve two arguments in the same role (meet, fight,
agree, recall the verbs of social interaction mentioned in 6.1.3): x is different from y/a
sibling of y/married to y means the same as y is different from x/a sibling of x/married
to x; similarly x meets/fights/agrees with y has the same meaning as y meets/fights/
agrees with x. Such expressions are called ‘reciprocal’.
One major group of converses is provided by the comparative forms of antonymous
adjectives:
(3)
thicker/thinner:
x is thicker than y
For transitive verbs, passive constructions with explicit
general means of conversion:
(4)
watch/be watched by:
x watches y
y is thinner than x
AGENT
provide another
y is watched by x
Related to the passive conversion is a productive pattern for pairs of converse nouns
in English: noun derivations from verbs with the suffixes -er and -ee. For example, x
is an employee of y means the same as y is an employer of x.
As some of the examples may have shown, one and the same pair of expressions
can constitute opposites of more than one type. Buy/sell and also pairs of spatial
prepositions and adverbs such as above/below can be considered converses as well
as directional opposites.
Table 8.1 shows a survey of the meaning relations treated here.
Table 8.1
Types of oppositions
Examples
Type
Characterization
Logical relation
big/small
war/peace
to love/to hate
antonyms
opposite extremes
on a scale
contraries
above/below
before/after
lock/unlock
directional
opposites
opposite directions
on an axis
contraries
even/odd
girl/boy
voter/non-voter
complementaries
either-or alternatives
within a given domain
complementaries
Monday/Tuesday/…
red/green/blue/…
heteronyms
more than two alternatives
within a given domain
contraries
buy/sell
wife/husband
bigger/smaller
employer/employee
converses
the same with
reversed roles
(relations only)
(various logical
relations)
Meaning relations
215
8.4 LEXICAL FIELDS
8.4.1 The notion of a lexical field
Most lexical items form groups with other lexemes. Antonyms and complementaries
belong together as opposites, as do pairs of words such as father and mother or adult
and child, or sets of words like the terms for the days of the week, for colours, for
numbers, for pieces of furniture or other kinds of thing within one superordinate
category.
Semantic theories of different orientations have tried to capture this phenomenon
by the notion of a lexical field. The literature offers very different definitions. Here,
the following informal characterization will be adopted.
DEFINITION 8 Lexical field
A lexical field is a group of lexemes that fulfils the following conditions:
∑ the lexemes are of the same word class
∑ their meanings have something in common
∑ they are interrelated by precisely definable meaning relations
∑ the group is complete in terms of the relevant meaning relations
Lexical fields are notated like sets by including a list of lexemes in set brackets { … }.
For example, the field of the terms for the days of the week would be written {Sunday,
Monday, …, Saturday}.
The meaning relations that constitute a lexical field may be rather particular (e.g.
sex and age oppositions). The six prepositional expressions in Fig. 8.4, above, below,
in front of, behind, right of, left of, form a field of three opposite pairs relating to three
axes in space that are orthogonal to each other. The names for the seven days of the
week are not merely heteronyms but are in addition interrelated by a cyclic order
Sunday Monday Tuesday Wednesday Thursday Friday Saturday Sunday … The relation of one name to the others in terms of this cycle is part of
its meaning. This is the reason why sentences such as (5) are logically true.
(5)
If today is Monday, tomorrow is Tuesday.
A similar relation of order interrelates the numerals. In addition, the field of numerals
is structured by specific meaning relations that capture the arithmetic properties of
numbers; for example the relation that thirty denotes a number that is three times
the number denoted by ten.
8.4.2 Small fields
Some fields are quite small. For example, each pair of antonyms, such as thick and
thin, forms a lexical field of two members. The meanings of the two antonyms have
216
Understanding semantics
in common that both relate to an end section of the same scale, they are related by
antonymy, and the group is complete because there are no other adjectives relating to
the same scale. Polysemous words with two different antonyms, such as old with its
antonyms new and young, belong to two different fields, {old1, young} and {old2, new}.
Strictly speaking, it is not words that belong to lexical fields but words in certain
meaning variants.
Complementary expressions such as girl and boy also form a field of two, viz. the
field of terms for children specified for sex. If we add the sex-neutral expression child
(in its non-relational meaning ›non-adult‹, not in its relational meaning ›offspring
of‹) to this field, we obtain the field of general terms for children. This field is
structured by two different meaning relations: hyponymy between child and boy/girl
and opposition in terms of sex between girl and boy. There are other lexical fields of
three elements with the same structure, e.g. adult/woman/man or horse/mare/stallion.
We can unite the field of general terms for children with the field for general terms for
adults. The resulting six expressions form a field of six words linked by the additional
meaning relation of opposition in terms of age that relates the members of the
pairs child/adult, girl/woman and boy/man. The result is shown in Fig. 8.5. Different
meaning relations are symbolized by different kinds of arrow.
Figure 8.5
The lexical field of general person terms
adult
woman
man
child
girl
boy
8.4.3 Taxonomies
Terms for animals, plants, food or artefacts such as furniture, vehicles, clothes,
musical instruments, etc. form lexical fields of considerable size. Their underlying
structure is a hierarchy with two or more levels: a topmost hyperonym like vehicle,
a level of general terms such as car, bicycle, boat, aeroplane, and further levels of
Meaning relations
217
more specific heteronymous terms for kinds of cars, bicycles, boats, planes, etc. Such
systems represent a special type of hierarchy called taxonomy:
DEFINITION 9 Taxonomy
A set of expressions is a taxonomy iff:
(i) they form a conceptual hierarchy in terms of hyponymy
(ii) hyponyms denote sub-kinds of what their hyperonyms denote
The hierarchy of word class terms in Fig. 7.2 is a taxonomy: nouns, verbs and
adjectives are kinds of words, count nouns and mass nouns are kinds of noun,
etc. The hierarchy of the three sibling terms sibling, sister and brother in Fig. 7.3,
however, is not a taxonomy: brothers and sisters are not kinds of siblings. Referents
of co-hyponyms in taxonomies differ in many properties. Referents of co-hyponyms
of small, often binary, hierarchies such as sibling/sister/brother differ only in one
property, e.g. in sex. Figure 8.6 contains a section of the animal taxonomy. We will
come back to taxonomies and their properties in 11.3.
Figure 8.6
An incomplete taxonomy of animal terms
animal
horse
cow
husky
pig
dog
terrier basset
cat
elephant tiger crocodile
collie
8.4.4 Meronymies
Many objects in the world are conceived as a whole consisting of different parts. And
correspondingly, our concepts for complex objects contain specifications of these
parts as elements. One of the best examples of a complex object is the human body
with its various parts, their subparts, and so on. The respective terms form what
is called a mereology, a system of terms for a whole and its parts, the parts of the
parts, and so on. Parts in a mereology are not only pieces of the body, like a butcher
might cut out of the body of a slaughtered animal. They also exhibit a certain degree
of autonomy; they serve a distinct function – it may be possible to move them
independently of other parts; in the case of artefacts, they may be replaceable; and
the parts may in turn be composed of other parts. For example, the head is the part
218
Understanding semantics
6
of the body that carries the most important sense organs, viz. eyes, ears, nose and
tongue; it contains the brain, etc. The face forms the front side of the head; it contains
the eyes, the nose and the mouth; it serves facial expression. Within the face, the
mouth is used for eating, drinking, speaking, breathing, biting, kissing, smiling, etc.;
the tongue serves for articulation, tasting, licking, swallowing and other purposes.
The items that form a mereology are necessarily relational nouns. In many
languages, body part terms are used with inalienable possessive constructions
(recall 5.4.1.2). A small section of the system of body part terms in English is given
in Fig. 8.7.
Figure 8.7
An incomplete meronymy of body part terms
body
neck
head
ear
belly
face
eye
mouth
lip
trunk
leg
arm
breast
hand
finger
forearm
palm
tongue
A system of this type is not to be confused with a hierarchy based on hyponymy.
The vertical lines stand for the part–whole relation between single potential referents
of the terms. For example, a potential referent of face is always part of a potential
referent of head. By contrast, the vertical lines in a logical hierarchy represent a
part–whole relation between the total denotations of the terms involved: the set of
all ducks is a subset of the set of all birds, but, of course, a single duck is not part of a
single bird. Hyponymy is based on a third kind of part–whole relation: the meaning
of the ‘higher’ (!) term, the hyperonym, is part of the meaning of the lower term, the
hyponym.
6 The word body is ambiguous because it can denote either the trunk or the trunk plus the head and
the limbs; the term body is used in the latter meaning here.
Meaning relations
219
The technical term for the constituting meaning relation is meronymy (from
ancient Greek meron ‘part’; the term partonomy is also used). A system based on
meronymies is called a mereological system, or mereology.
DEFINITION 10 Mereology, Meronymy
A set of expressions forms a mereology iff they form a hierarchy in terms of
holonyms and meronyms, where A is a meronym of B, and B a holonym of A,
iff A denotes constitutive parts of the kind of things that B denotes.
By a ‘constitutive part’ we mean a part that necessarily belongs to the whole and
contributes to making the whole what it is. Thus, the meaning of the meronym is
a concept that describes its potential referents as belonging to this kind of whole;
conversely, the meaning of the holonym describes its potential referents as having
this kind of part(s). For example, the concept ›hand‹ describes hands as having
fingers, and the concept ›finger‹ describes fingers as being certain parts of hands.
Meronymies involving nouns can be spelt out with the help of sentences of the
form ‘an A has a B’, e.g. a head has a face, a face has a mouth, a mouth has a tongue, a
tongue has a tip, etc. This is, however, not a reliable test for meronymy. Every person
has a mother, a weight, an age, and so on, but these are not parts of the person.
There is a further important difference between lexical hierarchies and meronymic
systems: unlike hyponymy and subordination, meronymy is not necessarily a
transitive relation. Although in Fig. 8.7 the term eye is linked to face and face is linked
to head, this does not result in the same kind of link between eye and head. If ‘part’ is
taken in the sense of ‘constitutive part’, i.e. something that essentially contributes to
making up the whole, then an eye is not a part of the head, as the tongue is not a part
of the face, or the face a part of the whole human body.
This much on meaning relations may suffice for the chapters to come. Meaning
relations will play a prominent role in the next chapter on meaning components.
EXERCISES
1. What are meaning relations? How do they differ from logical relations?
2. If A is a hyponym of B, what does this mean for the way in which
a. the meanings
b. the denotations
of A and B are related to each other?
3. Try to define the meaning relations between the modifier and the whole
compound for the following cases; find two analogous compounds each:
a. wood house
b. gold digger
c. soup spoon
220
Understanding semantics
4. Explain the difference between hyponymy and meronymy.
5. Which special feature distinguishes taxonomies from arbitrary hierarchies based
on hyponymy?
6. Discuss the meaning relations that hold between the following pairs of words; it
may be necessary to distinguish different meaning variants:
a. marry – divorce
b. red – orange
c. somewhere – nowhere
d. clean – dirty
e. friend – enemy
f. violin – string
7. The adjective light has several antonyms. Try to determine the pairs of opposites
and the meaning light has in each case. Consult the Oxford online dictionary for
meaning variants of the adjective light, http://oxforddictionaries.com/ (note that
there are three separate entries).
FURTHER READING
Cruse (1986) for an in-depth treatment of ‘sense relations’. For a recent survey
see Cann (2011). Foley (1997, ch. 6) on kinship terms from an anthropological
perspective (lexical fields of kinship terms were extensively studied in anthropological
linguistics.)
9
Meaning components
The assessment of semantic phenomena such as ambiguities and logical or meaning
relations forms a very important step of developing a semantic theory. However,
theorizing itself only starts when one tries to EXPLAIN the phenomena. For example,
the senses of uncle and aunt are related in the same way as those of sister and brother,
and Violetta is an aunt of Christopher logically entails Christopher is a nephew of
Violetta (provided Christopher refers to a male person and Violetta to a female one).
But how can this be derived from the meanings of the relevant terms? Likewise, verbs
carry selectional restrictions for their arguments (5.7). How can this be accounted
for in a principled way?
There are many different semantic theories. Each one defines the notion of
meaning differently and uses different means for the representation of meanings,
some formal, some informal. But almost all of them share a basic strategy of
explaining the semantic data, a strategy that ties in with our intuitive notion
of meaning: they assume that the meanings of most words are complex, composed of
more general components. The meaning of a lexical item is analysed by identifying
its components and the way in which they are combined. In a sense, the process
of semantic analysis is the converse of the process of composition by which we
determine the meaning of a complex expression on the basis of its components and
its structure. Therefore analysis into meaning components is called decomposition.
A decompositional theory allows the pursuit of the following objectives:
(M) Meaning. Providing models for word meanings.
What kind of entities are lexical meanings? What is their structure? How are
meanings to be represented?
(B)
Basic meanings. Reducing the vast variety of lexical meanings to a smaller set
of basic meanings.
Are there lexical items which are semantically basic? How can the meanings of nonbasic items be built up from more basic ones?
(P)
Precision. Providing a means of representation that allows a precise analysis of
lexical items.
What exactly are the components of lexical meanings? How can they be determined
and how can they be described in a precise way?
(R)
Meaning relations. Explaining meaning relations within and across lexical
fields.
Understanding semantics
222
How can meaning relations be accounted for on the basis of decompositional
meaning descriptions?
(C)
Composition. Explaining the compositional properties of lexical items.
With which kinds of expression can a given item be combined? How does its
meaning interact with the meanings of the expressions it is combined with?
Selectional restrictions represent one kind of phenomena relevant to the last point.
A more ambitious decompositional semantic theory would further strive to
develop an analytical apparatus that can be used for different languages. With such
an apparatus it is possible to compare the meanings of expressions from different
languages.
(L)
Language comparison. Determining the semantic relations between
expressions of different languages.
Are there expressions with the same meaning in other languages? If not, how are
the meanings of similar expressions related to each other?
For example, English and Japanese terms for ‘brother’ can be described as in Table
9.1. Japanese distinguishes between elder and younger siblings. Furthermore, there
are pairs of formal and informal terms with the same descriptive meaning. If the
relative is treated as superior, the formal terms (onīsan and otōtosan) are used.
Obviously none of the four Japanese terms is equivalent to the English word brother;
they all relate to brother like hyponyms.
Table 9.1
Meaning components of English and Japanese terms for ‘brother’
Meaning component specifying the referent’s
..kinship relation
to EGO
..sex
..age relation
to EGO
social
meaning
E
brother
›sibling‹
›male‹
–
–
J
ani
›sibling‹
›male‹
›elder‹
informal
J
onı-san
›sibling‹
›male‹
›elder‹
formal
J
oto-to
›sibling‹
›male‹
›younger‹
informal
J
oto-tosan
›sibling‹
›male‹
›younger‹
formal
Four different approaches to decomposition will be introduced in this chapter.
Each of them adopts a different conception of meaning with its own merits and
shortcomings. The first one to be introduced is the structuralist approach, now
a century old, which has exerted an enormous influence on the development of
modern linguistics and will therefore be treated in greater detail and without
restriction to semantics.
Meaning components
223
9.1 THE STRUCTURALIST APPROACH
9.1.1 Language as a system of signs
The Swiss linguist Ferdinand de Saussure (1857–1913) is generally credited with
the approach known as structuralism. The outlines of his approach appeared in
1916 under the title Cours de linguistique générale (‘Course of General Linguistics’), a
compilation of material from lectures Saussure held around 1910.
To Saussure, a language is an abstract complex system of relations and rules that
underlies all regularities to be observed in actual language use. The system is formed
by signs which are related in multiple ways. A sign, e.g. a word, consists of two parts.
One part is its sound form. The other part is its meaning. The association between
form and meaning of a sign is fixed by conventions of language use. The association
is ARBITRARY, i.e. any word could just as easily have a different meaning, and the
meaning could be associated with a different expression – provided the conventions
of the language were different. Saussure emphasizes that the sign and its parts are
real entities in the minds of the speakers – mental or cognitive entities in modern
terminology.
What distinguishes the structuralist approach is the method of analysis and the
resulting concept of linguistic signs. Saussure argues that language is to be studied
exclusively ‘from within’. The language system is a very complex structure formed by
units at different levels: single sounds, syllables, words, syntactic phrases, sentences.
Each unit is related to the smaller units it consists of and to the larger units it is a
part of, and also to all sorts of other units which are similar in some respect or other.
A sign constitutes a sign only as part of a system – only insofar as it is related to, and
different from, the other signs of the language. This holds for both the form and the
meaning of a sign. Form and meaning are only defined ‘negatively’, i.e. by differing
in specific ways from the other signs. They have no positively defined ‘substance’
and therefore cannot be described just for themselves; they can only be analysed by
describing their relationships to other signs. Neither the form nor the meaning of a
sign could exist independently of the language system it belongs to.
Let me illustrate these abstract considerations with a concrete example. The
1
French word rouge, standard pronunciation [], means ›red‹. Its sound form
is a string of three sound units (phonemes), /r/, /u/ and //, for which there are
certain conventions of articulation. Each of the phonemes allows for considerable
phonetic variation; /r/, for example is standardly pronounced []; but it can also
be pronounced as in Italian (rolled [r]) or even as [x] in German Buch. As long as
it can be distinguished from the other sounds of French, in particular /l/, its actual
articulation may vary. Japanese /r/ can even be pronounced [l] because Japanese
does not have two phonemes /l/ and /r/ but just one of this type (so-called liquids).
1 See Gussenhoven and Jacobs (2011, ch.1) for an explanation of the sounds. It is common practice
in linguistics to write the phonetic pronunciation in square brackets […] and with the symbols of
the IPA (International Phonetic Association) alphabet, while the phonological form and phonemes
(sound units of the language system) are written in slashes /… / using standard letters if possible.
224
Understanding semantics
Thus, Saussure argues, a sound unit of the system is not primarily defined by its
articulation or acoustic quality but by its relation to the other sound units.
The other component of the sign, the meaning of French rouge, is similar to the
meaning of English red or German rot (but see the next chapter for other languages).
The exact range of colours that the word denotes depends on which other colour
terms compete with rouge: orange (›orange‹), rose (›pink‹), violet (›purple‹) and brun
(›brown‹). If, for example, French lacked a term for purple, the respective range of
colours would be divided between the words for red, brown and blue, resulting in an
extension of the range of rouge. A sign, according to structuralism, IS CONSTITUTED BY
being different from the other signs of the system. Its form and its meaning are the
sum of their differences and relations to the other forms and meanings of the system.
The structuralist notion of meaning is thus radically relational. Strictly speaking, it
implies that we cannot determine the meaning of a lexeme independently, but only its
relations to the meanings of other lexemes. This notion of meaning is not commonly
accepted. While it is adopted by some semanticists (e.g. in Cruse 1986), alternative
theories have been developed, in particular cognitive semantics (chapter 11). The
cognitive approach does view the meaning of a lexeme as something that can in
principle be investigated and described independently. In this theory, meanings are
considered integral parts not of a system of linguistic meanings but of our overall
cognitive system, which can be studied by the methods of cognitive psychology.
In terms of the semiotic triangle, structuralism focuses exclusively on the left
side of the schema (Fig. 9.1). Semantics is not a matter of reference and denotation.
Rather it concerns the sign, i.e. a pair of form and meaning, and how it is related to
other signs.
Figure 9.1
The focus of structuralist semantics
expression
descriptive
meaning
denotation
9.1.2 Syntagmatic and paradigmatic relations
All linguistic units – sounds, syllables, words, phrases, sentences – can be combined
into more complex units. In the original terminology of structuralism, such a
Meaning components
225
complex is called a syntagm (from ancient Greek syntagma ›compound‹). A syllable
is a syntagm of phonemes, a word taken as a sound unit is a syntagm of syllables;
as a semantic unit, a word may be a syntagm of morphemes (e.g. un-natur-al, see
next section); a syntactic phrase like the NP a lucky decision is a syntagm of words,
a sentence is a syntagm of phrases. For each kind of syntagm there are specific
formation rules and within each syntagm the constituents are related to each other
in specific ways. The general term for the relations within a syntagm is syntagmatic
relations. The syntagmatic relations that a given unit bears to other constituents
are determined by its combinatorial properties. For example, an English NP may
take the form article + noun but not noun + article. Thus English articles have the
combinatorial property of preceding the noun.
Let us consider the syllable /set/ as an example (Fig. 9.2) of a syntagm of phonemes.
Figure 9.2
Three paradigms defined by the syntagm /set/
onset
/b/
bet
/p/
pet
/s/ /e/ /t/ set
/m/
met
/w/
wet
…
…
nucleus
/i/
/æ /
/s/ /e/ /t/
/u:/
/ae/
…
sit
sat
set
soot
cite
…
coda
/k/
/p/
/s/ /e/ /t/
/n/
/l/
…
sec-tion
sep-tet
set
cen-tury
cell
…
The syllable consists of the three phonemes /s/, /e/ and /t/, that occupy the
three basic positions of a syllable, commonly called onset, nucleus and coda. The
syntagmatic relations are first of all those of order: the onset precedes the nucleus
and the nucleus precedes the coda. In addition, nucleus and coda form a unit within
the syllable, the so-called rhyme. Therefore the relation that holds between nucleus
and coda is closer than, and different from, the relation between nucleus and onset.
For each position we can determine which units can occur in this position.
2
The set of all alternatives is called a paradigm (from ancient Greek paradeigma,
›something shown side by side‹). The onset position of the syllable /set/ can be
filled by certain, but not all consonant phonemes (e.g. //, the last sound of sing,
is not possible here) and by certain consonant clusters such as /fr/, which we will
disregard here. The nucleus position defines a paradigm that consists of all vowel
and diphthong phonemes. The coda paradigm consists of all consonant phonemes
that can occur in syllable-final position (excluding, for example, /h/, /y/ and /w/).
The units within a paradigm exhibit relations of difference and similarity. These
are called paradigmatic relations. For example, the opposition voiced vs voiceless
2 This is a special use of the term paradigm in structuralist terminology.
226
Understanding semantics
distinguishes /b/ and /p/ within the onset paradigm. Let us call the corresponding
properties of the units, e.g. voice (being voiced), their contrastive properties.
Table 9.2 displays the basic structuralist concepts. They allow the definition of
basic notions such as syntactic category (a paradigm of syntactic units), word class or
grammatical category (a paradigm of lexical units, 3.1) or lexeme (a unit with certain
paradigmatic and syntagmatic properties).
Table 9.2
Basic structuralist concepts
Structuralist concept
Definition
syntagm
complex unit
syntagmatic relations
relations between the constituents of a syntagm
combinatorial properties
syntagmatic properties
properties of a unit that determine its syntagmatic relations:
how it can be combined with other units
paradigm
set of all elements that can fill a certain position in a syntagm
paradigmatic relations
relations between the elements of a paradigm
contrastive properties
paradigmatic properties
distinctive properties of the elements of a paradigm
9.2 APPLYING THE STRUCTURALIST APPROACH TO MEANING
9.2.1 Semantic units: morphemes and lexemes
For applying the structuralist approach to meaning, one has to determine what the
semantic units of a language are, and describe them in terms of their combinatorial
and contrastive properties. Words may share meaningful parts with other lexemes.
Therefore, words (or lexemes) are not the smallest semantic units of a language.
Consider for example:
(1)
a.
meaningful, beautiful, harmful, successful, useful, sorrowful
b.
ugliness, cleverness, baldness, usefulness, carelessness, laziness, fitness
c.
clearing, hearing, cleaning, thinking, meeting, brainstorming
The meanings of these and many other words can be derived from the meaning
of the first part and the meaning of the suffix: -ful, -ness, -ing. The suffix -ful can
be attached to a large number of nouns and yields an adjective that, very roughly,
means ›having much of N‹ or ›being full of N‹. Thus, the suffix -ful has a meaning
of its own. Likewise, the meanings of the other two suffixes can roughly be defined
as ›the property of being A‹ for -ness attached to an adjective A, and ›situation
where someone V-s (something)‹ for words of the form V-ing. It should be noted,
though, that along with these examples where the meaning can be straightforwardly
Meaning components
227
decomposed into the meaning of the suffix and the meaning of the first part, there
are quite a number of lexemes that contain such a suffix without allowing for
straightforward decomposition. For example, meaning does not mean a situation
where somebody means something, wilderness does not denote the property of being
wilder and lawful does not mean ›full of law‹.
The two components of cleaning, the verb clean and the suffix -ing, cannot be
further divided into meaningful parts. Such minimal semantic units are called
3
morphemes. Clean can be used as a word on its own. It therefore constitutes what
is called a free morpheme. The suffix -ing is a bound morpheme as it can only occur
as a part of a word. The bound morphemes in (1) serve the derivation of words from
others. A second class of bound morphemes, the grammatical morphemes, is used
to build the different grammatical forms of words. Among the comparatively few
examples in English are the 3rd person singular present tense ending (eat-s), the
noun plural ending (horse-s), the possessive ending (king-’s), the past tense ending
(wash-ed), the comparative ending (high-er) and the ending -ing used to form the
progressive form of verbs (be wash-ing), the present participle (the ring-ing bell) and
the gerund (I hate iron-ing).
9.2.2 Paradigmatic and syntagmatic semantic relations
When meaning relations were discussed in chapter 8, we assumed that the related
items were of the same word class. Word classes are essentially paradigms in the
structuralist sense. Hence, meaning relations such as the oppositions, synonymy,
hyponymy, meronymy and the more specific relations structuring certain lexical
fields (8.3) are paradigmatic relations. For example, all terms for the days of the
week can be inserted into the empty position of the syntagm today is ___. Within the
resulting paradigm, we can assess the meaning relations between the terms.
Syntagmatic meaning relations hold between the constituents of a syntagm. They
consist in the way in which the meanings of the constituents are combined to yield
the meaning of the whole syntagm. Let us consider two simple examples. In (2) a
transitive verb is combined with a subject NP and an object NP to form a sentence:
(2)
Mary seized the bottle.
The syntagmatic meaning relation that holds between the subject NP and the verb
is that of the NP specifying the verb’s agent argument. The object NP is related to
the verb as its theme specification. The semantic relations of the NPs to the verb are
indicated by the linking rules (5.6). There is no direct semantic relation between the
subject and the object NP.
(3)
the red balloon
3 See Haspelmath (2010) on the notion of morpheme and other basic concepts of morphology.
228
Understanding semantics
The combination red balloon in (3) describes the referent of the NP as a red balloon.
Both the adjective and the noun are one-place predicates about the potential referent
of the NP (5.4.2.1). Thus the syntagmatic meaning relation between adjective and
noun is one of sharing their argument. This results in a conjunction of the two
predications. In 5.4.2.3 we encountered examples of non-predicational adjectives:
combinations such as alleged terrorist, future neighbour, atomic war, educational
disaster or doctoral dissertation. These adjectives interact with the noun in a different
way.
Predicate terms carry selectional restrictions. These conditions constrain the
choice of expression that can be meaningfully combined with the predicate terms
(5.7). Thus they constitute syntagmatic, or combinatorial, semantic properties of
verbs, adjectives and nouns. In addition to selectional restrictions, which impose
logical conditions on argument specifications, lexemes may have combinatorial
properties that restrict their usage further. For example, German has different
systems of terms for animals and people. Where people ‘essen’ (eat), animals ‘fressen’,
drinking of people is ‘trinken’, of animals ‘saufen’. People have a ‘Mund’ (mouth), but
4
animals a ‘Maul’, ‘Schnauze’, etc. Such distinctions give rise to language-specific
combinatorial meaning properties.
9.3 SEMANTIC FEATURES
9.3.1 Binary semantic features
One of the areas where the structuralist approach proved particularly fruitful was
phonology. As early as in the 1920s and 1930s the Prague School of structuralism
developed a phonological theory where the phonemes were described by sets of
distinctive binary features corresponding, for example, to the mode and place of
5
articulation. The use of features not only allowed the explanation of contrastive
and combinatorial properties of the phonemes within one language. Features could
also be established across languages, enabling a comparison of the sound systems of
different languages. Because of its great success, the feature approach was adopted in
other areas of linguistics, including semantics.
In the variety of approaches subsumed under the label of feature semantics,
semantic features are used as meaning components. For instance, the lexemes in (4a)
would receive a common semantic feature [FEMALE], those in (4b) receive the feature
[MALE], while neither feature is assigned to the expressions in (4c). (It is common
practice to write features within square brackets and with small capital letters.)
4 This only holds when the animal-related terms are used in their primary sense. Not surprisingly,
these expressions can also be used derogatively for people. For example, ‘saufen’ applied to people
means drinking greedily, excessively, etc. or heavily drinking alcohol. Halt dein Maul! (‘Shut up!’, lit.
‘hold your [animal] mouth’) has an extra negative expressive quality compared to the phrase Halt
deinen Mund! with the neutral term for mouth. (Both expressions are rude.)
5 See Gussenhoven and Jacobs (2005, ch. 5) for a phonological feature analysis.
Meaning components
(4)
229
a.
girl
woman
sister
wife
queen
[FEMALE]
b.
boy
man
brother
husband
king
[MALE]
c.
child
person
sibling
spouse
monarch (sex not specified)
The features [FEMALE] and [MALE] are not just different, but complementary. We
can therefore replace them with one binary feature, either [FEMALE] or [MALE], that
assumes the value + or the value –. For the sake of brevity and male chauvinism,
6
the [MALE] variant is chosen. The general feature [MALE] is written [±MALE] or
just [MALE], the fact that the feature has the value + or – is indicated by [+MALE]
and [–MALE], respectively. The term feature is used for both features without a value
([MALE]) and features with a value ([–MALE]).
The words in (4) share the property of denoting persons. This can be captured by
another feature [+HUMAN] that distinguishes them from the terms in (5). (We assume
the primary animal readings of the terms.)
(5)
a.
mare
bitch
cow
ewe
[–HUMAN]
[–MALE]
b.
stallion
cock
tomcat
bull
[–HUMAN]
[+MALE]
c.
horse
pig
sheep
fish
[–HUMAN]
With [HUMAN], [MALE] and an additional feature [ADULT], the meanings of the six
general person terms (8.4.2) can be described as in Table 9.3. The right side is a
so-called feature matrix.
Table 9.3
Features and a feature matrix for the six general person terms
Features and their values
[HUMAN]
[ADULT]
+
–
[MALE]
child
[+HUMAN]
[–ADULT]
girl
[+HUMAN]
[–ADULT]
[–MALE]
+
–
–
boy
[+HUMAN]
[–ADULT]
[+MALE]
+
–
+
adult
[+HUMAN]
[+ADULT]
+
+
woman
[+HUMAN]
[+ADULT]
[–MALE]
+
+
–
man
[+HUMAN]
[+ADULT]
[+MALE]
+
+
+
A feature with the value + or – constitutes a one-place predicate about the
potential referents of the lexeme. For example, if a noun carries the feature [+ADULT],
6 In the view of more recent feature theories, the choice is a matter of markedness. Of the two
alternatives, usually one is marked, or special, and the other unmarked, or normal, with clear
linguistic criteria for markedness. Features and feature values are then chosen in the way that [+a]
stands for the marked variant. In English, there is a certain tendency of marking expressions for
female persons or animals (cf. steward vs steward-ess) by adding a special morpheme. This would
provide a criterion for using the feature [±FEMALE] instead of [MALE].
230
Understanding semantics
its referents have to fulfil the condition of being adult; if it carries the feature
[–ADULT], they have to be not adult; if it has neither feature, no condition with respect
to being adult is imposed on its referents. Thus, in feature semantics word meanings
are considered combinations of a number of one-place predications. For example,
according to the analysis in Table 9.3, the meaning of the word boy is ›human and
male and not adult‹.
9.3.2 Application to paradigmatic relations
Binary features are directly related to certain logical relations and the corresponding
meaning relations (cf. Table 9.4). Two expressions with opposite values for some
feature [α], e.g. boy and mare, are logically incompatible, regardless of the rest of
their meanings. If the meanings of two expressions differ ONLY in the value of one
feature (boy vs girl), they are complementary opposites; the rest of their meaning
defines a semantic domain within which they are contradictory. Hyponymy holds
if two expressions have the same meaning except for an additional feature [+α] or
[–α] for one of them (e.g. child and boy). As will be argued later, hyponymy (of this
type) and complementary opposition are the only meaning relations that can be
captured with binary features.
Table 9.4
Binary features and logical relations
Meaning of A
Meaning of B
Logical relation
Meaning relation
X and [+a]
boy
vs.
Y and [–a]
mare
incompatibility
(undetermined)
X and [+a]
boy
vs.
X and [–a]
girl
incompatibility
complementary opposition
X and [+a]
X and [–a]
girl, boy
vs.
X
child
subordination
hyponymy
9.3.3 Application to combinatorial meaning properties
Features can be used to formulate selectional restrictions. For example, if a verb
(e.g. vaccinate) carries the selectional restriction ›human‹ for its subject referent, the
restriction can be described by the feature [+HUMAN]. It must, however, be carefully
observed what this is a feature of: it is a feature of a complement argument of the
verb, not a feature of the verb itself. If it were the latter, it would have to relate to the
referent of the verb, i.e. the event expressed (5.3.2). But the event, e.g. of vaccinating,
is not a human being; rather the agent is that performs the action. Conditions on
complements of the verb can, in fact, be captured by a decompositional analysis
of verb meanings, as we shall see in section 9.4.2, but not with the binary feature
approach.
Meaning components
231
9.3.4 Ideal properties of semantic features
So far only single features have been extracted from the meanings of lexemes. The
natural next step in feature analysis is the attempt of decomposing lexical meanings
COMPLETELY into binary features. The result for a particular lexeme would be a finite
list of features that captures all relevant properties of its potential referents and
differs from the feature list of any other non-synonymous lexeme in at least one
feature. This, however, has proved very difficult. It is by no means obvious how, for
example, the set of co-hyponyms of large taxonomies such as the animal or plant
taxonomy should be distinguished systematically by binary features.
Formally, a trivial feature analysis is always possible. We could, for example,
distinguish the animal terms bear, pig, rabbit, tiger, donkey, kangaroo, etc. by simply
introducing as many binary features [±BEAR], [±PIG], [±RABBIT], etc. and assigning to
each animal term a positive value for its ‘own’ feature and a negative value for each
other feature. Tiger would be [–BEAR], [–PIG], [–RABBIT], [+TIGER], [–DONKEY] and
[–…] for all the other features. The result is shown in Fig. 9.3.
Figure 9.3
A trivial feature matrix for co-hyponyms in the animal taxonomy
mammal
bear pig
[±BEAR]
[±PIG]
[±RABBIT]
[±TIGER]
[±DONKEY]
[±KANGAROO]
[±…]
rabbit
tiger donkey
kangaroo
…
+
–
–
–
–
–
–
–
+
–
–
–
–
–
–
–
+
–
–
–
–
–
–
–
+
–
–
–
–
–
–
–
+
–
–
–
–
–
–
–
+
–
–
–
–
–
–
–
…
This approach would capture in an extremely plain way one major aspect of
the structuralist notion of meaning: that the meaning of a unit is the sum of
its differences from the meanings of the other units of the system. But, clearly,
such an ‘analysis’ would not explain anything beyond the mere incompatibility
of the different animal terms (and this only if we tacitly assume that the features
themselves are mutually incompatible, i.e. if we already presuppose what should be
explained). The main point of decomposition is, of course, the REDUCTION of meaning
to more basic and ultimately minimal components. At least we would expect all the
expressions in Fig. 9.3 to carry a common feature [+ANIMAL]. The use of this feature
232
Understanding semantics
would then lead to the employment of features less specific than [TIGER], etc. This
leads us to a first important postulate for any feature approach, and in fact for any
decompositional theory:
∑ Features should be primitive, i.e. not further analysable.
The 1:1 analysis into as many features as there are expressions to be distinguished
suffers from a further principal drawback: the features are very special. For example,
the feature [±PIG] would only be of use when it takes the value +, i.e. with terms for
more specific kinds of pigs or for male, female or baby pigs. The feature value [–PIG],
though a necessary component of the meanings of all other animal terms contributes
almost no information (I do not know much about the meaning of rat if I know that
rats are [–PIG]). A feature system should provide distinctions that can be used more
widely, e.g. [±DOMESTIC]. Features must not be ‘ad hoc’, i.e. created just in order to
provide a solution to a particular problem, while disregarding more general aspects.
∑ Features should be general.
Of course, generality is a matter of degree. But what is certainly ruled out are features
that are distinctive for a handful of lexemes only.
9.3.5 Linguistic motivation
If we try to find more general features that allow us to structure the taxonomy of
animals, we will soon fall back upon our knowledge from biology lessons, trying to
apply the criteria of scientific classification such as the distinction between rodents
and other orders. But, as will be argued more thoroughly in 11.6, a distinction must
be drawn between semantic knowledge, i.e. the knowledge of meanings, and world
knowledge, i.e. knowledge about the entities we have words for. Semantic analysis has
to be based on LINGUISTIC data, namely the meaning relations observable within the
language system. After all, we are not doing the work of biologists, etc., but of linguists.
∑ Features should be linguistically motivated.
What would constitute relevant linguistic data that justify the assumption of
semantic features? So far we have only made use of instances of complementary
opposition or hyponymy. There are various other phenomena; examples will be taken
from the realm of animal terms. Of course, the phenomena are not restricted to this
part of the lexicon.
Selectional restrictions of verbs. There are specific verbs for the sounds certain
animals produce: e.g. miaow, bark, chirp, neigh or words for modes of running such
as gallop, trot, etc. The selectional restrictions of such verbs give rise to corresponding
features in the meanings of animal terms.
Verbs with implicit arguments. The German verb reiten (a cognate of the English
verb ride) is interpreted as ›to ride a horse‹ if the respective argument is not explicitly
specified otherwise. This justifies the assumption of a meaning component ›horse‹.
It is not clear, however, how in this approach a feature [+HORSE] can be built into the
Meaning components
233
meaning of the verb, since it is not the referent of the verb (i.e. the riding event) to
which the feature applies. Similarly there might be adjectives which are only used
when referring to properties or conditions of certain kinds of animals.
Lexicalizations of distinctions within certain kinds or groups. For some kinds of
animal there are general terms along with terms for the sexes (horse vs mare, stallion)
or their young (kitten, chicken, puppy, calf, lamb, cub, foal, etc.). There are specific
terms for horses of certain colours, e.g. sorrel for a horse of a light reddish-brown
colour. The more specific terms (mare, foal, sorrel) require features that they share
with the more general terms (horse). In other cases, the same word may be used to
form expressions for the males or females of different kinds of animal, e.g. hen and
cock for most kinds of bird. The terms for these animals too must share a feature.
Mereological relations. For many kinds of animal there are mostly analogous
meronymies (8.4.4) due to their similar anatomies. But often different terms are used
for corresponding body parts, for instance snout, muzzle, trunk or beak for mouth
and nose. Thus, animal terms can be grouped together according to the expressions
used for their body parts: terms for animals with a snout, animals with a beak, etc.
In such cases, a feature is linguistically motivated, e.g. [+BEAK] for animals with a
‘beak’. We also may distinguish terms for animals with special body parts, e.g. with
‘claws’, ‘hooves’, ‘fins’, ‘feathers’, ‘scales’, with ‘fur’ or a ‘shell’, or with a ‘tail’, a ‘mane’,
an ‘udder’ and so on. Although features like these border closely on extralinguistic
scientific characterizations, they nevertheless are linguistic features. We cannot
explain the meaning of fin, feather, udder, etc. without relating the meaning of the
words to the meanings of the words fish, bird, cow, etc. or hyperonyms of them.
There is a conceptual connection between, for example, ›bird‹ and ›feather‹: the fact
that birds have feathers is part of the meaning of bird as well as part of the meaning
of feather. A feature for the property of having feathers is therefore linguistically
motivated for the word bird as well as for all terms for kinds of birds with feathers.
Note that such a criterion is different from saying that, for example, seagulls can be
distinguished from other kinds of bird by the fact that their eggs are a certain colour:
there is no lexical relation between the word seagull and the respective colour word.
Metonymical relations. The semantic relations based on mereology are a special
case of metonymical relations. These apply more generally between terms for animals
or other kinds of objects and terms for things that belong to them. Examples are the
terms spawn for the eggs of fish or certain other water animals, nest for many animals,
or web for spiders. Between fish and spawn we have a meaning relation in which
referents of one term belong to referents of the other, i.e. a metonymical relation as
introduced in 3.4.2. The same kind of general relationship links the terms for certain
animals and for their meat when used as food: pig–pork, cow–beef or sheep–mutton.
Among the desirable properties of semantic features, or meaning components in
general, linguistic motivation is the most important. Unlike the other three properties
mentioned, it is absolutely necessary: only linguistic motivation guarantees that
the conditions are really meaning components, not just parts of our general
world knowledge. There may be a conflict between linguistic motivation and the
234
Understanding semantics
requirements that features be primitive and general. For example, the feature [HORSE]
is linguistically motivated due to the existence of words such as mare, stallion, colt or
neigh, but it is neither primitive nor general.
There is a linguistically motivated meaning relation between lexical items
whenever it is possible to form logically true sentences (7.1.4), called meaning
postulates, that express the relation.
(6)
a.
A mare is a female horse.
b.
A kitten is a baby cat.
c.
Pork is meat from pigs.
d.
Roe is the eggs of fish.
e.
Birds have feathers.
f.
Birds have a beak as their mouth.
9.3.6 Types of features
Although such linguistic data will take us a bit further along the way to an
exhaustive analysis, it is by no means clear whether it is possible to find a set
of appropriate features for distinguishing the great number of animal terms or
other large taxonomies. In fact, it is commonly accepted that for many lexemes
an exhaustive analysis into binary features that are at the same time primitive,
general and linguistically motivated is impossible. Features which exhibit all three
ideal properties are called classemes, or markers. It is markers which allow an
explanation of the combinatorial properties as well as the more general contrastive
properties. Markers may also play a role in grammar. For example the feature
[±HUMAN] is responsible for the choice of the pronouns who vs what, or [±MALE]
for the choice of he vs she. Ideally, markers also fulfil a further important condition,
related to the aim of language comparison:
∑ Features should be universal.
Indeed, features such as [±MALE], [±HUMAN], [±ANIMATE], etc. play an important role
in the structure of the lexica of very many languages.
For most lexemes, however, there will remain a residue of meaning that cannot
be captured by markers. For such cases, the notions of semes and distinguishers
were introduced. Semes serve for specific, more or less unique, distinctions within
particular lexical fields. For example, in the lexical field of the verbs buy, rent, take,
borrow, etc. one can use the semes [±PERMANENT] or [±FOR MONEY]. Semes are
primitive and linguistically motivated, but they are neither general nor, in most
cases, expected to be universal. Distinguishers are used for whole meaning residues.
If the meaning of mare were analysed just into the two features [–MALE] [+HORSE],
[–MALE] would be a marker and [+HORSE] a distinguisher. Distinguishers are neither
primitive nor general; they are improbable candidates for universality, but they may
be linguistically motivated.
Meaning components
235
9.3.7 Evaluation of the binary feature approach
While the analysis of lexical meanings into a set of binary features can be used for
explaining some semantic data, the limitations of the binary feature approach (BFA)
are in fact severe.
As stated above, binary features are essentially one-place predicates about the
referent of the lexeme. For example, the decomposition of the meaning of woman
into the three features [+HUMAN], [–MALE], [+ADULT] is equivalent to the following PL
definition (cf. 5.5):
(7)
x is a woman = human(x) Ÿ ¬male(x) Ÿ adult(x)
‘x is human and not male and adult’
The PL notation shows that any possible decomposition of meaning into binary
features has a uniform and very simple structure: a conjunction of a finite set of
one-place predicates, possibly negated, if the feature takes the value ‘–’ (minus). All
predicates have the same argument, namely the potential referent of the lexeme.
∑ In BFA there is only one kind of meaning component and only one way of
composing meanings out of them.
Consequently, the potential meaning components can only combine to form one
type of meaning: one-place predicates, since the conjunction of one-place predicates
always yields a one-place predicate.
Therefore, most verbs cannot be analysed along the approach of BFA, nor does it
apply to relational nouns or multi-place adjectives. Some aspects of the meanings
of multi-place verbs can be captured; for example, the difference between ›buy‹ and
›rent‹ by means of a feature [±PERMANENT]. But the core of the meaning, common
to both verbs, is a three-place concept about a change of possession which involves
the former and the future possessor and some object of possession. This three-place
predicate can never be composed out of one-place predicates.
The restriction to one-place predicate expressions is a severe limitation of BFA. In
addition, the approach is limited even in the case of one-place predicate expressions,
due to the principal difficulties of total decomposition discussed above.
∑ BFA is only applicable to a limited range of lexemes.
A further restriction results from the fact that the features of BFA can only relate to
the potential referents of a lexeme.
∑ In the case of multi-place predicate terms, the BFA can only capture conditions
concerning the potential referents of the lexeme.
As we have seen, certain phenomena, such as selectional restrictions, require the
possibility of imposing conditions on arguments of predicate terms. Similarly, we
might for example want to distinguish terms for items of clothing with respect to
the sex of the wearer, or terms for body parts of either female or male persons. But it
would be inappropriate to assign a feature [–MALE] or [+MALE] to the words bikini or
236
Understanding semantics
prostate: it is not the bikini or the prostate that is female or male, but the persons for
which bikinis are designed or which have this type of organ.
As is shown in Table 9.4, among the basic meaning relations only semantic
complementarity and certain cases of hyponymy can be explained in terms of
binary features. Conversion cannot be captured in principle, because this relation
involves two-or-more-place predicate terms. The BFA also has difficulties with other
meaning relations, e.g. antonymy. For example, in order to distinguish big and small,
one would have to use two features such as [± HIGH] and [± LOW] relating to the
7
intervals of the scale of size they denote (cf. Fig. 9.3). This is not particularly elegant.
More specific meaning relations that underlie certain lexical fields can provide
huge if not insurmountable problems for BFA, among them mereological relations,
the cyclic order of terms for the days of the weak or the months of the year, or the
various relationships between head and modifier of a regular compound (5.2.2). It is
impossible to capture the semantic relations between, say, three, four, twenty, hundred
8
and three hundred and twenty-four with binary features for these five numerals.
Thus, the last severe shortcoming of BFA is as follows:
∑ Most kinds of meaning relation cannot be accounted for with BFA.
To sum up, BFA is much too simple an approach to be able to meet the complexities
of semantic phenomena in natural languages. The use of binary features may be
appropriate for dealing with the small and limited areas of phonology or syntax,
where only a comparatively small set of units or categories has to be distinguished
and analysed with respect to their relations within the system. But the lexicon of a
language is of a complexity incommensurable to that of the sound system or system
of word classes or forms. It forms a large part of our utterly complex cognitive
system. It is not plausible to assume that the decomposition of word meanings is
restricted to structures as simple as those provided by BFA. It should, however, be
emphasized that the failure of BFA in semantics does not mean that the structuralist
approach as such has failed. Its strong emphasis on linguistic data and the systematic
relations within languages helped linguists to develop their theoretical methods of
today. The basic notions and methods of structuralism are compatible with semantic
theories quite different from BFA. In principle, it would be possible to develop
theories of lexical meaning within the framework of structuralism which are much
more adequate and powerful than BFA.
In the following, we will have a look at three approaches to decomposition that
provide alternatives to the structuralist tradition.
7 Big is assigned the feature [+HIGH], and small [+LOW], while not big would express [–HIGH]. In
addition, one would have to stipulate that [+HIGH] entails [–LOW] and [+LOW] [–HIGH]. However,
entailments should be derivable from the decomposition rather than being stipulated somewhere
else. The logical relationship between [HIGH] and [LOW] proves that these two ‘meaning components’
have common content; thus a decomposition in terms of [HIGH] and [LOW] is not of much
explanatory value.
8 These relations consist, among others, in the fact that three hundred and twenty-four denotes the
sum of the referents of three hundred and of twenty-four, three hundred denotes the product of the
referents of three and of hundred, and so on.
Meaning components
237
9.4 SEMANTIC FORMULAE
One phenomenon that gave rise to the development of alternative decompositional
approaches is the existence of systematic meaning relations between verbs, or
between verbs and adjectives, as given in (8). We encountered the phenomenon
above, when dealing with verb alternations in 6.1.
(8)
stative
x is open
x is hard
x is dead
x has y
inchoative
x opens
x hardens
x dies
x gets y
causative
y opens x
y hardens x
y kills x
z gives y to x
The verbs and adjectives in the first column denote a certain state of x. The inchoative
verbs in the second column express the coming about of this state of x. The third
column consists of causative verbs, used to express that someone (or something),
y (or z), brings about the state of x. While the stative and inchoative expressions
have the same number of arguments, the causative verbs have an additional causer
argument. There are hundreds of such triplets in English, and in addition many pairs
of verbs, or pairs of an adjective and a verb, which would fill two of the three places
in a row with the third term lacking in the lexicon, for example, wet/–/wet, –/lie down/
lay down. In some cases the expressions that enter this scheme have the same form
(but note that they represent three different lexemes as each belongs to a different
grammatical category), as in the case of open/open/open and harden/harden. In
others, the terms are related in form: hard/harden, and historically also dead/die. But
the forms of the terms may as well be morphologically unrelated (small/–/reduce).
There is a systematic meaning relation between the three members of each triplet. It
can be captured by logical entailments, such as:
(9)
a. y hardens x
fi x hardens and
x hardens
fi x will be hard
b. z gives y to x
fi x gets y
x gets y
fi x will have y
and
Clearly, a BFA approach is not able to explain these phenomena. One could, for
example, assign a common feature [+OPEN] to the three terms in the first group,
and distinguish the three elements by two additional features, say [–CHANGE] for the
adjective and [+CHANGE] for the two verbs and [+CAUSE] for distinguishing the third
item from the other two. This would help to keep the three elements of a group apart
and mark the terms of different groups in the same row as terms with common
semantic characteristics. But it would not allow the explanation of the exact meaning
relations.
9.4.1 Dowty’s decompositional semantics
A classical, though not the earliest, decompositional approach to this phenomenon
is Dowty (1979), a theory developed in the framework of Montague Grammar, which
238
Understanding semantics
will be introduced in chapter 13. Dowty uses decomposition in order to model
aspectual classes (6.2). His method is to give a general formula for each aspectual
class. The general formulae have a fixed structure in which individual elements are
inserted for individual verbs. The formulae are defined in predicate logic (5.5).
Dowty analyses the adjective openA as a stative expression not further decomposed
into meaning components. Using the notation of predicate logic, he represents the
meaning of x is open by a simple formula in which the state term open is used as a
one-place predicate constant:
x is openA:
(10)
open(x)
The meaning of the intransitive verb openIV is then represented by means of an
operator BECOME which can be applied to any stative formula:
x opensIV:
(11)
BECOME(open(x))
The component BECOME is used to express a change from a situation where the stative
formula is false to a situation where it is true. If ‘BECOME(open(x))’ is true at a certain
time t, then ‘open(x)’ is false before t and true afterwards. It thus follows from the
meaning representations of openA and openIV that x opensIV entails x will be openA.
For the causative verb openTV , Dowty employs an operator CAUSE which takes
two argument terms, an individual term for the causer, and a formula for the event
caused. The meaning of y opensTV x is represented as in (12) and interpreted as: ›y
causes x to become open‹. Hence, y opensTV x logically entails x opensIV.
y opensTV x:
(12)
9
CAUSE (y, BECOME(open(x)))
Crucially, the meaning of openTV does not simply contain the three meaning
components CAUSE, BECOME and open in the way the meaning of woman is thought
to contain the three meaning components [+HUMAN], [–MALE] and [+ADULT] in BFA
(three one-place predicates linked by conjunction). Rather, the three components
combine with each other in a specific way. The component CAUSE is a two-place
predicate; its first argument is the agent argument of the verb; its second argument
is BECOME(open(x)), a complex containing the two meaning components BECOME and
open. BECOME is used as a one-place predicate with a (stative) situation argument.
This argument, in turn, consists of the one-place predicate open applied to the theme
argument of the original verb.
An analogous analysis can be given for the other groups in (13). For example,
(13) a.
x has y:
have(x,y)
b.
x gets y:
BECOME(have(x,y))
c.
z gives y to x:
CAUSE (z, BECOME(have(x,y))
9 The formulae used here are slightly simplified versions of the original ones given in Dowty (1979).
Meaning components
239
The explanatory potential of the approach is obvious. There are hundreds if not
thousands of stative, inchoative and causative verbs and adjectives whose meanings
can be decomposed in this way. The decomposition completely uncovers the meaning
relations that hold within and across the groups. This particular analysis is less
concerned with the question whether meaning components such as open or have are
primitives. It might well be the case that some of them can be further decomposed.
But this would not reduce the explanatory power of this kind of approach.
9.4.2 Jackendoff’s Conceptual Semantics
10
Jackendoff ’s Conceptual Semantics is another theory that deals with semantic
phenomena concerning verbs (but not exclusively). The foundations of his theory
are quite different from Dowty’s. Unlike Dowty, Jackendoff assumes that meanings
are concepts represented in our mind. To him, basic semantic components are at
the same time conceptual components out of which the human mind is capable of
composing all kinds of concepts it operates with. The way in which meanings are
decomposed in Jackendoff ’s approach is however similar to Dowty’s. We cannot go
deeper into the foundations of his theory nor enter into the details of his semantic
representations, but will confine the discussion to a few examples that give an idea of
one more way in which meanings can be decomposed.
With tense neglected, the sentence John went home would be analysed as (14):
(14) John went home:
[Event GO ([Thing JOHN], [Path TO ([Place HOME])])]
This looks more complicated than it actually is. Let us first omit the square brackets
together with their subscripts. We then obtain a structure of the same kind as those
used by Dowty:
(15) John went home:
GO(JOHN, TO(HOME))
is a primitive concept, and meaning component, for a movement of an object
along a path. Correspondingly, the concept GO has an object argument and a path
argument. In (15) the primitive concept JOHN fills the first argument position. The
second position is occupied by TO(HOME), a specification of the path in terms of its
goal. The one-place operator TO is another primitive concept, which takes a (concept
for a) place as its argument, in this case HOME, and yields a concept for a path to this
place. If we now return to the more complex notation in (14), we see that each pair
of square brackets [ ] encloses one concept; the subscript at the left bracket assigns
the concept to one of a few basic categories: Event, Thing (including persons), Path
and Place, others being Action, State, Time, Amount or Property. From (14) we can
abstract a semantic representation of the verb go by omitting the arguments of the
GO concept, replacing them by empty slots.
GO
10 The following examples are taken from Jackendoff (1990).
240
Understanding semantics
go: [Event GO ([Thing __ ], [Path __ ])]
(16)
It could be argued that (16) does not really constitute an instance of decomposition
since the meaning of go is essentially represented by a conceptual primitive GO. This
is not much better than representing the meaning of pig by a single binary feature
[+PIG]. Jackendoff, however, argues that the concept GO happens to be a conceptual
primitive. This makes the verb go a semantic primitive. The conceptual primitive GO
is a component of concepts for all kinds of transitions. (Jackendoff ’s is a so-called
localistic theory. Localism assumes that the linguistic concepts of space and motion
are basic and serve as patterns for non-local concepts.) For instance, John gave Bill
$5 is represented as in (17a), brackets and subscripts omitted. The meaning of give
itself would be (17b), with three empty argument places to be filled with concepts for
the giver, the given and the recipient:
(17) a.
b.
John gave Bill $5:
CAUSE(JOHN, GO($5, TO(BILL)))
give:
CAUSE( __ , GO( __ , TO( __ )))
Embedded within the CAUSE concept, which has the same structure as the predicate
cause in Dowty’s analyses, we find the concept GO used to represent a transition of
some object into the possession of Bill: the concept of motion in space is applied
to a transfer of possession. The example may convey a first impression of the
kind of generalizations a decompositional approach like this allows. In fact, it
appears possible to reduce the meanings of thousands of verbs in English or any
other language to a very limited number of general patterns composed out of a
comparatively small set of basic conceptual components.
Again, the main asset of this approach lies in the possibility of forming complex
meaning representations by way of embedding components into each other. Let me
give you two more examples. The first one illustrates how selectional restrictions can
be accounted for. The verb drink requires the object to be liquid. Jackendoff gives the
analysis in (18a), simplified in (18b):
(18) a.
drink:
[Event CAUSE ([Thing __ ]i , [Event GO ([Thing LIQUID]j ,
[Path TO ([Place IN ([Thing MOUTH OF ([Thing __ ]i)] )] )] )] )]
b.
drink:
CAUSE ( __ i , GO (LIQUIDj , TO (IN-MOUTH-OF __ i )))
The structure of the meaning of drink is very similar to the structure of give in
11
(17b). The place for the ‘moving’ thing, i.e. the object that is drunk, is not just left
open and categorized as ‘thing’ but specified by the concept LIQUID. The subscript ‘j’
11 The subscripts i and j are so-called referential indices. Identical indices indicate identical referents.
Thus, the indexing in (18a) indicates that the drinking agent causes the object to ‘go’ into their
own mouth. According to this analysis, the verb drink cannot be used for the action of pouring
something into someone other’s mouth, which is correct.
Meaning components
241
indicates that the argument is not implicit but has to be specified in the sentence.
The specification of the object term is then ‘fused’ with the specification LIQUID (recall
5.7.2 for the process of fusion).
The other example is Jackendoff ’s analysis of the verb butter. The verb is obviously
semantically related to the noun butter: it means ›spread (something) with butter‹.
In this concept, ›butter‹ is an implicit argument of an underlying predicate (›spread
with‹). Jackendoff gives the following analysis for butter in a sentence like Harry
buttered the bread, simplified in (19b):
(19) a.
butter:
[Event CAUSE ([Thing __ ]i ,
[Event GO ([Thing BUTTER], [Path TO ([Place ON ([Thing __ ]j )] )] )] )]
b.
butter:
CAUSE ( __ i, GO ( BUTTER , TO (ON ( __ j ))))
The sentence Harry buttered the bread can then be represented as in (20):
(20)
butter:
CAUSE (HARRY, GO ( BUTTER , TO (ON (BREAD))))
There are very many noun-verb pairs of this type in English. Of course, not all such
pairs can be analysed in exactly the same way as butter/butter: think of hammer/
hammer or host/host; obviously there are various patterns of decomposition and
relations between the noun and the verb.
This may suffice as a sketch of the basic mechanisms and the explanatory power
of what I would like to call ‘formula’ approaches to lexical decomposition. They
represent lexical meanings by means of predicate logic formulae which are able to
combine various types of meaning components in different ways. Both Dowty’s
and Jackendoff ’s approaches are parts of elaborated theories dealing with much
more ambitious questions than just lexical decomposition. In both theories,
decomposition serves as a means for explaining the combinatorial properties of the
analysed expressions: how they can be combined with other elements in a sentence
and how their meanings interact. For example, Jackendoff ’s approach also allows
a derivation of the thematic roles of verbs from their meaning descriptions and a
theory of linking.
Although their respective theoretical foundations and objectives are different,
they use similar decompositional structures. Both work with abstract semantic
primitives that have one or more arguments and thereby allow the embedding of
other components. It is therefore possible to build up complex structures capable
of capturing the meanings or multi-place predicate expressions. Unlike in feature
combinations in BFA, these structures can express conditions on arguments other
than the referential argument.
An apparent restriction of the two approaches is the fact that they are only applied
to verb meanings. There are, however, theories in a similar vein that deal with other
lexical categories (cf. Bierwisch and Lang 1989 for adjectives, and Pustejovsky 1995
for his unified ‘qualia’ approach to nouns and verbs). What all formula approaches
242
Understanding semantics
have in common is a formal apparatus (technically, a ‘formal language’) with a
vocabulary of primitives and a precise syntax and semantics of its own that allows
the formation of formulae as meaning representations.
9.5 SEMANTIC PRIMES: WIERZBICKA’S NATURAL SEMANTIC
METALANGUAGE
The last approach to decomposition to be discussed was developed by Wierzbicka.
Her aim is the development of a system that allows the description of all meanings in
all languages by means of a strictly limited set of semantic primitives (or ‘semantic
primes’, like ‘prime numbers’, in her terminology). She calls this system ‘Natural
Semantic Metalanguage’, or NSM for short. NSM consists of a set of semantic
primitives and a syntax for their combination. Let me give you her definition of the
meaning of envy as a first example (Wierzbicka 1996, p. 161):
(21) X feels envy =
sometimes a person thinks something like this:
something good happened to this other person
it did not happen to me
I want things like this to happen to me
because of this, this person feels something bad
X feels something like this
The defining part contains the semantic primitives SOME, PERSON (= SOMEONE), THINK,
(SOME)THING, LIKE, THIS, GOOD, HAPPEN, OTHER, NOT, I, WANT, BECAUSE, FEEL, BAD. They are
combined into simple sentences following the syntactic rules of NSM.
For Wierzbicka and her followers the careful choice of semantic primitives is of
central concern. An item qualifies as a semantic primitive if
∑ it is undefinable, i.e. its meaning cannot be defined in terms of other expressions
∑ it is universal, i.e. it is lexicalized in all natural languages
Both requirements can only be met approximately. For example, many concepts
appear to be interdefinable, for example ›eye‹ and ›see‹ (›eye‹ = ›part of the body
used for seeing‹ and ›see‹ = ›perceive with the eyes‹). Which one is primitive can only
be decided in practice. Universality is, of course, not verifiable. But the researchers
adopting the approach did investigate a great number of languages, from Europe,
Africa, Asia and Australia. So the primitives they sorted out on this basis are good
candidates for universal primitives. Wierzbicka (1996, pp. 35f., 73f.) lists fifty-five
primitives, here rearranged; the list has been constantly growing:
(22)
I, YOU, SOMEONE (PERSON), SOMETHING (THING), PEOPLE; WORD
THIS, THE SAME, OTHER, PART (OF), KIND (OF)
ONE, TWO, MANY (MUCH), MORE, VERY, ALL, SOME (OF)
Meaning components
243
THINK, KNOW, WANT, FEEL, SEE, HEAR, SAY
GOOD, BAD; BIG, SMALL
DO, HAPPEN; MOVE, THERE IS, (BE) ALIVE
WHEN, BEFORE, AFTER; A LONG TIME, A SHORT TIME; NOW
WHERE, UNDER, ABOVE; FAR, NEAR; SIDE; INSIDE; HERE
NOT, CAN; IF, BECAUSE, LIKE, IF … WOULD, MAYBE
To the extent that the primitives are universal, every language is supposed to
possess a corresponding set of lexemes, for example the words I, you, someone, etc.
in English. If these lexemes are polysemous, they have to be taken in the relevant
reading. For example, the English verb want, when considered the lexicalization of
the NSM primitive WANT, must be taken as meaning ›wish‹, not ›lack‹. In addition,
every language is supposed to possess syntactic rules for combining the lexical
counterparts of the semantic primitives to the same effect as the primitives are
combined in NSM. Thus the NSM meaning definitions can be translated into
particular languages.
A few more examples may serve to show the kind of decompositional analysis
possible in this framework. ABOVE and FAR are primitives that allow, among others,
the definition of the meaning of sky:
(23) sky (Wierzbicka 1996, p. 220)
something very big
people can see it
people can think like this about this something
it is a place
it is above all other places
it is far from people
Once this concept is available, one can proceed to define the meaning of sun and then
of blue:
(24) sun (Wierzbicka 1996, p. 220)
something
people can often see that something in the sky
when this something is in the sky people can see other things because of this
when this something is in the sky people often feel something because of this
(25) X is blue = (Wierzbicka 1996, p. 309)
at some times people can see the sun above them in the sky
when one sees things like X one can think of the sky at these times
Let me finally illustrate how the NSM approach works for the kind of examples
treated by Dowty and Jackendoff. Goddard gives the following NSM definition of
causative break, a verb that is of the same type as openTV, discussed above.
244
Understanding semantics
(26) xPerson break(s) y (e.g. Howard broke the window) =
a.
x does something to y
b.
because of this, something happens to y at this time
c.
because of this, after this y is not one thing any more
12
Along the lines of Dowty and Jackendoff the verb would be analysed as in (27a) and
(27b), respectively:
(27) a.
x breaks y:
CAUSE(x, BECOME(broken(y)))
b.
x breaks y:
CAUSE(x, GO(y, TO(BROKEN)))
The CAUSE component relates to an action of x that causes the change of y into the
state of being broken. (The predicate broken/BROKEN in the two formulas is chosen
just for convenience; for a non-circular definition it would have to be replaced
appropriately.) This action, and the fact that it affects y, is expressed by the first two
clauses in (26). The change of state corresponding to BECOME(broken(y)) in (27a)
and GO(y, TO(BROKEN)) in (27b) is expressed in a different but equivalent way in the
third clause of the NSM definition. Thus the analyses are very similar, although they
look different.
Unlike the other approaches to decomposition, Wierzbicka’s is not restricted to
the analysis of certain word classes. It even offers meaning descriptions of emotion
words like envy or colour terms or interjections (like yuk!), which other theories
are unable to decompose. It is able to capture all sorts of meaning relations such
as those that make up taxonomies (captured by means of the primitive KIND OF),
or meronymies (captured by PART OF). More than any other approach, it allows the
definition and comparison of word meanings in different languages. But the NSM
method has its shortcomings too. First, it must be stated that NSM descriptions
suffer from a certain lack of precision, as the primitives of NSM are very general and
the definitions therefore rather vague. For example, the definition for envy given in
(21) would also apply to self-pity or feeling neglected, etc. and the definition given
for break also covers cut, split, shatter and similar verbs of destruction. For many
NSM definitions, it is questionable whether they are sufficiently specific to capture
the meaning of the terms defined. While this deficiency might be overcome by more
elaborate definitions, a further drawback cannot be remedied: the approach is unable
to explain meaning relations that hold between the primitives such as the antonymy
of big and small, good and bad, far and near, or the directional opposition between
after and before, or under and above. Their meaning relations are ‘non-compositional’
in this theory, i.e. they cannot be explained on the basis of meaning definitions in
NSM, because these expressions have the status of undefinability.
12 The example is taken from Goddard (1998, p. 281). The original definition is in the past tense (‘X
broke Y …’). The subscript PERSON on the variable X is used to distinguish this variant of the verb
from the one in which the subject is an event, as in the storm broke the window.
Meaning components
245
9.6 SUMMARY AND EVALUATION OF THE APPROACHES TO
DECOMPOSITION
In this chapter we have taken a look at different theoretical approaches to meaning,
in particular lexical meaning. First, essentials of the structuralist approach were
presented. Although no longer a leading paradigm in present-day linguistics, it still
is very important. The method of analysing language as an autonomous system,
closed in itself, helped to establish linguistics as a science of its own. It also provided
students of language and languages with the basic technique of determining the
contrastive and combinatorial properties of units at different levels (from phonemes
to sentences) in a systematic way. These methods are still valid.
Four approaches to decomposition were then introduced: BFA, which was developed
in the tradition of structuralism, the formula approaches of Dowty and Jackendoff,
and Wierzbicka’s NSM.
BFA, which is still enjoying some popularity, is the one with the most severe
limitations. Being confined to a very simple pattern of decomposition (i.e.
conjunctions of one-place predicates), it is unable to account for all other types of
meaning structures and, consequently, unable to capture meaning relations other
than simple incompatibility, complementarity and hyponymy. Due to its insufficiency
for semantic analysis in general it is of equally limited use for language comparison.
The two formula approaches perform equally well, due to their rich structures of
meaning representations. They do not lay too much emphasis on the basicness of their
primitives. Nevertheless, central components of their theories, such as CAUSE, BECOME
and GO, are plausible candidates for universal meaning components. Their strength
lies in the explanation of the compositional properties of verbs. Wierzbicka’s is the
only approach that makes reduction to basic meanings a central issue and focuses on
language comparison. It is, however, less clear how the somewhat amorphous NSM
definitions can be used for rigidly deriving meaning relations between lexical items,
including the basic logical relations, and it remains open to what extent it is able to
account for the compositional properties of lexical items. These drawbacks are all
due to the lack of precision of the meaning descriptions.
There are, of course, a number of other approaches to decomposition in addition to
those discussed here. One approach which has gained some influence is Pustejovsky’s
‘Generative Lexicon’. I consider the general meaning format of this theory too
complex for a discussion in this book. In a later chapter, we will discuss a proposal
which was developed in cognitive psychology – ‘frames’ in chapter 12. The general
format of frames is at the same time simple, general and very powerful. I will not
introduce it yet, as it requires a certain background in cognition.
246
Understanding semantics
EXERCISES
1. In which way is a linguistic unit integrated into, and defined by, the language
system? What kinds of relations and properties determine the character of a
linguistic unit?
2. What is the meaning of a lexeme according to the structuralist doctrine?
3. What is meant by the terms syntagmatic relations and paradigmatic relations?
What, in particular, are syntagmatic and paradigmatic meaning relations? What
do selectional restrictions have to do with these relations?
4. What is the difference between a lexeme and a morpheme?
5. Give a feature analysis of the nine members of the system of German personal
pronouns (Table 4.1). Describe the properties of the referent which the features
you use are to capture. If a feature is not relevant for a certain case, leave the feature
open; the value ‘–’ is reserved for those positively negative cases where ‘+’ is to be
excluded. Reread the section on the meaning of personal pronouns in 4.1.1.
6. How many and which morphemes make up the following words? For each
morpheme (i) determine its category (N, V, A) if it is basic, or its input and
output category if it is an affix and (ii) find two more words that contain the same
morpheme in the same function. Indicate the structure of the word by bracketing.
a. localization
b. unsuccessful
c. regrettable
7. Which types of semantic features are distinguished in BFA? How do they differ?
Try to give examples of expressions for which the different kinds of features can
be considered meaning components.
8. Find ten more pairs or triplets of stative, inchoative and causative verbs (or
adjectives, for the stative part). Check the entailments illustrated in (9).
9. Assign the three types of verbs in Dowty’s examples to the aspectual classes
mentioned in 6.2.
10. Discuss the advantages of formula approaches over BFA.
11. Discuss the differences between NSM and other approaches to decomposition.
FURTHER READING
Saussure himself on structuralism (= Harris 1983), also Lyons (1977, ch. 12);
more comprehensively Matthews (2001). Foley (1997, ch. 4) on structuralism and
its relation to linguistic anthropology. Dillon (1977) for an application of feature
semantics. Lyons (1977, ch. 9.9) for a criticism of feature semantics. Dowty (1979)
requires a background in Montague Grammar, Jackendoff (1990) is somewhat
easier to read, but cannot be fully understood without knowledge of Chomsky-type
Meaning components
247
language theory and grammar. For a short introduction, see Riemer (2010: 8.1); more
elaborately Jackendoff (2011). Goddard (1998) gives a survey of decompositional
approaches and a very readable introduction into NSM analysis. Wierzbicka (1996)
offers the essentials of NSM theory and its relation to semantic and cultural studies.
For recent surveys on decomposition see Engelberg (2011a, 2011b). Cann (2011)
offers a survey of structuralist semantics.
10
?
Meaning and language comparison
If structuralism is right in assuming that every individual language is a system of
its own, we should expect that languages can be very different, that they have terms
for different things or different ways of naming the same things and that similar
situations are expressed in different ways and from different perspectives. But
how big are the differences between languages? Are all languages essentially alike,
allowing the expression of the same thoughts, the communication of the same things,
only differing in their respective ways of formulation? Or are languages different to
an extent that it may be altogether impossible to express in one language what can
be expressed in another? Are the semantic systems completely arbitrary, or are they
constrained by universal principles?
We will start with simple examples that illustrate problems of translation. A closer
look at different ways of expressing that someone has a headache will take us a bit
deeper into the matter of semantic comparison. The second part will be devoted
to studies of colour term systems in various languages. These studies played an
important role not only in the understanding of the relationship between languages
but also in the development of semantics in general.
10.1 TRANSLATION PROBLEMS
Everybody who tries seriously to learn a foreign language will sooner or later realize
how different languages are, not only with respect to grammar and pronunciation
but also with respect to their vocabulary and how it is organized. When we naively
begin to learn a new language, we will probably start with the unconscious working
hypothesis that for each word in our native tongue there is a corresponding word
in the target language. But the more we get into the new language, the more this
hypothesis will crumble. People who have managed to master a foreign language to
a degree close to native-speaker competence would probably say that the semantic
equivalence of two lexemes is the exception rather than the rule.
There are different types of mismatch. One language may have two or more words
where another language has only one.
(1)
English
Japanese
mouse rat
nezumi
finger toe
yubi
water
mizu
yu
›cold w.‹
›warm w.‹
Meaning and language comparison
249
But the relations can be more complicated. Consider the following lexical field in
English and Japanese, and German for comparison:
(2)
German
English
Japanese
Wald
wood1
mori
Holz
wood2
ki1
Baum
tree
ki2
English has a count noun wood1 and a mass noun wood2. A wood1 consists of trees,
while wood2 is the substance trees largely consist of. Hence, the two meanings can
be clearly distinguished on linguistic grounds. Likewise, the Japanese meanings of
ki1 and ki2 are clearly different. It follows that neither of the two English expressions
matches with either of the two Japanese words. They only match in one meaning
variant, respectively. In general, the ubiquitous polysemy of the vast majority of
words suggests that (with the exception of number terms) there are almost no two
expressions from different languages that have the same overall meaning (cf. 8.1 for
the parallel problem of total vs partial synonymy within one language).
In (2), the meanings of the English, German and Japanese terms at least (roughly)
match in some of their possible readings. But in many cases, correspondences turn
out to be more complicated. Even apparently basic and universal concepts such as
›eat‹ and ›drink‹ are differently lexicalized. A first glance at dictionaries will give us
the following correspondences:
(3)
German
English
Japanese
essen
eat
taberu
trinken
drink
nomu
But as was mentioned in 9.2.2, the German terms essen and trinken are reserved
for people, unlike their English and Japanese counterparts. Even if we restrict the
verbs to human agents, they do not match perfectly. In English and German, a
(thin) soup can be either ‘eaten’ (essen) or ‘drunk’ (trinken), depending on whether
a spoon is used or the soup is directly taken into the mouth. Japanese use the verb
nomu regardless of whether the soup is drunk (which is the traditional technique)
or consumed with a spoon. One might now think that the crucial criterion for using
nomu is that the object be liquid. But this is not the case. The verb is also used for oral
medicine, including pills. The crucial point appears to be that the object is directly
swallowed without being chewed. This is a different criterion from the one that
regulates the use of English drink or German trinken. We cannot ‘drink’ a pill and we
can eat a soup even if we do not chew it.
Still, these are cases where for every context appropriate translations are possible.
But often a certain term in one language does not correspond to any term at all in the
other language. This may be due to the fact that the things the term refers to simply
do not exist where the other language is spoken, e.g. plants or animals or meals or
artefacts or social institutions. But it also happens in semantic areas that are shared.
Consider the area of working. English has a verb to work which, like German arbeiten,
covers a broad range of activities, paid work as well as unpaid work (e.g. in one’s own
250
Understanding semantics
kitchen or garden). The work may be physical or intellectual work such as reading
this book, learning German or studying linguistics. Japanese has no term that covers
all these activities. The closest equivalent to the verb work is hataraku. But it does
not cover intellectual work. On the other hand, Japanese has a verb asobu that can
be used for any kind of activity that is not work: playing, any kind of entertainment
or leisure including doing nothing at all. No simple verb with this meaning exists in
English or German.
An insurmountable problem for adequate translations can be posed by differences
in social meaning. Imagine an American movie of the type ‘boy meets girl’. John, a
yuppie real estate agent, meets Mary, a tough customer, and falls in love with her. If
the dialogues are translated into German, the translators have to decide whether John
and Mary address each other with Sie or du (2.3). They will have to start with Sie,
because that would be adequate, to a German audience, for the business relationship
between them, and will end up addressing each other with du once they have become
intimate. The problem for the translator will be: when do they have to switch from
Sie to du? As German personal pronouns of address, unlike English, carry social
meaning, speakers of German are forced to express distinctions which English does
not provide.
With Japanese, things are much more complex. First, there are several expressions
available to render English I and you, respectively. Each one has a different social
meaning, i.e. indicates different social relations. Second, in very many cases the use
of pronoun-like expressions for self-reference and address is altogether inadequate
in Japanese. For example, in a dialogue between mother and son, the son would
normally address his mother with ‘Mrs Mother’ (okāsan), not with any variant of
‘you’. Referring to himself he would either use a variant of ‘I’ or his own name (!),
but not ‘Mr Son’. His mother would refer to herself preferably as ‘Mrs Mother’ and
address her son with a variant of ‘you’ or his name. For example, little Taro could ask
his mother something of the form (4a) and his mother could answer (4b).
(4)
a.
‘Will Mrs Mother buy Taro an ice-cream?’
b.
‘Mrs Mother will buy Taro an ice-cream.’
Compare these forms to the adequate ways of expression in Western languages with
the characteristic switch between the pronouns I and you for the same arguments of
the predicate.
‘Will you
b.
‘
I
buy me
an ice-cream?’
⇔
a.
⇔
(5)
will buy you an ice-cream.’
For Western ears, the sentences in (4) are formulated as though they were said about
third persons. In fact, (4a) can as well be used, for example, by Taro’s elder sister for
asking a question of her father ABOUT Taro and her mother, and her father could use
(4b) as an answer.
Meaning and language comparison
251
Similar rules govern the forms of address and self-reference outside the family.
Students would normally address their teachers as ‘Teacher’ (sensei) or ‘name +
Teacher’ (e.g. Yamamoto Sensei), but not with ‘you’. A further crucial difference
between English and Japanese is the fact that in Japanese pronouns are normally
completely omitted. For example, in informal speech the question ‘Will you buy me
an ice-cream’ would be formulated as ‘Ice-cream buy-give?’ Omitting pronouns in
subject position is not allowed by the grammar of standard English. Due to all these
fundamental differences between Japanese and English, a direct, literal translation
of Japanese dialogues into English would be inadequate, or even ungrammatical. If,
however, the translation were in accordance with the natural means of expression
in English, it would not convey the ways in which social interaction is handled in
Japanese by choosing terms of address and self-reference.
10.2 HEADACHE, INTERNATIONAL
10.2.1 Grammatical constructions
Deeper in the language system we find differences in grammar. Examples are
abundant once we compare historically unrelated languages such as English and
Japanese. In terms of structural differences, Japanese is not as ‘exotic’, in comparison
with English, as many other languages, for example the native languages of the
Americas, Australia or Papua New Guinea. But with respect to some semantic
phenomena Japanese grammar differs considerably from English or other Western
languages.
For an illustration, we will compare the ways in which having a headache is
expressed in some European languages and in Japanese. A headache situation
involves three ingredients: (i) an experiencer E who feels the headache; (ii) the
sensation S, a pain that E feels; (iii) the body part B that aches: the head. Languages
express the situations by a variety of grammatical patterns in which these three
ingredients figure in different ways:
(6)
a.
English:
1
I have a headache
The English construction ties S and B, the pain and where it is felt, together into
one concept ›headache‹. What is expressed by have (6a) is that the headache is in
some way associated with E. Note that the relation would be much more specific if
one were to say I feel a headache instead. The English construction is paralleled in
German, with the slight difference that the sensation is in the plural, but there is also
the synonymous ich habe Kopfweh (singular, mass). The standard French phrase too
is similar.
1 This is the most common construction. Alternatively, headaches can also be expressed by
the variants my head is aching or my head is hurting me, which are similar to the Hungarian
construction (6f) and German 2 (6d), respectively.
252
(6)
Understanding semantics
b.
German 1:
ich
hab-e
1S.NOM
have-1S.PS
lit. ‘I have headaches’
Kopfschmerz-en
headache-PL
c.
French:
j’
ai
have.1S.PS
1S.NOM
lit. ‘I’ve pain at the head’
mal à
pain at
la
DEFART
tête
head
French has a possessive construction with three argument terms, E in the subject
position, S in the object position and B in a locative PP.
There is an alternative German construction that is similar to the standard Russian
way of expressing the situation.
(6)
d.
German 2:
mir
tu-t
der
1S.DAT ache-3S.PS DEFART
2
lit. ‘[to] me aches the head’
Kopf
head.S.NOM
e.
Russian:
u
menya
at
1S.GEN
lit. ‘at me aches head’
golova
head.SG.NOM
bolit
ache.3S.PS
weh
ache
The main predicate, i.e. the verb, directly expresses the sensation itself, not some
abstract association of E with S (the German verb wehtun ‘ache/hurt’ splits into
two parts in this sentence type). The subject specifies the aching part B, while
the experiencer term is put in an indirect object or PP position. Thus the head is
what the statement focuses on, not the experiencer. A similar variant is used in
Hungarian:
(6)
f.
Hungarian: fáj
a
fej-em
3
DEFART
head-POSS1S, SG
aches
lit. ‘aches the head-of-mine’
Here the verb is used as a one-place predicate with B as its only argument. B, in turn,
is specified with the possessive suffix as the B of E. B and E are tied into one.
The usual Japanese equivalent of the sentences in (6a–f) is (7a):
(7)
a.
Japanese
4
atama
ga
ita-i
head
NOM
feel hurting-PS
lit. ‘head feels hurting’
2 The two German constructions are not completely equivalent, the latter being more generally
applicable to any kind of pain felt in or at the head, but that does not matter for the present
discussion.
3 -em/-am is a noun suffix that indicates possession of the speaker and singular number of the noun.
4 In order to keep the grammatical structures of the sentences as transparent as possible, Japanese
examples are given in the so-called plain style lacking all formality markers (cf. 2.3.2).
Meaning and language comparison
253
Itai is a so-called verbal adjective (VA). Verbal adjectives carry a tense ending and
can form the VP in predicative use. They function like a copula plus predicative
adjective in English. Interestingly, the VA itai has only one argument, B, specified by
the subject of the sentence. But what about the experiencer E? How do I know from
a sentence like (7a) whose head it is that hurts? The really interesting thing about
Japanese is that one does know from the sentence, and this is a point where Japanese
differs radically from English, German, French, Russian, Hungarian and many other
languages.
The answer is simple: the experiencer in the case of (7a) is the speaker. There are
only two exceptions. (i) The sentence could, with appropriate intonation, also be
used as a question; it would then mean ›Do you have a headache?‹. (ii) In a literary
text with a central protagonist, through whose eyes and mind the narration is
told, it could relate to that person. The crucial point is that itai means a subjective
sensation, not an objective phenomenon. Subjective sensations, like the feelings
of pain, itching, appetite, grief or longing, only exist for the experiencer. (7a) says
something like ‘the head feels aching’. E is the only one who can state this or answer
the corresponding question. A specification of E can be explicitly added to the
sentence, as in (7b, c), but it has to be semantically redundant. The experiencer
specification is added as a topic of the sentence, not as a complement of ita-i.
(7d) is unacceptable, except for the special literary reading mentioned. (TOP is the
Japanese topic marker, which is hard to translate. It means something like ‘as for …’,
but is less heavy.)
(7)
b.
c.
d.
Watashi
I
TOP …
wa atama ga itai.
Anata
you
TOP …
wa atama ga itai?
*Mary wa atama ga itai.
‘As for me, head feels hurting’
‘As for you, head feels hurting?’
5
‘As for Mary, head feels hurting.’
Sentence (7a) is syntactically complete. Words such as itai have an implicit
argument for the direct experiencer. Thus Japanese represents the third logical
possibility of tying together two of the three ingredients, this time S and E. It
is worth noting that, wherever two ingredients are tied together, one of them
is dominant, and the expression denotes the dominant one. Headache and
Kopfschmerzen denote S and locate it at B; fejem denotes B and assigns it to E; itai
denotes S and assigns it to E.
The five patterns are compared in Fig. 10.1. A rectangle indicates the central
predicate expressed by the verb, circles represent its arguments; the subject is shaded.
Where two ingredients are tied together, the dominating one is indicated by larger
type.
5 The sentence may, perhaps, be read as meaning something like ‘when I think of Mary, I’m getting
a headache’, but it cannot mean ‘Mary has a headache’.
254
Understanding semantics
Figure 10.1
Headache constructions
English, German 1
E
have
French
E
have
S-B
S
LOC
B
S
B
Hungarian
S
B-E
Japanese
E-S
Russian, German 2
E
B
These are not all the types that are logically possible. For example, our sample
contains no constructions with B or E in predicate position. These, however, are
unlikely constructions. Given the typical function of verbs as denoting events
and states of temporary nature as opposed to nouns that tend to denote the more
permanent objects of experience, we will expect that neither E nor B will figure as the
verb of the construction. But there remain further plausible possibilities. Instead of
a verb of the type have there might be constructions using be (‘a pain is in my head’)
that put the sensation into the subject position. Also there may be languages where
the pain is expressed by a two-place predicate with E in subject position (‘I ache-feel
my head’). And there are yet other languages that tie all three ingredients together
into one complex predication of type S-B-E.
10.2.2 Evidentials
As we have seen, the Japanese construction does not provide a syntactic position for
the specification of the experiencer. How, then, is it possible to express that someone
else has a headache? In the other languages considered, we can simply replace the
experiencer ‘I’ by an appropriate NP and change the verb form if necessary. The
resulting expressions are perfectly analogous to the 1st person variants (with the
exception that the Hungarian construction needs an additional dative position for
the possessor NP if it is not pronominal; the result is a construction very similar to
German 2).
Meaning and language comparison
(8)
a.
English:
Ken has a headache.
d.
German 2:
Ken
Ken.DAT
tut der Kopf weh.
aches the head
f.
Hungarian:
Ken-nak
Ken.-DAT
fáj a fej-e.
aches the head-POSS3SG,SG
255
Strictly speaking, Japanese has no way of saying this. But of course it has means
of talking about third persons’ headaches. The ways in which it can be done are
interesting because they reflect a point that usually goes unnoticed to native speakers
of languages such as English, Russian or Hungarian: we can never really KNOW that
someone else has a headache, not in the sense of knowing as the experiencers
themselves do. Every person has their own body and their own sensations which
cannot be immediately perceived by anybody else. Knowledge about someone else’s
sensations is necessarily second-hand and of a different quality than knowledge from
sources that are immediately accessible. Consequently, statements that are based on
second-hand knowledge are less reliable. In Japanese, this difference is obligatorily
expressed; second-hand knowledge is always marked as indirect evidence. (9) shows
the three most common possibilities for the case of others having a headache:
(9)
a.
b.
c.
John
wa
atama
J.
TOP
head
‘John says he has a headache’
John wa atama ga
…
‘John seems to have a headache’
ga
NOM
ita-i-sō
ache-PS-EVD1
ita-sō
ache-EVD2
da.
COP.PS
da.
COP.PS
John wa atama ga
ita-gat-te i-ru.
…
ache-EVD3-PROG-PS
‘John is displaying symptoms of having a headache’
In each case, the NP John is added to the basic construction as a topic and/or subject
marked with the topic particle wa. In (9a) the nominal suffix -sō plus the copula
verb da is added to the full form of the adjective itai, including the tense ending
-i, to render the sense of John saying he has a headache. In (9b) the nominal suffix
-sō plus copula is added to the bare stem ita- of the adjective, meaning something
like the English seem to construction. The third pattern puts the experiencer into a
more active role of displaying symptoms of, or behaving like, having a headache. The
construction involves a verbal suffix -garu that turns a verbal adjective into a verb. It
is here used in its progressive present tense form -gatte iru.
The grammar of Japanese strictly distinguishes between statements that we
are entitled to make on the basis of first-hand knowledge, and statements for
which we rely on second-hand evidence. The so-called evidential constructions
in (9) are widely used. There are about seventy adjectives for emotions, feelings,
sensations and mental states that are used in the same way as itai, e.g. for feeling
256
Understanding semantics
happy (ureshii), sad (kanashii), lonesome (sabishii), cold (samui) or hot (atsui), for
liking (suki) and hating (iya, kirai), for being afraid of something (kowai), finding
6
something cute (kawaii) or delicious (oishii). One can only say of something that it
is ‘oishii’ if one has actually tasted it. Something that appears delicious would be said
to be ‘oishi-sō’ (cf. (9b)). This type of adjectives also includes the volitional forms
of verb that are formed with the suffix -tai; it turns verbs into verbal adjectives
rendering the meaning ›want to V‹. One can say ‘I want to go’ (iki-tai, from iku ‘go’),
but not with the same construction ‘John wants to go’ (*John wa iki-tai). Again, one
would have to use an evidential construction meaning ‘John seems to want to go’
(John wa iki-ta-sō da).
The comparison of the way in which bodily sensations, feelings and wishes are
expressed offers an insight into the nature of knowledge. Languages such as Japanese
force upon the language user the distinction between first-hand and second-hand
evidence. This insight could not be obtained from the analysis of languages such as
English, which treat feelings of oneself and the feelings of other persons alike and in
the same way as any kind of objective observation, e.g. John has red hair.
10.3 RELATIVISM AND UNIVERSALISM
How different are languages? The possible positions regarding this question range
between two extremes. One is called universalism. According to the universalist
position, all languages obey the same principles. The structure of every language is
some variant of universal grammar. And universal grammar is part of the human
genetic equipment. Likewise, the cognitive system is genetically determined. For
biological reasons, all human beings perceive themselves and their environments
essentially in the same way; they form the same kind of concepts and organize them
into the same kind of complex model of the world. Consequently, languages can
only differ within a limited range of variation. Considerable differences between
languages do exist. The environment in which a linguistic society lives and the
culture it has developed will be reflected in the language. For example, each language
community will have a particularly elaborate vocabulary in those areas of life that
are of central importance. But ultimately, the universalist would argue, all languages
make use of the same mental resources. They differ only in the way in which these
resources are used.
The opposite position is known as linguistic relativism. To an extreme relativist,
each language is radically different. Due to its unique grammar and its uniquely
structured lexicon, it represents a unique way of talking about the world and
corresponds to a particular way of thinking. Each language represents, and creates,
a worldview of its own. The relativist position is connected with the names of two
American linguists who worked on indigenous North American languages in the
first half of the twentieth century, Edward Sapir (1884–1939) and Benjamin Whorf
(1897–1941). The following passage from Whorf is often quoted as a formulation of
6 See Martin (1975: 361) for a survey.
Meaning and language comparison
257
the so-called Sapir-Whorf hypothesis (Whorf 1956: 212–14, the original date of
publication is 1940):
Formulation of ideas is not an independent process, strictly rational in the old
sense, but is part of a particular grammar, and differs, from slightly to greatly,
between different grammars. We dissect nature along lines laid down by our
native languages. The categories and types that we isolate from the world of
phenomena we do not find there because they stare every observer in the face;
on the contrary, the world is presented in a kaleidoscopic flux of impressions
which has to be organized by our minds – and this means largely by the linguistic
systems in our minds. We cut nature up, organize it into concepts, and ascribe
significances as we do, largely because we are parties to an agreement to organize
it in this way – an agreement that holds throughout our speech community and is
codified in the patterns of our language. The agreement is, of course, an implicit
and unstated one, but its terms are absolutely obligatory; we cannot talk at all
except by subscribing to the organization and classification of data which the
agreement decrees.
Applied to the Japanese evidentials, this view would mean that the members of the
Japanese speech community implicitly agree that second-hand evidence of emotions
and perceptions is to be distinguished from first-hand knowledge. This hypothesis
is plausible, and there are many other phenomena where linguistic structures can be
linked to the culture of a speech community. Take, for example, the Japanese terms
for siblings which, in usual colloquial Japanese, force the language user to distinguish
between younger and elder siblings. This trait of Japanese is not accidental. Japanese
society is extremely hierarchical. No two persons that are in some social relation to
each other are of the same rank. For example, those who are older rank higher than
those who are younger and men rank higher than women. The ranking rules are
deeply rooted in social behaviour AND in language use and structure. For example,
while siblings in Western societies address each other mutually by their first names,
the mode of address between Japanese siblings of different age is asymmetric. Elder
siblings are addressed by their title, as it were, namely the polite versions of ane and
ani (onê-san and onî-san, respectively, the latter literally meaning ›Mr elder brother‹),
but younger siblings are just called by their names.
The domain of sexuality provides another case of correspondence between culture
and language. Due to a long history of sexual taboo enforced by the Christian church,
European languages exhibit remarkable lexical peculiarities. The taboo is reflected
in the massive use of circumscriptions, i.e. indirect ways of expression. The words
for ›bra‹ mentioned in 7.6.1 may illustrate the point. The French expression soutiensgorge, literally meaning ›throat support‹, avoids mentioning what is really ‘supported’
by referring instead to a neighbouring part of the body that is not taboo. The English
term bra is even more indirect. It is an abbreviation of brassiere (a French loan word
deriving from bras ›arm‹) and originally means a short vest or top. The Spanish term
sujetador (lit. ›subjugator‹) reflects the sexual taboo in not mentioning the object of
subjugation and in a negative attitude towards it as something to be subjugated. By
contrast, the Tok Pisin notion ›prison of the breasts‹ (kalabus bilong susu) mentions
258
Understanding semantics
the breasts directly and reflects the loss of freedom that came with this particular
7
item of Western culture and the foreign taboo that made it necessary.
Such phenomena, however, do not directly support the strong relativist position. Is
it really language that forces the Japanese into their hierarchical social thinking, or
Westerners into observing sexual taboos? Rather, it appears, that language REFLECTS
social structure and cultural standards. It is true that it is not possible to talk directly
about sexuality if the language does not provide the appropriate words. But the
so-called sexual revolution in Western culture in the last decades of the twentieth
century has shown that speech communities rapidly develop socially acceptable
expressions when the taboo loses force.
It must also be observed that particular grammatical traits of a language need
not influence the worldview of its users. For example, many European languages
have grammatical gender. In French and Spanish each noun is either masculine or
feminine, in German and Russian nouns are either masculine, feminine or neuter.
Gender determines, among other things, the form of the definite article (le/la in
French, der/die/das in German). It would, however, be absurd to conclude that gender
classification imposes any kind of different worldviews upon the user. The fact that
the German nouns for ‘government’, ‘street’ and ‘banana’ (Regierung, Straße, Banane)
are of feminine gender does not make speakers think of these things as having
8
anything in common, in particular not female sex.
Although these observations point towards a differentiated position somewhere
between universalism and relativism, the central questions remain. How different
are languages? How deep are these differences? What do the observable differences
indicate? Do they correspond to different ways of thinking? The questions are far
from being settled. We will now turn to one field where extensive comparative
research has been done: the field of colour term systems.
10.4 BERLIN AND KAY’S INVESTIGATION OF COLOUR TERMS
The spectrum of colours, the same for all human beings with normal vision, forms
a continuum with no natural boundaries between the colours: red shades into
orange, pink, purple, brown; orange into yellow and red; and so on. Thus it is to be
expected that languages cut the colours up in different ways, into a different number
of colour terms that denote different parts of the colour space. In fact it was held that
9
languages lexicalize colours arbitrarily. It was the investigation reported in Berlin
and Kay (1969) which changed the scene dramatically.
The study covered approximately a hundred languages from all over the world. For
twenty languages native speakers were interviewed. For the rest, Berlin and Kay drew
7 For many more examples of prudery in language see Bryson (1990, ch. 14) on English.
8 There are slight gender effects for speakers of gender languages. For example, it has been shown
that subjects can remember male proper names for objects better, if the noun used for them is
masculine. But findings such as these can hardly be considered evidence for a different worldview
(see Deutscher 2010, ch. 8).
9 See Berlin and Kay (1969, n. 1, p. 159f.) for quotations on this position.
Meaning and language comparison
259
on previous studies, grammars and dictionaries. They investigated how many colour
terms these languages possess and to which range of colours each term refers. Such
an investigation is only possible if it is restricted to what Berlin and Kay called basic
colour terms (BCT, for short). English, for example, has hundreds of lexicalized
colour terms (think of the vocabulary for the colours of lipsticks, eye-shadows,
fashion accessories, cars, etc.). Only a few of them are basic. The main criterion for
basicness is:
∑ BCTs are not subordinates of other colour terms.
This excludes terms like olive (a shade of green), crimson (red) or chocolate (brown).
In addition, Berlin and Kay used the following criteria: BCTs are simple words, not
compounds or derivations (like greenish, dark green, mint-green, blue-green); BCTs
are not restricted to a narrow class of objects (cf. blond for hair); BCTs (not the
colours they denote) are ‘psychologically salient’, i.e. they ‘[tend] to occur at the
beginning of elicited lists of color terms’; their use is stable across informants and
10
across occasions and they are used by all informants. Berlin and Kay also ruled
out recent loan words (such as aubergine) and colour terms that are derived from
the name of an object characteristically having that colour, such as gold, silver, olive,
etc. Orange originally belonged to this class, but it qualifies as a genuine BCT since
it fulfils the main criterion of being a term on a par with red and yellow rather than
a subordinate of either. As we will see, these criteria yield a set of eleven BCTs for
English.
Berlin and Kay had their subjects perform two tests. First, they elicited the set
of BCTs in their respective native languages. Then the informants were shown a
chart of 329 colour chips (from the Munsell Color Company), 9 for black and white
and different shades of grey, and 320 others, arranged in a field of 8 by 40 chips,
that horizontally range from red via orange, yellow, green, blue, purple to red and
vertically from near-white to near-black (see Fig. 10.2). The original colour chart is
depicted in Berlin and Kay 1969 and in Palmer 1996; it is online at http://www.icsi.
berkeley.edu/wcs/data.html. For each BCT the subjects were asked (a) to point out
the focal colour chip, i.e. the chip that was considered the best example for this BCT,
and (b) to indicate the outer boundary of the range of colours denoted by the BCT.
It turned out that informants of the same language narrowly agreed on the focal
colours, but indicated the boundaries differently. The main result, however, was this:
11
the languages vary in the number of BCTs from two to eleven. The focal colours for
the respective BCTs are not distributed randomly within the field of colours; rather
they concentrate on just eleven colours which were indicated by the informants
consistently, and with very little variation, as the best examples for the BCTs.
Apparently, speakers of different languages, with different numbers of BCTs, agree
what is the best ‘red’/‘rouge’ (French)/‘akai’ (Japanese), etc. These eleven colours
are focal (i.e. ‘pure’, ‘real’, ‘typical’) white, grey, black, red, orange, yellow, green, blue,
purple, pink and brown (indicated by the dots in Fig. 10.2). These are the foci of the
10 See Berlin and Kay (1969, p. 6) for the details.
11 The maximum number was later corrected to twelve (see below).
260
Understanding semantics
English BCTs white, grey, black, red, orange, etc. and of the corresponding words in all
other languages with eleven BCTs. What is much more important is the finding that
the foci of BCTs in languages with fewer than eleven are also among this set of eleven
focal colours. Apparently these colours form a set of fixed points of orientation for
the denotation of BCTs.
Fig. 10.2
Munsell colour chart used by Berlin and Kay
white
yellow
grey
pink
pink
orange
green
red
blue
brown
puiple
red
black
The second important finding is this: languages with the same number of BCTs
have basic colour terms with the same foci, and if a language has n + 1 BCTs, it has
BCTs with all those foci that are denoted in languages with n BCTs, plus an additional
BCT with a new focal colour. In this way the existing systems of BCTs form a sequence
where a BCT for a new focus is added step by step; at the final stage there are eleven
BCTs that have their focus on the eleven focal colours. The first foci in this sequence
are focal white and black (Stage I); the third focus added is red (Stage II); red is
followed by green and yellow in arbitrary order (Stages III and IV), then blue (V),
then brown (VI). For the remaining four foci, grey, pink, purple and orange, there
is no fixed order. According to this order of appearance of BCTs, there will be no
language that, for example, has a BCT for brown, but not for blue and green.
Of course, if a language has just three BCTs with focus on focal white, black and
red, the red term does not have the same narrow denotation as the red term in a
language with eleven BCTs such as English. The smaller the number of BCTs in a
language, the larger are the respective ranges of colours they denote. For example,
the red term in a language with five BCTs (with focus on black, white, red, yellow and
green) will also cover the reddish parts of the denotations of English orange, brown,
pink and purple because there are no more specific BCTs for these hues; lighter
orange and yellowish brown will fall under the yellow term; blue and bluish purple
will be covered by the green term or the black term, if the colours are very dark; a
very light pink will be called ‘white’ (for details see charts in Berlin and Kay 1969).
Thus the range of colours covered by a BCT with a given focus depends on how
many other BCTs the language has and what their foci are. A black term in a system
of Stage I does not have the same meaning as a black term in a system of another
Meaning and language comparison
261
stage; therefore the total number of BCTs is added to the names of the colour terms
in Table 10.1. Berlin and Kay distinguish seven stages of BCT systems; the last stage
includes all languages with eight to eleven BCTs; for BCTs of languages in this last
stage, ‘8’ is added to the BCT names. For example, there are no languages with three
BCTs that focus on yellow, red and blue which might make sense; let alone such
unlikely variants as foci on pink, purple and turquoise, seemingly the most popular
colours with children.
Table 10.1
The seven stages of colour term systems in Berlin and Kay (1969)
Stage
BCTs
I
BLK/2
WHI/2
II
BLK/3
WHI/3
RED/3
IIIa
BLK/4
WHI/4
RED/4
YEL/4
IIIb
BLK/4
WHI/4
RED/4
IV
BLK/5
WHI/5
RED/5
YEL/5
GRN/4
GRN/5
V
BLK/6
WHI/6
RED/6
YEL/6
GRN/6
BLU/6
VI
BLK/7
WHI/7
RED/7
YEL/7
GRN/7
BLU/7
BRN/7
VII
BLK/8
WHI/8
RED/8
YEL/8
GRN/8
BLU/8
BRN/8
GREY/8
or
ORANGE/8
or
PINK/8
or
PURPLE/8
Focus
black
white
red
yellow
green
blue
brown
grey etc.
From the distribution of BCT systems, Berlin and Kay drew the conclusion that
the possible systems form evolutionary stages. The two BTCs of Stage I systems,
WHITE/2 and BLACK/2, are terms for light/warm colours (white, yellow, red) and dark/
cool colours (black, blue, green), respectively. In the transition to Stage II, WHITE/2
splits into WHITE/3 for light colours (white) and RED/3 for warm colours (red, yellow),
while BLACK/3 continues to denote dark/cool colours. When we proceed to Stage III,
in one variant yellow splits from RED/3 to form a category of its own, resulting in a
system with WHITE/4 for light (white), RED/4 for red, YELLOW/4 for yellow, and BLACK/4
still covering all dark/cool colours. Alternatively, green and blue together split from
BLACK/3, forming a new colour category often called ‘grue’ that focuses on either focal
green or focal blue (but never on some blue-green midway between focal blue and
focal green). In the resulting system, yellow is still mostly denoted by RED/4. In the
transitions to later stages up to V, the denotations of multicolour BCTs split further,
e.g. grue into blue and green. The higher BCTs with foci on brown, grey, orange, pink
and purple are not the result of splittings but of establishing terms for hues that
lie between the foci of the six terms of Stage V: grey between white and black, pink
between white and red, orange between yellow and red, purple between red and blue,
and brown between yellow and black.
262
Understanding semantics
The findings of Berlin and Kay triggered a great number of follow-up studies, in
particular the ‘World Color Survey’ where Berlin and Kay’s experiments were applied
12
to 110 unwritten languages. As a result, their findings were slightly revised. The
larger categories of the early stages may have more than one focus, e.g. white, yellow,
and red for WHITE/2 or green and blue for GREEN/4. It turned out that there are certain
types that do not fit into the sequential chain, e.g. languages of Stage III with a BCT
for grue but yellow still going with white, while in Stage II systems yellow is already
separated from white. In addition the order of appearance of the ‘higher’ colours
is less strict: a BCT for brown may appear before a term for blue. Some languages
13
have twelve BCTs. But these are minor modifications. The main tendencies were
14
confirmed:
∑ The denotations of BCTs are best described in terms of focal colours.
∑ The range of colours that a given BCT denotes depends on the total number of
BCTs in the language and on the foci they have, respectively.
∑ There is a limited universal set of eleven (or twelve) focal colours. The best
representatives of BCTs invariably are among these focal colours.
∑ The possible types of system form a sequence – starting with a contrast between
WHITE/WARM and BLACK/COOL – in which the higher systems are extensions of the
lower systems.
10.5 CONSEQUENCES
What follows from the findings on colour terms for the structuralist view of
language? The results certainly prove that the arbitrariness of lexicalization can be
constrained. Still, languages differ largely in their colour terminology. Also, another
doctrine of structuralism has been confirmed: the denotations of terms within the
same field are interdependent and delimit each other: the range of a BCT with a given
focus depends on how many other BCTs compete with it.
And what do the findings mean for the debate between relativism and universalism?
Originally, the results were considered evidence against the Sapir-Whorf hypothesis.
But are they really? It was claimed (Kay and McDaniel 1978) that the universal
constraints on colour vocabularies are directly rooted in the nature of human colour
perception. Even if this is right, this cannot be generalized to most other semantic
15
domains. Most of the things languages have words for are not like colours. For
12
13
14
15
See the World Color Survey homepage www.icsi.berkeley.edu/wcs/.
Russian has two BCTs for lighter and darker blue, Hungarian has two for red.
See Kay and Regier (2003).
Meanwhile, there is evidence that even for colour terms there are what are called ‘Whorfian effects’.
According to Drivonikou et al. (2007), language influences the discrimination of colours: similar
colours represented to the right eye (which is connected to the left hemisphere where the linguistic
lexicon resides) are more easily discriminated if there are two different BCTs for the colours in the
native language of the subjects.
Meaning and language comparison
263
example, we can see animals, but we do not have specialized dog cells and cat cells.
Animal concepts, e.g. the meaning of the English word pig, are not based on sense
data alone such as visual appearance – although visual shape is certainly part of the
concept. In addition, animal concepts are based on cultural aspects. For example,
the English ›pig‹ concept will reflect that pigs are domesticated and eaten (witness
the meaning relation between the words pig and pork). The concepts are related to
the culture, i.e. to the way speakers of a language ‘look at’ things. For other semantic
domains, the findings on colour terms and colour perception have no significance at
all. Many words denote things that cannot be immediately perceived with our senses;
it would not be possible to ask subjects to point to the best examples from a selection
of cases. Consider a word like mistake. The notion ›mistake‹ appears quite natural
and basic to us. But it presupposes complex cultural systems of rules that determine
what is right and what is wrong and in which regard. Such concepts are completely
culture-dependent.
Relativism is certainly right in emphasizing the differences between languages.
They do exist, they are the rule rather than the exception, they are deep – and they
are fascinating. A relativist attitude is absolutely necessary for all who seriously try
to understand other languages. It provides the only chance to escape the natural
tendency of applying one’s old views and categories to new things. Only when we
expect other languages to be different from our native language, will we be able to
recognize and understand the differences. Thus, relativism is the better working
hypothesis. Once we are aware of the differences between languages, we may set
out to satisfy the universalist by trying to determine the common denominator that
makes a comparison possible. After all, it must exist, because otherwise we would
not have been able to understand our native as well as another language with our
one and only mind.
EXERCISES
1. When words are imported from other languages, their meaning is often changed
more or less. For example, the German word Arbeit (›work‹) was borrowed
into Japanese, where it was shortened to baito. Japanese baito has a narrower
meaning, denoting part-time students’ jobs such as giving private lessons.
a. Try to find three words in your native language that are borrowed from
other languages. Consult dictionaries to determine their meanings in both
languages.
b. Try to find an explanation why the meanings of loanwords so often differ
from their origins – despite the fact that they are apparently borrowed for
their meaning.
2. Try to determine the meaning relations that hold between the following pairs of
words. First look up the translation of the English term in a bilingual dictionary,
then crosscheck with a reverse dictionary:
a. English cook (verb) vs
German kochen
b. English hot
vs
German heiß
264
Understanding semantics
4. Ask three people to spontaneously list twenty or more colour terms in the order
in which they come to their minds. Compare the results. Are the eleven BCTs
among the words listed? Do they appear towards the beginning of the list? Which
are the first five items?
5. Given that a BLACK/3 BCT covers not only black but also blue, green and other
colours, why is it considered a case of BLACK/3 rather than, say, BLUE/3?
6. What are the essentials of the relativist and the universalist position? Which
position do you think is realistic? What is your own position in view of your
personal experience with foreign languages?
FURTHER READING
Suzuki (1978, ch. 5) for Japanese terms of address and self-reference; Kuroda (1973)
for Japanese evidentials; Palmer (2001, ch. 2.2) on evidentials in general. Whorf
(1940) for a basic outline of the relativist position. Salzmann (1993, ch. 8) for a
discussion of the Sapir-Whorf Hypothesis and the relation between language and
culture. Palmer (1996, ch. 6) on language and worldview. Berlin and Kay (1969) for
the original report of their investigations of colour term systems; Kay and McDaniel
(1978) for an explanation of the universals in terms of colour perception; Wierzbicka
(1996) for an alternative account in the framework of NSM. Foley (1997, ch. 7) for a
more recent discussion of the research in colour terminology, ch. 6 on kinship term
systems in the context of the relativism vs universalism debate. Lee (1996) offers
a careful reconstruction of Whorf ’s original theory and its fate in the relativistuniversalist debate. For general discussions of the relationship between language and
thought see Lucy (1992) and Deutscher (2010); both contain a thorough discussion
and evaluation of the research about colour terms; on the latter, also see Dedrick
(1998). For methods in cross-linguistic semantics see Matthewson (2011).
11
Meaning and cognition
Over the last couple of decades, the development of a new branch of science, cognitive
psychology, or more generally cognitive science, has given important fresh impulses
to linguistics in general and semantics in particular. Cognitive science is concerned
with how the human mind works, how it receives information from the environment
via the senses and processes this information, recognizing what is perceived,
comparing it to former data, classifying it and storing it in memory. It tries to account
for the complex ways in which the vast amount of information is structured in our
minds, how we can operate with it when we think and reason. Language plays a
central role in these theories. On the one hand, speech perception and production
and the underlying mental structures are major objects of investigation. On the other
hand, the way in which we use language to express what we ‘have in mind’ can tell
much about how the human mind is organized.
The importance of cognitive psychology for semantics lies in its emphasis on the
exploration of the concepts and categories we use. While semantics in the wider
tradition of structuralism aims at a description of meaning relations, the cognitive
approach focuses on meanings themselves. It tries to provide positive descriptions
of the word meanings, and thereby to account for why and how words denote what
they denote. In terms of the semiotic triangle, a cognitive approach to semantics
can be characterized as focusing on the base of the triangle, the meaning and how it
determines the denotation (Fig. 11.1).
Figure 11.1
The focus of cognitive semantics
expression
meaning
denotation
266
Understanding semantics
11.1 CATEGORIES AND CONCEPTS
The fundamental notion of cognitive science is categorization. In his introduction
to cognitive psychology, Barsalou (1992a: 15) describes it as follows:
Upon walking into a home, people know instantly what is present. They recognize
chairs, stereos, plants, dogs, friends, guacamole, wine, and just about anything
else they happen to perceive. […] When people interact, they recognize friends,
facial expressions, actions, and activities. When people read, they categorize
letters and words. Categorization occurs in all sensory modalities, not just vision.
People categorize the sounds of animals and artifacts, as well as the sounds of
speech; they categorize smells, tastes, skin sensations, and physical movements;
and they categorize subjective experiences, including emotions and thoughts.
Categorization provides the gateway between perception and cognition. After a
perceptual system acquires information about an entity in the environment, the
cognitive system places the entity into a category.
Categorizing something that we perceive (or imagine, or remember) means to
perceive it as something of a kind. If we look at a photograph of a person, we will
categorize it as an object of the kind ‘photograph’ that displays something of the
kind ‘person’, and if we recognize the person, we will categorize him or her as that
person. We assign everything that enters our minds to one or more categories. The
category DOG (we will use SMALL CAPITALS for categories) consists of all those things we
would categorize as dogs. It includes not only the real existing dogs but also former
and future dogs or fictitious dogs in novels, jokes, etc. Entities belong to different
categories at the same time. Our common friend John belongs to the categories
PERSON, MAN, FICTITIOUS, BICYCLE OWNER and maybe (we just do not know) LOUSY DANCER
along with countless other categories. The single entities that belong to a category are
called exemplars or members of the category. (John is an exemplar of the category
BICYCLE OWNER.) Larger, more general categories include smaller, more specific ones.
These are subcategories of the former: MAN and WOMAN as well as BICYCLE OWNER are
subcategories of the category PERSON. All members of a subcategory are also members
of the more general category. Note that the subcategories of a category are not
members of it. Rather both share members. If you are familiar with set theory, it may
help to realize that categories are sets, subcategories subsets and members elements.
We encountered categories before: denotations (2.2.2) are categories. The
descriptive meaning of a word defines a category, the set of all its potential referents.
When we refer to Donald using the term duck, we treat him as a member of the
category DUCK.
Categorization is only possible if the respective categories are in some way
available in the cognitive system, i.e. in our mind. In the terminology of cognitive
science, categorization requires mental representations of the categories. There are
various theories concerning the nature of category representations. Here the view
will be adopted that categories are represented by concepts for their exemplars. The
Meaning and cognition
267
1
category DOG is represented in the mind by the concept ›dog‹. When we encounter
an object, our cognitive apparatus will produce a preliminary description of it, which
consists of what we perceive of the object, e.g. its size, shape, colour, smell, etc. The
description will be compared with the concepts we have in our mind and if the
description happens to match with the concept ›dog‹, the object will be categorized
as a dog.
Two things must be kept in mind when we think of word meanings and concepts.
First, word meanings do not coincide with our concepts for actual categories. For
example, the meaning of the word bicycle is a relatively abstract concept that is just
rich enough for defining the category of bicycles. But you will probably have a richer
concept depending on your individual knowledge of bicycles and your personal
experiences. Thus, the bicycle concept that constitutes the meaning of the word is
only part of the concept that in your mind defines your personal category of bicycles.
The meaning of the word must be a much leaner concept that is shared by all English
speakers who know the word and its meaning.
Second, we do not have a word for every category we have a concept for. Trivially
there are infinitely many concepts which can be expressed only by complex
expressions, e.g. the concept ›expensive sushi restaurant closed on Saturdays‹ – the
syntactic possibilities of language reflect this potential. But there are also categories
and concepts which cannot be verbalized at all, or only insufficiently. Many concepts
for bodily sensations, feelings, emotions, for facial expressions and physiognomies,
for flavours and odours, for melodies and harmonies, etc. can hardly be put into
words, if they can at all. For example, a verbal description of a face can never be as
accurate as a photograph. Words can never fully describe the taste of an orange, the
smell of carnations or the sound of a violin. In fact, it is plausible to assume that only
the lesser part of our concepts can be expressed in language.
The system of lexical meanings is only part of the overall system of concepts. And
a lexicalized concept is only part of the richer concept that defines the category we
actually connect to the word. In this regard, a distinction will later be drawn between
semantic concepts (word meanings), cultural concepts (richer concepts shared by a
cultural community) and personal concepts (cf. 11.6).
11.2 PROTOTYPE THEORY
11.2.1 The traditional model of categorization
The traditional view of categorization is shaped by the model of ‘necessary and
sufficient conditions’ (NSC model for short) that goes back to Aristotle. According to
the NSC model, a category is defined by a set of necessary conditions, which together
are sufficient. For example, if we assume that the category WOMAN is defined by the
1 Henceforth ›…‹ quotes are used not only for meanings but for everything on the conceptual level,
i.e. meanings, concepts and components of concepts.
268
Understanding semantics
three conditions of being human, female and adult, each one is necessary. If someone
is not human or not female or not adult, he or she is not a woman. On the other
hand, the conditions of being human and female and adult are jointly sufficient for
membership in the category WOMAN. It does not matter what other conditions someone
or something may fulfil. Being a woman or not depends on these three conditions.
As you will have noticed, the NSC model of categorization, also called the checklist model, is implemented by the BFA notion of meaning (9.3). The binary features
that make up the meaning of a word according to BFA are jointly sufficient necessary
conditions of the NSC model.
The Aristotelian model can be characterized by the following points:
∑ Categorization depends on a fixed set of conditions or features.
∑ Each condition is absolutely necessary.
∑ The conditions are binary (yes-or-no) conditions.
∑ Category membership is a binary (yes-or-no) issue.
∑ Categories have clear boundaries.
∑ All members of a category are of equal status.
That categories have clear boundaries is a direct consequence of the fact that the
defining conditions are binary. Everything either fulfils this set of conditions or it
does not. If it does, it belongs to the category, otherwise it does not. As a consequence,
categories have clear boundaries, and within their boundaries all members enjoy the
same status of full members. Each single point of the NSC model was challenged in
prototype theory, the result of the first extensive studies in categorization undertaken
by cognitive psychologists and semanticists.
11.2.2 Prototypes
The findings on colour terms (10.4), which at the same time were findings on colour
categorization, did not seem to fit the NSC model at all. Clearly, the subjects did not
categorize the colour chips by checking a list of binary features. The basic colour
categories are primarily defined by focal colours. A definite boundary between
neighbouring categories cannot be drawn.‘Category boundaries […] are not reliable,
even for repeated trials with the same informant’ (Berlin and Kay 1969: 15). When
one moves from focal red to focal brown in a system that has BCTs for both, one
somewhere leaves the range of red hues and enters that of brown hues. Furthermore,
the categories RED and BROWN intuitively overlap. Brownish red is predominantly
red but still more or less brown, and reddish brown is vice versa. Thus, it appears, a
colour belongs to a category if it is sufficiently similar to the focal hue, i.e. the best
example. Since, however, similarity is a matter of degree, category membership too is
a matter of degree, rather than a yes-or-no issue.
Linguists and psychologists set out to investigate other areas for similar
phenomena. They found that for many different categories best examples can be
empirically established. These come to mind first and are consistently considered
Meaning and cognition
269
‘better members’ than less typical cases. Such central examples came to be called
prototypes. Experiments were carried out to examine whether category membership
in general is a matter of similarity to the prototype and hence a matter of degree;
whether categories have fuzzy boundaries; whether the defining conditions, or
features, are binary and always strictly necessary.
It soon turned out that, apparently, many categories have a ‘graded structure’.
They contain prototypical exemplars that represent the category best and other
exemplars that do it to a lesser extent but are still good examples, while others only
enjoy a marginal status. For example, in a famous study cited in every textbook on
cognitive semantics, Eleanor Rosch, whose name is inextricably linked with the early
development of prototype theory, established a ranking within the general category
BIRD. The subjects participating in the study were asked to rate on a scale from 1
(for best) to 7 the ‘goodness-as-example’ of different kinds of birds. The results
were surprisingly consistent. Robins were considered the best examples, followed
by doves, sparrows and canaries; owls, parrots, pheasants and toucans occupied a
medium position, ducks and peacocks were considered even less good examples,
while penguins and ostriches ranked lowest. Similar results were obtained for the
categories FURNITURE, FRUIT, CLOTHING, etc. Furthermore, the rankings were consistent
with other findings, such as the time it took the subjects to answer questions of the
type ‘Is a penguin a bird?’, ‘Is an eagle a bird?’, etc.; the less typical the examples were,
the longer the reaction time.
Since prototypical examples are what we think of first, we will exclude other cases
when a category is mentioned, as long as there is no reason to do otherwise. Thus
prototypes play an important role in what is called default reasoning, i.e. reasoning
in terms of assumptions which replace specific actual information as long as none is
provided. For example, when Mary tells John,
(1)
Look, there’s a bird on the window sill.
John will think of a prototypical bird, not of an owl, a condor, an ostrich or a penguin.
If someone mentions a ‘car’, we will not think of a truck or a veteran car. It is therefore
misleading to use the general terms for non-prototypical cases, misleading – but
not semantically incorrect. Penguins are birds and we can refer to them as birds in
appropriate contexts. For example, (2a) and (2b) are perfectly acceptable:
(2)
a.
The only birds that live in the Antarctic are penguins.
b.
Penguins come ashore to nest. The birds lay one to three eggs.
11.2.3 Fuzzy boundaries
Other experiments were performed in order to assess the fuzziness of category
boundaries. The linguist William Labov presented pictures similar to those in Fig.
11.2 to students and asked them to name the objects depicted.
270
Understanding semantics
Figure 11.2
Cups, bowls, vases
1
2
8
3
4
5
9
10
7
6
The subjects categorized objects like 3 as a cup, 10 as a vase and 6 as a bowl, but
produced inconsistent answers for vases or bowls with a handle – 1 and 5 – or
exemplars with intermediate height-width ratio such as 2, 4, 7 and 9. The subjects
were also asked to imagine that the objects were filled with coffee or standing on a
shelf with cut flowers in them, etc. This had a strong influence on the categorization
(for example, more objects were categorized as cups when they contained coffee). The
experiment showed that prototypical cups are about as high as they are wide, have a
handle and are used for drinking hot drinks such as coffee, while prototypical vases
are higher than they are wide, lack a handle and are used for putting flowers in. If we
encounter an object where these features are shuffled into a different combination,
the criteria for categorizing it as a cup, a vase or a bowl may come into conflict
with each other. Is object 1 a cup because it has a handle or is it a vase because it is
much higher than wide? Likewise the criterion of the height-width ratio may cause
problems if an object’s ratio lies somewhere between that of cups and vases (2 and
9) or between that of cups and bowls (4 and 7).
11.2.4 Family resemblance
The prototype phenomena cast a new light on the features that define a category. If
a category is primarily defined through its prototype, the conditions that define the
prototype need not be necessary conditions for the rest of the category. (If they were,
all members of the category would be equally prototypical.) For example, members
of the category CUP may lack a handle although the prototype has one; the prototype
of the category BIRD is small and able to fly, but other members may be unable to fly
or much bigger. In the case of cups, let us assume that they are distinguished by (i)
being used for drinking coffee or other hot drinks, (ii) having a handle, and (iii) a
balanced height-width ratio. In Fig. 11.3 the ten objects are arranged in accordance
with these features.
If we accept that all objects except 6 and 10 (the prototypical bowl and vase) are
in some sense cups, the resulting category is not defined by any common condition.
Some of the objects fulfil one condition, others two, prototypical cups all three. What
ties the objects together in one category is what is called a family resemblance.
Meaning and cognition
271
Figure 11.3
Distribution of cup features and family resemblance
balanced ratio
used for drinking coffee
1
2
3
4
5
10
9
8
7
6
handle
The philosopher Wittgenstein (1958) introduced the notion in connection with his
famous example of the category GAME. He argued that there is no property shared by
all games. Rather, some games share properties with certain others, e.g. the property
of being played on a board or being competitive. Others may share no properties at
all, but the whole domain of games is interconnected by similarities between the
members. Applied to our simple example, we see that object 1 resembles object 2 in
having a handle, object 2 resembles object 9 in that we would drink coffee from it. But
object 1 and object 9 share none of the three properties, yet they belong to the same
category, due to the linking element 2.
11.2.5 Graded membership
Still, the ten objects do not seem to be simply cups or not. We may count object 1
as a cup, but it certainly resembles a vase more than a cup. Other objects such as 7
seem to fall right between the categories CUP and BOWL. The many cases of uncertain
category membership and the ranking results in the tests mentioned suggested that
category membership is not a yes-or-no question but a matter of degree. Degrees of
membership can be determined by ranking experiments. Or they could be calculated
on the basis of the prototypical properties, weighting the features and reducing the
degree of membership if a certain property is missing. For example, one might rate
the prototypical cup 3 as a 1.0 cup, and 2, 4 and 8 as 0.8 cups (= pretty good cups).
Cups 7 and 9 might be 0.5 cups (half-cups), and 1 and 5 0.2 cups (very marginal
cups). But the prototypical bowl 6 and the prototypical vase 10 would be 0.0
members of the category, i.e. definitely non-cups.
11.2.6 The prototype model of categorization
The characteristics of the resulting new model of categorization can be summarized
as follows:
∑ Graded structure
The members of a category are not of equal status.
272
Understanding semantics
∑ Prototypes are best examples
There are prototypical members that are consistently considered the best
examples of the category.
∑ No set of necessary conditions
Category membership is not a matter of a fixed set of necessary conditions. The
prototype of a category may be defined by properties absent with less typical
examples.
∑ Family resemblance
Category members are connected by family resemblance.
∑ Prototypes are reference points
Prototypes serve as reference points for categorization.
Category membership is a matter of similarity to the prototype.
∑ Graded membership
Category membership is a matter of degree.
∑ Fuzzy boundaries
Categories have fuzzy boundaries.
As we shall see in the next section, most of these claims are not unproblematic. But
when they were first established, they appeared suggestive and plausible and proved
very influential. In fact they still are, despite the fact that researchers like Rosch
herself soon revised her initial views. Before we turn to a critique of PT (henceforth
short for prototype theory), we will take a closer look at the notion of the prototype.
In order to make the argumentation in the following more transparent, Fig. 11.4
displays the central points and how they are interconnected in PT.
Figure 11.4
Central claims of Prototype Theory and how they are connected
There are
best examples:
the prototypes
therefore
therefore
Graded structure:
members differ
in status
Prototypes are
the reference points
for categorization
therefore
therefore
Graded membership
Fuzzy boundaries
According to PT, prototypes have two crucial properties. First, they are the best
examples for their category. This point is due to the graded structure of categories,
i.e. the existence of better and poorer examples. Conversely, the existence of best
Meaning and cognition
273
examples, along with less good examples, entails graded structure. Second, PT
claims, prototypes serve as reference points for categorization, the crucial criterion
for membership in the category being similarity to the prototype. The two properties
are seen as mutually dependent: prototypes are the best examples because they serve
as reference points, and they do so because they are the best examples. Similarity
itself is a matter of degree. Hence the second property of prototypes implies graded
membership: category membership is not a yes-or-no matter but a matter of
degree. As a consequence of graded membership, categories have fuzzy boundaries.
Conversely, fuzzy boundaries are understood as giving rise to graded structure: the
clear members of a category are better members than those that form the fuzzy
boundary area of the category. Graded structure and graded membership are seen
essentially as two sides of the same coin. If something is a poorer member, it is less
similar to the prototype and hence belongs to the category to a lesser degree, and
vice versa. For the sake of simplicity, family resemblance and the role of necessary
conditions are not included in the schema.
In the following, the network of assumptions and claims of PT will be subjected
to a critical evaluation from a semantic point of view. As it will turn out, the picture
gives rise to a couple of serious questions.
11.2.7 What kinds of entity are prototypes?
The informal definition of prototypes as the ‘best examples’ suggests that prototypes
are distinguished individual members of the category. This makes sense in some
cases, for example for the colour categories. The prototype of GREEN is one particular
colour, i.e. one member of the category. But if prototypes are to be not only best
examples but also reference points for categorization, then for most categories
they cannot be exemplars. If, for example, a certain dog exemplar had to serve as
the prototype of the category DOG, one would have to know this particular dog in
order to be able to categorize anything as a dog, because potential members would
have to be compared with it. Exemplar prototypes may work in very early stages of
language and concept acquisition where we first learn what a dog is by being shown
a particular dog. But later we will replace the representation of the first dog we knew
by a more abstract general representation of ‘the’ prototypical dog.
In the ranking experiments by Rosch and others, the subjects were asked to
judge the representativity not of exemplars but of subcategories of a given category.
For example, they were asked to rank subcategories of birds. But identifying the
prototype with a particular subcategory does not work either. First, the problem
is only shifted. If ROBIN is the prototype of BIRD, what is the prototype of ROBIN?
Most of us do not know subcategories of robins, but certainly there are better and
poorer examples of robins too. Also, the question arises as to whether all robins
are prototypical examples of the category BIRD or only prototypical robins. If the
latter, the prototype of ROBIN would be identical with the prototype of BIRD and the
categories would coincide. If we decide instead that any robin whatsoever counts as
a prototypical case of a bird, we have to include extraordinary big fat monster robins
unable to fly among the prototypical birds. Second, robins exhibit distinctive features
274
Understanding semantics
such as having a red breast which we would not include among the features of the
prototypical bird.
It makes much more sense to assume that the prototype is an abstract case defined
by a concept that fixes certain features and leaves others open. The concept for the
prototypical bird contains a specification of its appearance and anatomy and of the
most prominent features of its behaviour. But other traits may be unspecified, such
as their diet, the exact shape of the beak, the colour of the feet or the markings of
the feathers. Such a concept is different from the concept for an individual bird,
which would have to be specified for all features. Therefore, the prototype is not an
exemplar. It is not a subcategory either, because these would also be specified for
some of the features left open, e.g. the four mentioned. An individual bird will count
as prototypical if it exhibits all the features that make up the prototype. Prototypical
birds may differ in the features left open. If a certain subcategory, e.g. ROBIN or
SPARROW, matches the concept of the prototype for its normal members, it may be
roughly identified with the prototype but, strictly speaking, it is something different.
It must be noted, however, that for more abstract categories, such as the biological
category ANIMAL that includes not only mammals, prototypes cannot be defined in
2
this way. There are certainly best examples for the category, probably cats and dogs.
But dinosaurs, birds, fish, insects, worms, coral polyps and amoebae are also animals.
The category is much too general to fix any concrete anatomical or behavioural
properties of the kind that make up the prototype. Rather, for such categories the
choice of best examples is not a matter of exemplariness but of familiarity. It is
therefore questionable if this kind of prototype can possibly serve as a reference
point for categorization with respect to such categories.
11.2.8 Which features make up the prototype?
One is tempted to answer: the typical features. Some features of birds are intuitively
considered essential and typical features of birds, e.g. having wings and feathers
and being able to fly as well as the properties of singing, having a beak, laying eggs
and hatching them in a nest. Others, like having a particular colour, a certain weight
and size may be characteristic for certain kinds of birds but not for birds in general.
Having feathers is a ‘good’ feature because it well distinguishes birds from non-birds.
The feature is therefore said to have a high cue validity for the category BIRD which
means that it applies to a high proportion of members and to a low proportion of
non-members. The features of having wings, being able to fly and laying eggs have
lower cue validity because they are shared by other kinds of animal, e.g. most insects.
They are, however, of higher cue validity within the narrower domain of vertebrates.
Here, the feature of having wings and being able to fly is very distinctive, since except
for birds only bats share it. Thus, a prototype must be defined by a combination of
features that together maximize the cue validity. Features of low cue validity that
2 Along with this category, there is a much narrower category ANIMAL, that equates roughly to
QUADRUPED and does not include birds, fish, snakes, etc., let alone insects and amoebae. Here and
below, the category ANIMAL is not meant in the narrow sense.
Meaning and cognition
275
distinguish birds and other vertebrates from insects have to be included in order to
enhance the cue validity of the features such as having wings and being able to fly.
11.2.9 Similarity
Prototype theory claims that membership in a given category is a matter of similarity
to the prototype. While this seems a clear and simple criterion, a closer look
shows that it is anything but that. When thinking of applications, we will think of
comparing, for example, various kinds of birds to a prototypical bird with respect to
physical appearance, primarily shape, size and colour. But, in fact, similarity in this
concrete sense of resemblance is not relevant in most cases. Consider, for example,
a wolf, a husky and a poodle. Wolves and huskies are physically very similar and
both very different from poodles. Yet the husky and the poodle belong to the same
category DOG and the wolf does not. The categorization of dogs and wolves is not only
a matter of physical similarity. Membership in the category DOG must be a matter
of similarity to the prototypical dog in some other respects. The question, then,
arises: in which regard is a would-be member of a given category to be similar to
the prototype? Which of its properties are relevant for the comparison? The answer
is far from trivial.
Consider a second case. Mary has two cousins, Marc and Tom (now you know).
They are brothers and physically very similar. Let us assume we know that Marc is
a fairly prototypical member of the category MAN. The physical resemblance of Tom
would then entitle us to categorize him too as a man. But if Marc is a butcher, Tom’s
physical similarity to Marc does not entitle us to conclude that he has the same
profession. Rather, the relevant aspect of Marc and Tom would be the kind of job they
have or they are qualified for. Categorization in the category MAN requires similarity
to the prototype in quite different respects from categorization in the category
BUTCHER, or in the category BACHELOR or NEIGHBOUR or HIP-HOP FREAK, for that matter.
These considerations show again that in order for prototypes to serve as reference
points of categorization they must be defined by a set of crucial features. And it is
these features that must be checked in order to judge the similarity to the prototype.
The features may be of different weight or interdependent. For example, for birds the
feature of being able to fly may depend on the body weight and the form and relative
size of the wings. In this sense, the resulting model of categorization is not the same
as the simple checklist model of NSC, where each feature is supposed to be necessary,
independent of the other features. But the difference is not as radical as PT makes it
appear.
Another difficult question concerns the relevant scale of similarity on which the
degree of membership depends. Let us assume similarity is measured in values
between 0 and 1. Given a prototype and the set of criteria relevant to the comparison,
it is clear which cases receive the value 1 on the scale of similarity: all those that
completely agree with the prototype in all relevant aspects. But in the other direction
the scale is undefined and, in fact, indefinable. Consider again the case of the category
DOG. Since wolves are definitely non-members of the category DOG, they must be
assigned the value 0.0 on the scale of similarity. But some of the crucial defining
276
Understanding semantics
properties of the prototypical dog must also be properties of wolves. Certainly
wolves are more similar to dogs than cows. For that reason, the categories WOLF and
DOG belong to the same superordinate category CANINE while COW does not. It then
appears we should assign cows the value zero and wolves some value reasonably
greater than 0.0. But, of course, this kind of adjustment of the scale can be repeated
arbitrarily. Cows are more similar to dogs than crabs are, because both cows and
dogs are mammals; crabs are more similar to dogs than potatoes, potatoes more
similar than stones, and stones more similar than, say, French personal pronouns.
The more kinds of cases we include in the assignment of similarity values, the
greater the value of wolves becomes and the more similar (in terms of the scale) they
become to dogs. It is just impossible to fix, for a given category, a general zero point
of the scale and degrees of membership for potential and actual members – except
for uncontroversial members. Apparently the degree of similarity and membership
depends on the given context, namely on the range of rival categories.
The considerations about the nature of the prototype, its defining properties and
the notion of similarity show that PT is not as unproblematic as it appears at first
sight. The claim that categorization is a matter of similarity to the prototype raises a
lot of non-trivial questions.
11.3 THE HIERARCHICAL ORGANIZATION OF CATEGORIES
11.3.1 The basic level
Look at Fig. 11.5.
Figure 11.5
What do you see? Your spontaneous answer will probably be, ‘A lion!’ Although
3
your answer is not hard to guess, the fact that it is not is far from trivial. The reason
why the unanimous answer is not trivial is that ‘A lion’ is by no means the only
possible answer to the question. You could have categorized the entity referred to
more specifically as a ‘male lion’, an ‘adult male lion’, a ‘prototypical male lion’ or a
‘marching male lion with a big mane’, ‘a marching big lion in side view’ and so on.
Alternatively it would be correct to use more general terms such as feline, mammal,
animal, living thing or entity. In the given context, the question can also be taken as
referring to the picture rather than what it depicts. So what you see when you look
3 You can test it yourself: Show twenty people a toothbrush, ask them to answer spontaneously to the
question ‘What is this?’, and twenty people will answer ‘A toothbrush.’
Meaning and cognition
277
at Fig. 11.5 is also a ‘picture’, a ‘picture of a lion’, the ‘MS Office™ ClipArt’s lion in
black and white’, an ‘illustration’, an ‘example’, etc. Even if we disregard the latter
kind of answer, the phenomenon remains that an object can always be categorized
at different levels of generality. What the little experiment shows is that there is
obviously one level of categorization that we prefer, an intermediate level somewhere
between very specific and very general categorization.
Berlin and others studied large plant taxonomies in a Mexican language and
managed to establish this preferred intermediate level. Rosch applied this kind
of research to other areas such as clothing and furniture. The main result is this:
there is an intermediate level, the so-called basic level which is in several regards
distinguished and privileged. Fig. 11.6 shows a minor part of the taxonomy of
English terms for musical instruments.
Figure 11.6
Categories of musical instruments – basic level
MUSICAL INSTRUMENT
basic level
TRUMPET
SAXOPHONE
JAZZ
BACH
TRUMPET TRUMPET
SOPRANO
SAX.
ALTO
SAX.
PIANO
…
GRAND
PIANO
UPRIGHT
PIANO
The basic level is occupied by the categories TRUMPET, SAXOPHONE, PIANO (as well
as all those not depicted here, such as VIOLIN, FLUTE, GUITAR, DRUM, ORGAN, CLARINET,
etc.). The subordinate level is by no means complete. Of course, it is not the lowest
level that exists. On the second subordinate level there are special categories of jazz
trumpets or soprano saxophones, and so on. But it should be noted that there is a
limit to levels of categorizations as far as lexicalized categories are concerned. The
subordinate level displayed here is probably the lowest level of categories in nonexpert terminology. Table 11.1 gives some further examples of the basic level.
Table 11.1
Examples for the basic level
Level
superordinate
basic level
Categories
GARMENT
VEHICLE
TROUSERS,
SKIRT, SHIRT,
DRESS, BRA,
JACKET,
subordinate
ANIMAL
COLOUR
BICYCLE, CAR,
DOG, CAT,
WHITE, BLACK,
BUS, TRUCK,
HORSE BEAR,
RED, YELLOW,
PIG, RABBIT,
GREEN, BLUE,
MOTORBIKE
…
…
BLUE JEANS
…
TIGER, LION,
RACING BIKE
…
COLLIE
…
…
BROWN,
…
TURQUOISE…
278
Understanding semantics
11.3.2 Properties of the basic level
11.3.2.1 The psychological aspect: properties of basic level categories
The basic level (BL) is privileged in the cognitive system in various ways. It operates
faster than higher and lower levels: in psychological experiments, response times
for BL categorizations are the shortest. Apparently it is the level at which most of
our general knowledge is organized. For example, the knowledge of what people
normally wear is in terms of BL garment categories, the composition of a jazz band
is in terms of BL instruments, and so on.
The BL is the highest level at which category members have a similar overall shape.
(Trumpets have a common shape and saxophones have, but not musical instruments
in general.) The overall shape is not only a matter of the visual appearance. An
important aspect of shape is the characteristic parts the object consists of. Most BL
terms have their own mereology (8.4.4). For example, a ‘piano’ has a ‘keyboard’; the
keyboard in turn consists of white and black ‘keys’ which each serve to move a felt
‘hammer’ inside the piano that hits one to three ‘strings’ thereby producing a tone of a
certain pitch. Similarly, the category of birds is characterized by having ‘feathers’ and
a ‘beak’ among other parts of their anatomy. No such common parts necessarily exist
for higher level categories such as ANIMAL or MUSICAL INSTRUMENT.
In the case of visible categories, it is possible to draw a general picture of the
members of BL and lower categories, e.g. a picture of a pig, a piano or a car, but not
so for any higher level categories.
BL categories are partly defined in terms of the ways in which we interact with
their members. For example, a trumpet, a piano and a harp are each played in their
own way and used for different kinds of musical activity. Shirts are put on and worn
in particular ways that distinguish them from coats or socks. For artefacts, the
characteristic parts (e.g. keys, wheels, sleeves) are intrinsically connected to the ways
in which we use the objects or in which they function. (Therefore not any arbitrary
piece of an object counts as a part of it, but only such that are linked to the way in
which the object functions or is handled. A key of a piano is a part of it, but not the
middle part of its lid.) Likewise, members of natural categories play a different role
in human life. In Western culture, pigs are held for slaughter, and cats as pets. Tigers
mostly live in the wild and if we encounter one, we will act differently than if facing
a reindeer, an ant or a snail.
11.3.2.2 The linguistic aspect: properties of basic level terms
The preference for BL categories in our cognitive system is reflected by the terms we
have for BL categories and the privileged role they play in communication. They are
mostly simple, short, native and old (except for terms for new categories of artefacts).
If BL terms are complex, long or of foreign origin, they are likely to be shortened and
assimilated (note the origins of bra < brassiere, bus < omnibus, piano < pianoforte,
bike < bicycle). In contrast, lower level category terms are often composite (cf.
wonder-bra, jazz trumpet, racing bike, olive green) and sometimes borrowed from
other languages. BL terms are learned first. And they are the ones used most
Meaning and cognition
279
frequently. Their use complies with one of the fundamental rules of communication,
the demand to be as informative as required and not to be more informative than
4
necessary.
11.3.2.3 The basic level and prototypes
The concept of prototypes works best for BL categories. The common overall shape
and the uniform ways in which we interact with the members of BL categories
make up the greater part of the prototype. BL categories combine a high degree of
distinctness from others with a rich specification of characteristics, thus exhibiting
a high cue validity.
11.4 CHALLENGES TO PROTOTYPE THEORY
In this section and the next, some issues will be pointed out that present challenges to
PT. The discussion offers the opportunity of addressing a number of basic questions.
The critical points are those of the right part of the schema in Fig. 11.4:
∑ the role of prototypes as reference points for category membership and,
consequently, the abandonment of necessary conditions
∑ the notion of graded membership
∑ the fuzziness of category boundaries
It will be argued that these points are not as firmly established and as closely
interconnected as was suggested in early PT. The phenomena PT addresses allow for
alternative interpretations and explanations. It should be emphasized, however, that
the critique does not concern the results concerning the hierarchical organization of
categories.
11.4.1 Graded membership vs graded structure
Some experiments of cognitive psychologists produced evidence in conflict with
PT. One such finding is that there are best examples even for categories with clear
boundaries and binary membership. For example, it was proved by experiments
that for the category ODD NUMBER the smaller odd numbers 1, 3, 5, 7 and 9 were
considered the best examples. But there are no borderline cases, and in no sense is
18764098376542141 less of an odd number than 3. There are two defining conditions:
an odd number must be a natural number > 0, i.e. a member of the set {1, 2, 3, 4,
…} and it must not be divisible by 2. Both conditions are necessary. Hence there is
no room for more or less similarity to the prototype(s). The number 4 is somehow
similar to an odd number in that it is not divisible by 3 (instead of 2), but this does
not count. Likewise, 12.999999 is almost identical with an odd natural number, but it
4 The so-called ‘Maxim of Quantity’, one of four conversational ‘maxims’ from Paul Grice. See chapter
1, Further reading, for references.
280
Understanding semantics
is no more of an odd number than 2 or π. Thus, the concept of similarity boils down
to the trivial borderline case of absolute equivalence. And as for that, any member
of the category would obviously do as an object of comparison. The main point, of
course, is that categorization in this case does not involve comparison at all. If we
want to decide whether a number is odd or even, we usually think of its decimal form
and check if the last digit is 1, 3, 5, 7 or 9. Thus, these five cases do play a crucial role
in categorization but not the role PT provides for them, i.e. not the role of serving as
reference points of comparison. The best examples of the category ODD NUMBER may be
considered prototypes (they certainly come to mind first and represent the category
best) and consequently the category itself has a graded structure. But, if these cases
are prototypes, they do not play the role of reference points for categorization. And
although the category has a graded structure, it has clear boundaries.
Similarly, in the case of the much-cited category BIRD one must admit that
penguins, although certainly odd birds, nevertheless are birds, not quarter-birds or
something. Likewise, for superordinate categories such as the broad category ANIMAL,
membership cannot be a matter of comparison with the prototype. There is no
similarity between an amoeba and a dog, not in any intuitive sense, yet an amoeba is
100 per cent an animal. For such categories, membership is a matter of definition, i.e.
of necessary conditions, although there are prototypes in the sense of best examples.
In general, the prototype model of categorization seems to work at the basic level and
below, but not above it. As a consequence, we have to state the following qualifications
of PT:
∑ A category may have prototypes, but they need not be reference points for
categorization.
∑ Graded structure is not necessarily linked with graded membership.
∑ Category membership is not necessarily a matter of similarity to the prototype.
∑ Category membership may be a matter of necessary conditions as assumed in
the NSC model.
The case of categories such as ODD NUMBER and ANIMAL shows that the two criteria
for prototypes – being the best examples and serving as reference points for
categorization – are, in fact, independent. Consequently, graded structure and
graded membership are not two sides of the same coin. Even if some members of
a category are more representative than others, all members may be 100 per cent
members separated from non-members by a well-defined boundary. Figure 11.7
schematically illustrates the case of a category with graded membership on the left,
which is necessarily accompanied by graded structure and a fuzzy boundary. The
right shows a category with yes-or-no membership and a resulting clear boundary,
yet with graded structure (such as ODD NUMBER or BIRD).
11.4.2 Fuzzy boundaries
From a semantic point of view, the question of fuzzy category boundaries and graded
membership is more important than the existence, nature and role of prototypes.
Meaning and cognition
281
Figure 11.7
Graded membership vs. graded structure
NO
YES
NO
graded membership
membership
NO
←
YES
→
NO
graded structure
It will be argued in 11.5 that for semantic categories, i.e. word and sentence
denotations, the assumptions of graded membership and fuzzy boundaries are
generally unacceptable. Since all examples hitherto mentioned (BIRD, CUP, etc.) are
semantic categories, this is indeed a strong claim. We will therefore take a closer look
at this part of PT.
There seem to be two reasons for the wide acceptance of the fuzziness claim.
(A) The claim appears to be in accordance with our intuitive impression that
category boundaries very often are in some sense fuzzy, unclear, variable or fluid. The
impression has several sources and, in fact, concerns different phenomena, which do
not all bear on the issue in the same way. The sources include:
∑ variation of word meanings within a language community (meaning variation)
∑ partial knowledge of word meanings (ignorance)
∑ sloppy use (pragmatism)
∑ inherent flexibility of word meanings (vagueness)
(B) The second reason is, of course, the evidence produced in experiments such as
the cups test by Labov. The evidence, however, is not as conclusive as it appears at
first sight. First, the outcomes of the experiments are in part a consequence of the test
design. Second, the results allow for alternative interpretations.
11.4.2.1 Sources of fuzziness
Meaning variation. As you will certainly have experienced countless times, people’s
use of terms actually differs. For example, one might have an argument with others
about where blue ends and green begins. Disagreement is even the rule for all
subjective categories such as BEAUTIFUL, BORING or FUN, which depend on personal
attitudes and evaluations. The application of personal standards for what is boring,
fun, etc. does, however, not mean that the resulting individual categories are fuzzy.
Apart from such subjective notions, there are words with which people connect
similar but different concepts. Think of categories such as ADULT. For some people,
282
Understanding semantics
age is the main criterion for adulthood, for others maturity, for yet others economic
independence.
Meaning variation plays a role in our intuitive notion of fluid category boundaries,
because in addition to the knowledge of word meanings we have a ‘theory of common
usage’, i.e. a theory of how other people use the words and how their uses vary. Such
a theory is an important part of our communicative competence, because in order
to make ourselves understood we must be able to anticipate how other people will
interpret what we say.
Ignorance. We all use so many words whose meaning we do not really understand. But
obviously, the resulting kind of uncertainty about the exact denotations of words is not
at issue in a general theory of categorization. PT is meant as a model for the categories
we have, not for those we have not. Nevertheless, the fact that there are so many
words we do not really understand contributes strongly to our subjective notion that
categories are fuzzy, and hence to the plausibility of the general fuzziness claim of PT.
Pragmatism. Whether something belongs to a category is one question, whether
it can be successfully referred to with the corresponding word is another. The first
issue concerns the order of things in one’s mind, the other is a matter of successful
communication. Since we do not have categories and words for everything we
may happen to want to talk about, we often find ourselves in need of referring
to something outside our system of semantic categories. If we then choose an
expression that in the strict sense does not apply to the intended referent but yields
a description that fits it approximately, this may well pass as regular language use. In
actual communication, we can expand the range of application for a given word well
beyond its original limits, provided we do not come close to the denotations of other
words. We often choose our terms rather sloppily. Thus, the practical denotation of
a word, i.e. the range of objects it can actually be used for, depends on the CoU and
5
is wider than the denotation defined by its lexicalized meaning.
Vagueness. If all these sources of fuzziness are excluded, the question remains of how
the proper source of fuzziness, variability of a particular, fully known, uncontroversial
category is to be modelled. PT has provided one model: categorization in terms of
graded similarity to the prototype. In 11.5.3 we will see that there is an alternative
explanation – vagueness – which is reconcilable with simple yes-or-no membership.
11.4.2.2 Test design and interpretation of the experiments on category fuzziness
In part, the results of the crucial experiments were predetermined by the very test
designs. When subjects are asked in a scientific experiment to assign membership
ratings to different types of furniture, they will produce a scale of the kind they
feel is required. The results will, however, not so much reflect their individual
categorizations as their willingness to compromise (pragmatism) and their theories
5 For a much more elaborate and systematic account of the ways in which we make use of our
linguistic abilities in actual communication, see the discussion of ‘variability’, ‘negotiability’ and
‘adaptability’ in Verschueren (1999), ch. 2.
Meaning and cognition
283
of common usage. Such an experiment is not suitable for testing whether for the
subjects membership is a yes-or-no matter or a matter of degree.
Another criticism concerns the nature of the attributes chosen for the tests. It is
questionable if the shape criterion which played such a prominent role in Labov’s
cup experiments is of much importance for the categorization of cups, bowls and
vases. Human artefacts are created for certain purposes. Cups, for instance, are made
for drinking coffee, tea, hot chocolate and other hot drinks. This is the central and
crucial feature of cups: ›for drinking hot drinks‹. From this single feature all other
characteristics of cups can be derived. As these drinks are hot, the material must
be sufficiently heat-resistant and insulating. A handle is a good idea if one does
not want to burn one’s fingers. The normal quantity of such beverages consumed
at a time determines the size of cups. The wish to have the drink while it is still
hot restricts the diameter of the container, and so on. The derivative features follow
from practical experience. Their presence optimizes the suitability for drinking hot
drinks from the objects. That is why they are features of the prototypical cup. None
of the derivative features constitutes a necessary condition. And there is no need to
consider any of them an extra part of the concept that defines the category CUP. But
then the question of whether one or more of them are absent has nothing to do with
the categorization. Likewise, the category VASE is defined by the feature that vases
are for putting cut flowers in. Thus, the impression of a continuum of membership
degrees between different categories such as CUP and VASE is an artefact of an
experiment that emphasized the derivative feature of shape rather than the primary
feature of function. The test design also suggested that the categories tested should
be mutually exclusive. But, in fact, the same object can be used both for drinking
coffee and putting flowers in (a fact that was confirmed by Labov’s results). When it
is used for coffee, it is a cup; when it hosts a bunch of flowers, it is a vase. Thus what
the experiments really show is that categorization is flexible and context-dependent
because it is, for things like cups, not primarily a matter of physical features. But they
do not prove that category boundaries are fuzzy.
11.4.3 Summary
What do the considerations in this section amount to? One of the basic observations
of PT cannot be denied. Certainly for many categories there are prototypes in the
sense of best representatives. More generally, many categories have more central and
more marginal members. In this respect the findings on colour terms carry over
to a large range of other categories. In addition, category boundaries are variable
and flexible. But we have to distinguish between variability in different senses.
Meaning variation, ignorance and sloppy use may contribute to our subjective
notion of category fuzziness. But only inherent variability of single categories in our
individual cognitive systems is relevant for a model of categorization. And this type
of variability can be explained differently, as will be argued in the next section.
Thus, while the observations are valid and challenging, the conclusions are not
necessarily those drawn in early PT. The experimental findings do not compel us to
accept that:
284
Understanding semantics
∑ ?? necessary conditions play no role in defining categories
∑ ?? category membership is generally a question of similarity to the prototype
∑ ?? category membership is generally a matter of degree
∑ ?? category boundaries are generally fuzzy.
Meanwhile, many cognitive semanticists have abandoned the original PT model.
Instead of prototypes they talk of ‘prototype effects’ in connection with graded
category structure. Richer and subtler models of categorization account for the
graded structure of categories (cf. the notion of ICM, ‘idealized cognitive model ’ in
Lakoff 1987). The structure then gives rise to prototype effects, for example typicality
of certain members of the category. In this volume, we will not go into these more
detailed developments of cognitive semantics.
11.5 SEMANTICS AND PROTOTYPE THEORY
11.5.1 Cognitive semantics
When we transfer the notions of cognitive science to semantics, we obtain the version
of the semiotic triangle displayed in Fig. 11.8.
Figure 11.8
Cognitive version of the semiotic triangle
expression
a concept
referent/
situation
The proposition of the sentence is a concept for the category of situations
potentially referred to, and the meanings of the referential phrases are concepts for
their potential referents. In relating a sentence to a particular situation and particular
referents, we categorize them accordingly. For example, if Mary says to John,
(3)
The tea is lukewarm.
she categorizes the situation as a the-tea-is-lukewarm situation, the beverage she
refers to as tea, its temperature as lukewarm. In general, each predication contained
in a sentence represents a categorization of its arguments, and the complex
predication expressed by the whole sentence amounts to a categorization of the
situation (recall 5.8).
In a very general sense, cognitive semantics can be used as a cover term for any
semantic approach that adopts this perspective on meaning and reference (i.e. for
Meaning and cognition
285
mentalist semantics in general). The range of cognitive semantics is smaller than the
range of cognitive science because, as was noted above, the categories and concepts
that can be expressed by words are only a subset of the categories and concepts that
form our cognitive systems. It is therefore useful to have specific terms for semantic
concepts, i.e. meanings, and semantic categories, i.e. denotations.
So-called prototype semantics (PS) applies PT directly to semantic categories.
If the notion is adopted that category membership is a matter of degree, then the
suitability of certain words for certain things is also a matter of degree. In the PS
perspective, we cannot say that we can or cannot use the word cup for referring to
some object x, but rather that x can well or hardly be called a cup. Graded membership
carries over to the whole sentence. A sentence is therefore not simply true or false.
If we identify true with 1.0 and false with 0.0, truth values lie somewhere on a scale
between 0.0 and 1.0 (including the two endpoints). For example, if Fido passes for
a 0.7 member of the category MONSTER, the sentence Fido is a monster will receive
a truth value of 0.7. As will be argued, this consequence is unacceptable from a
semantic and pragmatic point of view.
11.5.2 Polarization
Back in chapter 7, the Principle of Polarity was introduced: ‘In a given CoU where
its presuppositions are fulfilled, with a given reading, a declarative sentence is either
6
true or false.’ The principle now appears in a new light. As it stands, it plainly rules
out graded membership. Is it wrong? A simplification? Or is there evidence for it?
Yes, there is evidence. First of all, if there is any semantic universal, then it is
negation. Every language has grammatical and lexical means for negating a given
sentence. Thus, the possible sentences of a language, apart from a few exceptions
which need not bother us here, come in pairs of positive sentences and their
respective negations. As we saw in 7.1.3, determining the respective pairs is not
always trivial; we may tend to mix up negation with contraries, e.g. when we first
think of the negation of always as being never rather than not always. But it is possible
to determine such pairs, if the methods of logical analysis are properly applied. The
availability of negations for all positive sentences has the effect that whatever we say
constitutes a choice between the two alternatives, Yes or No (recall the discussion in
7.1.3 on negation). We called this phenomenon polarization.
Polarization is not restricted to declarative sentences. It also applies to all other
sentence types such as interrogative or imperative sentences:
(4)
a.
Why is there beer in the fridge? vs Why is there no beer in the fridge?
b.
Please, put beer into the fridge! vs Please, don’t put beer into the fridge!
6 The Principle of Polarity, as formulated here, provides for the possibility that a sentence lacks
a truth value because its presuppositions are not fulfilled. As we saw in 7.7.1, lack of a truth
value does not amount to an in-between truth value, but just to a truth value gap. Therefore, the
possibility of truth value gaps does not affect the argument we are going to develop here.
286
Understanding semantics
In this respect, it is even more general than the Principle of Polarity, which only
applies to declarative sentences. Polarization does, of course, not only apply at
sentence level, but it concerns every single predication contained in a sentence.
Imagine, for example, Mary telling John:
(5)
a.
The mouse is eating a big grain of corn.
In doing so, she not only opts for the positive alternative of the sentence as a whole,
but also for the positive alternative of the predications ‘is a mouse’, ‘is eating’, ‘is a
grain of corn’ and ‘is big’. (5) would not be true unless each of the predications is
true. Thus, when using predicate terms in composing a sentence, one puts a plus or
minus sign before each one and as a result also before the total predication expressed
with the sentence.
The general availability of negation is one piece of evidence for polarization. The
other, equally universal, is the possibility of transforming any declarative sentence into
an interrogative sentence. For Donald is a duck there is the interrogative counterpart
Is Donald a duck? and so on. The resulting questions require a simple type of answer:
Yes or No. Since the interrogative sentence has the same proposition as its declarative
counterpart (2.2.3) the fact that the possible answers to such questions are just Yes
or No proves that there are only two truth values for propositions, TRUE and FALSE, and
not a whole scale between them.
Consider now what would happen to polarization if membership in semantic
categories were a matter of degree. For example, let us assume that the creature
referred to in (5) is only a 0.8 mouse, that it is only 0.7 eating and the grain of corn is
only 0.3 big. Weighting the single predications appropriately, we might obtain a value
of, say 0.5, for the whole situation. (In models with graded membership, 0.5 is usually
considered a pretty high value, which justifies the categorization.) But under these
circumstances, we could with the same right claim the contrary of (5), namely that
the mouse is not eating a big grain of corn. Clearly this is not what we mean when we
opt for the positive sentence. We mean that it is true and not its negation. Even if the
situation we refer to is not exactly as we say, the way we put it is as we say it and not
the other way round. And this is what we will be taken to have said and is therefore
to be considered its meaning proper.
Polarization is inescapable. But it would not be, and probably would not even exist,
if membership in semantic categories were graded. It is therefore concluded that
semantic categories are binary.
Two questions arise now: (i) How can we account for flexible category boundaries,
if membership is yes-or-no? (ii) How does language cope with the obvious fact that
the world does not come divided into black and white but rather as a continuum of
phenomena without natural partitions? The answer to the first question is inherent
vagueness; the answer to the second is the availability of various linguistic means of
differentiation.
Meaning and cognition
287
11.5.3 Flexible concepts: vagueness
The crucial point concerning the flexibility of semantic categories is the fact that it
is built into the very word meanings. Let us consider a simple example, the adjective
big. It presents a prototypical case of what was already introduced as vague meaning
in 3.2.4. The vagueness of the adjective has two sources. First, whether we categorize
something as big or not big is a matter of a norm. If we consider the ‘big grain of
corn’ in (5) once more, the object may be called big if it is big for a grain of corn or
big for a mouse to eat or big for a grain of corn for this mouse to eat (the mouse may
prefer grains of corn of a certain size). These are three different norms. Once we have
chosen a particular norm, we next have to fix a borderline that separates big grains
from grains that are not big, i.e. we have to fix the criteria for a categorization by size
in terms of the chosen norm. Size is an attribute of objects that allows ordering on
a scale, with smaller objects ranging lower than bigger objects. Somewhere on this
scale there is a point beyond which everything is big, while everything below it is not.
This is illustrated in Fig. 11.9.
Figure 11.9
The big categorization
d
f
a
b
c
e
not big …
c
a
bigger
←NO
YES→
f
b
e
bigger
d
… big
bigger
Six objects of different size are shown in the upper part. In the lower part, they are
ordered on a scale of increasing size. We may draw the critical line between f and b as
well as at any other point of the scale. This is a matter of contextual appropriateness.
But wherever the line is drawn, it creates a yes-or-no division on the scale. Thus,
semantic concepts may be vague in the sense that the boundaries of the resulting
categories can be fixed in a flexible way. But in every given context they must be fixed
somehow and will then yield a simple yes-or-no categorization.
In the case of colour terms, too, categorization is a matter of norms. In the
neutral case, these are defined by the focal colours, but the norms may differ for
288
Understanding semantics
certain kinds of object (cf. red ears, blue lips, green apples, yellow leather, etc.).
When we use a colour term in a particular context, we choose the appropriate focus
hue and draw a circle, as it were, around it that marks the boundary of the resulting
denotation.
Many concepts owe their vagueness to single vague meaning components. For
example, among the three components of the semantic concept ›boy‹ – ›human‹,
›male‹, ›not adult‹ – the component ›not adult‹ is vague since it requires, and
allows, a line to be drawn within the continuum between youths and adults. In the
case of ›cup‹, the central criterion of suitability for hot drinks is vague, because
the suitability depends on secondary features that may vary considerably. The
mathematical concept ›odd‹ is not vague, while the everyday concept ›odd‹ is. In fact,
probably the majority of semantic concepts are in some way vague. This is not a flaw
of natural language but a great achievement because it allows a very flexible use of
the semantic material.
Vague concepts are not deficient. On the contrary, the meaning of big is quite
precise. Its flexibility is due to its adjustability to context. Like an adjustable lens,
it can be focused for the given purpose. But whatever focus we choose in a given
context, the concept then has a certain focus and polarizes accordingly.
The observation of inherent vagueness allows a different explanation for the
flexibility of category boundaries. The PT model explains it with a fixed prototype
and allowance for more or less similarity to it, with flexible use of fixed means, so
to speak. The inherent-vagueness model explains it as variable use of adjustable
means. A vague semantic concept such as ›big‹ allows us to choose an appropriate
norm in the given context (e.g. the average size of grains of corn or the average size
of things mice eat). Consequently, the concept ›big‹ determines a different category
in each context of use, e.g. BIG FOR A GRAIN OF CORN or BIG FOR A MOUSE TO EAT. Thus,
the concept ›big‹, in practice, provides us not with one category but with clusters
of categories, each cluster corresponding to the choice of a norm and alternative
decisions as to where to draw the exact borderline. Such concepts can cope both with
the requirement of flexibility and of polarization.
The PT model and the vague concepts model may be symbolized as in Fig. 11.10.
Figure 11.10
Prototypes vs. vagueness
category defined by
prototype and similarity
category clusters defined by
an adjustable concept
Meaning and cognition
289
In the PT model (left side), the arrows point to single candidate members, thicker
arrows indicating a higher degree of membership. In the other model, circles and
ellipses represent possibilities of drawing boundaries for the category, reflecting, for
example, different choices of norms for size combined with choices of drawing the
borderline between ‘big’ and ‘not big’.
If PS is to maintain the position of widespread graded category membership, the
phenomenon of polarization presents a serious challenge. PS would have to postulate
and model a general mechanism that turns the assumed underlying fuzzy categories
into binary semantic categories, whenever the respective words are used in a natural
language sentence in a particular CoU. This challenge, however, has not yet been
addressed in PT or later developments.
11.5.4 Means of differentiation
Language has several means to cope with the black-and-white nature of the semantic
category system. Three of them will briefly be mentioned.
Hedges. So-called hedges are often cited in PT as evidence for fuzzy category
boundaries. These are expressions that allow a general modification of a categorization
yielding either an expansion, (6a), or a narrowing of the categorization, (6b):
(6)
a.
A futon is kind of a mattress/ … is something like a mattress.
b.
John is a real bike freak.
Such phenomena do not prove that category boundaries are fuzzy and memberships
graded. (6a) does not simply mean that a futon is a mattress; rather it describes a
futon as something that is not a mattress, albeit similar to one. ‘Kind of a mattress‹ is
a vague concept derived from the concept ›mattress‹.
Lexical differentiation. Apart from these general means of differentiation, a
language community and individual language users have the possibility of adapting
the lexicon to the needs of categorization in communication. For example in the
domain of colours, the system of basic colour terms is supplemented in various ways.
There are:
∑ expressions for categories half-way between others: green-blue, blue-grey, etc.
∑ lexicalized hedge categories: whitish, reddish, bluish, greenish
∑ terms for more special categories: crimson, scarlet, vermilion, fiery red, blood-red
and hundreds more
Quantification. Another general means of differentiation is provided by
quantification (recall 4.5). Let us consider first a simple example without quantification:
(7)
a.
The eggs are raw.
b.
The eggs are not raw.
Understanding semantics
290
Together, the sentences form a clear alternative. The eggs referred to either fall
into the category RAW or they do not (recall the discussion on the Presupposition
of Indivisibility in 4.7.2.2). The alternative represents the simple all-or-nothing
contrast symbolized in Fig. 11.11 (where we assume that there are nine eggs; black
dots represent raw eggs).
Figure 11.11
All-or-nothing contrast
the eggs
are not
raw
the eggs
are raw
0
1
2
3
4
5
6
7
8
9
The alternative of (7a) vs (7b) obviously suits only the cases 0 and 9. Pragmatism
may allow the use of the sentences also for cases like 1 and 8, but this is not what is
literally asserted. The intermediate cases with some of the eggs raw and some not
are simply cut out, as it were. The alternative constitutes a system of two categories
that includes only the extreme cases. These are at the same time the simplest cases
in that all the eggs are alike. Although this may seem rather drastic, the strategy
of focusing on simple phenomena and disregarding the more complex is quite
representative of the way in which we make use of semantic categories. A certain
degree of simplification serves communicative economy.
We are not, however, forced to use this black-and-white mode of speaking. For the
grey shades, we can make use of quantification, by specifying the quantity of raw
eggs, or more generally, the number or portion of cases for which the VP is true.
(8)
a.
Some of the eggs are raw.
b.
Many of the eggs are raw.
(8a) is true if at least one egg is raw and false if none is. (8b) specifies the number as
great. Like big, the adjective many is vague and its truth depends on the choice of a
norm. If we compare the number of raw eggs to the rest, we may fix ›many‹ as ›five
or more‹. In other cases it might be appropriate to draw the division line at a lower or
7
higher number. In any event, two things happen when we quantify. First, the whole
range of possible cases between all and nothing is opened up and arranged on a
scale of quantity. Second, at some point of the resulting scale a division is introduced
between truth and falsity of the new sentence. Thus we receive a binary alternative of
categorizations that together cover the intermediate cases as well. This is illustrated
for many in Fig. 11.12.
7 You may wonder why case 0 is included within ‘not many’ and case 9 in ‘many’. If one knows the
exact number, one would express case 0 with no and case 9 with all. But the fact that it is acceptable
to say ‘not many, if any, …’ and ‘many, if not all, …’ proves that the two extremes no and all are
logically compatible with not many and many, respectively, and hence covered by their meanings.
Meaning and cognition
291
Figure 11.12
Quantification with many
not many
many
0
1
2
3
4
more
5
6
7
8
9
more
Note that the resulting picture is very similar to the one for big in Fig. 11.9. The
location of the crucial line that separates true from false, or membership from nonmembership for the quantifier category, depends on the choice of the quantifier. If
we replace many by some, the borderline would shift to between 0 and 1; for all, it
would have to be placed between the maximum and the one immediately below (8
and 9 here).
Modalization. Categorization is very often fraught with uncertainty. In order to
verbalize conditions in the world, we are often forced to make decisions about
how to categorize something without knowing exactly what we ought to know
for that purpose. Of course, language furnishes us with the means of expressing
degrees of certainty in connection with categorization. If I am not certain about
the appropriateness of putting things this way or the other, there are a lot of ‘modal’
expressions at my disposal: adverbs like perhaps, maybe, probably, certainly or sure;
modal verbs like may, will, could or must; or evidentials that qualify the source of
what I express.
(9)
a.
The T-shirt may be dirty.
b.
The T-shirt is perhaps dirty.
But even sentences containing such modal elements are polarized: it is or it is not
the case that the T-shirt is perhaps dirty, etc.
11.5.5 Summary
The adoption of the cognitive perspective as such is no doubt very useful. However,
one central point of PT, graded membership and the resulting concept of fuzzy
category boundaries is inappropriate for semantic categories. When people speak,
human language forces upon them the application of verbalized categorization in
a binary yes-or-no mode (polarization). Word and sentence meanings nevertheless
allow a highly flexible use because many meanings are vague and therefore adjustable
to the particular CoU. Since flexibility is built into the meanings themselves, variable
boundaries are compatible with yes-or-no membership. Apart from the availability
of expressions with a vague meaning, language has several devices that allow for a
differentiation of available categories.
292
Understanding semantics
Figure 11.13
Revision of the central claims of prototype theory
9
Prototypes are
the best examples
but
NO categorization by prototype
therefore
and
NO graded membership
9 Graded structure
but
NO fuzzy boundaries
Figure 11.13 shows the main results of our discussion of PT. If you compare this to
Fig. 11.4 above, you will realize that the claims of PT represented on the right side of
the figure are questioned. What remains of the original picture is the notion of graded
structure, reflected in the existence of better and less good examples. This observation
is very valuable and its explanation is the objective of much, and promising, research
in the area. What is to be refuted, however, is the idea that the graded structure is a
direct reflection of different degrees of representativity. Furthermore, it is certainly
necessary to provide for more than one model of categorization. There may be
categories for which membership is a matter of matching with the prototype (in
certain, relevant respects). For other categories, the NSC model may be more
adequate. Yet other models may be required for further types of categorys. Likewise,
the notion of fuzziness is not altogether inadequate for semantic categories. Of
course, vagueness and context-dependence give rise to ‘fuzziness’ in some sense.
But, at least for semantic concepts, the variability of category boundaries must be
reconciled with polarization, i.e. with binary membership.
11.6 SEMANTIC KNOWLEDGE
11.6.1 Personal knowledge vs cultural knowledge
In his textbook Language, written long before the emergence of cognitive science,
Sapir characterized the role of semantic categories in language use as follows:
The world of our experiences must be enormously simplified and generalized
before it is possible to make a symbolic inventory of all our experiences of things
and relations and this inventory is imperative before we can convey ideas. The
elements of language, the symbols that ticket off experience, must therefore
be associated with whole groups, delimited classes, of experience rather than
with the single experiences themselves. Only so is communication possible,
for the single experience lodges in an individual consciousness and is, strictly
Meaning and cognition
293
speaking, incommunicable. To be communicated it needs to be referred to a class
[ = category, S.L.] which is tacitly accepted by the community as an identity.
(Sapir 1921: 12f.)
The passage illustrates the necessity of distinguishing between personal knowledge
of a given category and the knowledge that makes up the category for the community
we are a member of. For example, every one of us knows several things about apples.
Some of what we know is personal knowledge, e.g. how the particular apples we ate
tasted, which role which sorts of apples play in our personal diet, where we can buy
apples in our neighbourhood, etc. But part of our knowledge is common property
of the cultural community we belong to. It includes the knowledge that apples are
fruits, what they look like, how they taste, that they contain vitamins, where they
come from, how they are eaten or prepared, how they are traded, what they cost,
that they are used to produce apple juice, etc. Let us call this knowledge cultural
knowledge as opposed to our personal knowledge. Cultural knowledge defines
what can be called cultural categories, the categories which are ‘tacitly accepted by
the community as an identity’. Names of cultural categories will be marked with
a subscript ‘C’. For example, APPLEC is the cultural category of apples, defined by
our present cultural knowledge of apples. Cultural knowledge is not to be taken as
‘everything that is known about the category’. It does not include expert knowledge.
For example, knowledge of the complex biochemistry of apples is not part of the
cultural concept. Personal knowledge usually includes the greater part of cultural
knowledge, depending on the range of experience of the individual. But we are all
some way apart from commanding all cultural knowledge, as for us all there are
8
many areas of everyday life where we have little experience.
Given that the speech community makes use of cultural categories, the question
arises of how much of it constitutes the meaning of words. Is the meaning of the word
apple the concept for the cultural category APPLEC? If so, every detail that belongs
to the cultural knowledge of apples is part of the meaning of the word. If not, the
meaning of a word is only part of the cultural concept. Aspects such as what apples
normally cost in the supermarket, what sorts of apple one can buy there and how
one prepares baked apples might then not be considered components of the word
meaning.
The distinction between semantic knowledge and ‘world knowledge’ (i.e. cultural
and personal knowledge) is a doctrine of traditional semantics. It can be plausibly
argued that we do not need to have the total cultural knowledge about apples,
computers, mice or alcohol in order to know the meaning of the words apple,
computer, mouse and alcohol. Many researchers of the cognitive orientation are,
however, pleading for abandoning the distinction. They argue that our semantic
categories are interwoven with our overall cognitive system in a way that does not
8 We have chosen the term cultural knowledge instead of world knowledge and encyclopedic
knowledge, either of which you may frequently encounter. Often the notion world knowledge is used
indiscriminately for cultural knowledge and personal knowledge, while encyclopedic knowledge is
understood as including expert knowledge.
294
Understanding semantics
allow their separation. Also, they argue, the full meaning of a word can only be
grasped in the context of larger parts of the cognitive system. This view resembles
the structuralist position in emphasizing the interdependence of categories, but
it does not primarily relate to semantic categories. Rather it aims at the way in
which our categories, among them the semantic categories, are integrated into a
full model of the world, the world as we perceive and interpret it, in which we are
placed ourselves, interacting with it. In the following, we will nevertheless argue
that a distinction between cultural knowledge and semantic knowledge is necessary,
important and feasible.
11.6.2 The apple juice question
Ungerer and Schmid (2006), a recent introduction to cognitive linguistics, which
focuses on cognitive semantics, explicitly subscribes to an approach that tries to
capture meaning in terms of the experiential knowledge that laypersons connect
with words. The approach is representative for a position contrary to a distinction
between cultural knowledge and semantic knowledge. Let us call it the ‘cultural
knowledge approach’ (CKA, for short). The authors report on an experiment for
determining what makes up the category APPLE JUICE (Ungerer and Schmid 2006:
92f.). Lay subjects were asked to list features of apple juice. It was then checked
whether these features could also be assigned to either or both or neither of the
categories APPLE and JUICE, in order to determine where the features stem from. The
result is given in Table 11.2. In addition to these properties, ›made from apples‹ is
listed as ‘salient specific attribute’ (Ungerer and Schmid’s term attribute is replaced
by feature in order to keep to the terminology chosen here).
Table 11.2
Features of the category APPLE JUICE according to Ungerer and Schmid (2006: 93)
Salient specific feature: ›made from apples‹
features that the category APPLE JUICE shares with the categories …
1 … JUICE
›liquid‹
›no alcohol‹
›thirst-quenching‹
›supplied in bottles or cartons‹
›served in glasses‹, etc.
2 … JUICE and APPLE
›sweet or sour-sweet‹
›healthy‹
›tastes good‹
3 … APPLE
›yellow or similar colour‹
›fruity‹, etc.
4 with neither
›mixed with soda water‹
›naturally cloudy‹
Meaning and cognition
295
The word apple juice is a regular compound of the two nouns apple and juice.
According to the ‘classical’ analysis briefly sketched in (8.1.2), its meaning is the
meaning of the head noun plus a specification added by the modifier. In favour
of CKA, Ungerer and Schmid argue that their results provide evidence against the
traditional analysis. If it were correct, they claim, most features of the category APPLE
JUICE should stem from the head category, only one should be contributed by the
modifier (›made of apples‹) and no extra features should hold for the compound only.
The latter requirement follows from the fact that the meaning of regular compounds
should completely derive from the meanings of their parts. Ungerer and Schmid
point out that, however, the category APPLE JUICE is linked to the category APPLE
by quite a number of further common features (rows 2 and 3) and exhibits some
genuine features of its own (row 4).
While the results of the experiment are as they are, it must be questioned if the
subjects described what for them constitutes the meaning of the word apple juice.
Apparently they described the cultural knowledge they commanded and held
relevant. However, a critical look at the features listed shows that most of the features
elicited cannot be considered parts of the meaning of the word.
How can we decide if a particular feature of a category is part of the meaning of
the corresponding word? There are (at least) two tests. First we can check if it is
necessary to know that this is a feature of the category in order to know what the
word means. Second, we can carry out the following thought experiment: if the actual
members of the category happened to lack the feature in question, could we still
use the word for referring to them? Such thought experiments are relevant because
the meaning of a word must fit all potential referents. There is no way of exploring
concepts other than by asking oneself what they would possibly cover.
Let us apply the tests to the critical features in Table 11.2, i.e. those in rows 3 and
4, which according to Ungerer and Schmid contradict the classical analysis. For the
thought experiment, imagine the food industry decides that apple juice needs a
radical change in order to sell better. Thus they filter the stuff, add blue colour and
heavy peppermint flavour and promote it as a definitely-no-soda drink. It would
cease to have any of the critical features and would yet be apple juice, because it
would still be made from apples. The result of the experiment is in accordance with
what we find when we ask ourselves what one needs to know about apple juice (the
stuff) in order to be entitled to claim knowledge of the meaning of the word apple
juice. One does not need to know its colour or if it is cloudy. In fact, as we all know,
there is cloudy as well as clear apple juice. Did you know that potato juice from
ordinary potatoes turns reddish in a few minutes? If not, would you say you did not
understand the word potato juice in the previous sentence? The other features in the
table have the same status. One need not know the taste of apple juice in order to
know the word meaning. (Have you ever tasted potato juice?) And one need not know
the soda water thing. There is one feature that really carries the whole load: ›made of
apples‹. The word apple juice means ›juice made of apples‹. Period. If one knows that,
one knows the meaning of the word. Apple juice being quite a common sort of drink,
the other features are likely candidates for the cultural knowledge about apple juice,
but they do not form components of the meaning of the word apple juice.
296
Understanding semantics
Thus the classical semantic analysis of compounds is not disproved by the
experiment. Rather, the discussion proves the importance of distinguishing between
cultural knowledge and semantic knowledge.
11.6.3 Cultural knowledge vs semantic knowledge
From this the following picture emerges. To any word we know (e.g. apple juice), we
connect in our personal cognitive system a semantic concept, its meaning (›apple
juice‹ = ›juice made of apples‹). The meaning determines a fairly broad category, the
denotation of the word. Let us mark denotations, or semantic categories, with the
subscript S for ‘semantic’. The category APPLE JUICES contains all possible variants of
apple juice including the clear, blue, mint-flavoured, no-soda variant of our thought
experiment. The category APPLE JUICES is broader than the presently valid cultural
category APPLE JUICEC which is characterized by additional features such as those
elicited in the experiment reported in Ungerer and Schmid (2006). The cultural
concept that represents APPLE JUICEC is correspondingly more specific than the leaner
concept that constitutes the word meaning. This is an iron law of cognition: the more
specific a concept, the narrower the category it represents, and vice versa. Thus, the
meaning of the word apple juice, the semantic concept ›apple juice‹, is part of the
cultural concept for apple juice, but the cultural category APPLE JUICEC is a subcategory
(!) of the semantic category APPLE JUICES. In Fig. 11.14, the cultural concept and the
cultural category are integrated into the semiotic triangle. Note how the more specific
cultural concept includes the leaner semantic concept, while, conversely, the cultural
category is only a subset of the semantic category.
Figure 11.14
The semiotic triangle integrating cultural knowledge
expression
cultural concept
meaning
determines
cultural
category
represents
The cultural category can be roughly equated with the actual denotation of the
word, i.e. those members of the total denotation that we encounter, or expect to
encounter, in actual life. These form only a subset of those we might conceive of, i.e.
Meaning and cognition
297
all potential referents of the word. Theorists that equate semantic knowledge with
cultural knowledge equate the actual denotation with the total denotation.
An approach that distinguishes between semantic knowledge and cultural
knowledge, or between the meaning of a word and the cultural concept into which it
is embedded, is superior to CKA in offering an explanation for the following points:
∑ the stability of word meanings as compared to cultural concepts
∑ the abstractness of word meanings as compared to cultural concepts
∑ the communicational economy of word meanings
∑ the simplicity of meaning relations between words
Let us conclude this section with a brief illustration of these points.
Stability. Cultural knowledge is subject to constant change. For example, certain
categories of artefacts such as TELEPHONE or COMPUTER have changed to a degree
that people of, say, two generations back, would have been unable to anticipate.
Nevertheless, the telephones and computers of the 1950s are no less members of the
respective semantic categories than today’s telephones and computers. Both are 100
per cent members of the semantic categories TELEPHONES and COMPUTERS. While the
cultural categories have shifted, e.g. with the invention of the mobile telephone or
the personal computer, we do not feel that the words telephone and computer have
changed their meanings. Rather word meanings form a stable core of the changing
cultural concepts. Constant meanings allow for communication with constant means
in an ever-changing world. Of course, meanings are not completely constant. The
lexicon keeps adapting to the changing communicational needs, but at a much larger
temporal scale.
Abstractness. The lack of many concrete features that narrow down the cultural
categories also accounts for the abstractness of semantic concepts, which allows their
application to uncommon cases outside the familiar cultural categories. One example
was the applicability of the semantic concept ›apple juice‹ to exotic fictitious varieties
of apple juice. Similar thought experiments could be conducted for almost all other
semantic categories.
Economy. Although the greater part of cultural knowledge does not form part of
the word meaning, it is nevertheless relevant to communication. In conversation,
we draw heavily on all sorts of common knowledge. This allows us to keep our
messages semantically to a minimum. For example, if John says (10a) to Mary,
indicating thereby that Mary should bring him some apple juice, he will rely upon
his and Mary’s common cultural knowledge about the packaging of apple juice. If
John’s apple juice is usually packed in bottles, he need not say (10b), yet Mary will
understand as much and be able to locate the juice among the things on the balcony.
(10) a.
b.
There is apple juice on the balcony.
There is apple juice in bottles on the balcony.
298
Understanding semantics
The fact that Mary understands something like (10b) when John says (10a) does
not mean that ›is in bottles‹ is part of the meaning of the word apple juice. On the
contrary, the feature need not be part of the word meaning because it is part of the
additional cultural knowledge.
Meaning relations. Last but not least it must be stated that CKA approaches blur
rather than explain meaning relations such as those between a compound and its
parts. The simple meaning relation between the head and the compound is buried
under a heap of cultural features.
11.7 SUMMARY
This chapter took us through a series of fundamental issues. First, the notions
of meaning as a concept and denotation as a category were embedded into the
larger context of cognitive science. Here the most basic issue is the general nature
of categorization. The traditional view, represented, for example, by the BFA,
considers membership in a particular category a matter of a fixed set of necessary
conditions to be fulfilled by its members. The view implies that membership in
a category is a clear yes-or-no issue. Given a particular category, anything either
belongs to it or it does not. Hence, all members of a category enjoy equal member
status. As a consequence, a sharp category boundary separates members from
non-members (11.2.1). This view was challenged by PT. No! PT says, category
boundaries are not sharp, but fuzzy; there are better and poorer members; in
particular there are best examples: the prototypes. On the other hand, categories
may have very poor members. Thus, no clear distinction can be drawn between
members and non-members, rather membership is a matter of degree. As for
the defining ‘necessary’ conditions, even the most typical ones such as ability to
fly in the case of birds may be violated. So: dispense with necessary conditions!
Categorization is not a matter of checking conditions but is carried out by a
comparison with the prototype (11.2).
While all this has its appeal, the PT approach proved problematic under closer
scrutiny. For one thing, its results do not apply to categorization as generally as
was claimed and are not as cogent as it first appeared (11.4). When PT is applied to
semantics, severe problems arise with the notion of graded membership: it is in direct
conflict with polarization, probably the most fundamental semantic phenomenon
(11.5.2). Polarization totally pervades word and sentence meaning. It underlies each
predication – and therefore almost all word meanings. And since each sentence
contains at least one predication (because it must contain a finite verb and all verbs
are predicate terms), each single sentence is subject to polarization. Consequently,
categorization by verbal means is always binary, i.e. a yes-or-no matter. Despite such
problems, PT did contribute valuable results to semantics. The observation that
for most categories we operate with prototypes, for example as default cases when
no further specification is available, is an important result. A further important
contribution of PT is the detection and specification of the basic level (11.3).
Meaning and cognition
299
In tackling the PT claims about fuzzy category boundaries and graded membership
we obtained insights into the nature of word meanings and the ways in which
language, this polarizing (or polarized) cognitive equipment, copes with the
non-binary character of the ‘world’. For one thing, vague meanings enable a
flexible categorization. There is no need for concepts that would provide fuzzy
category boundaries, because the word meanings available enable us to adapt
category boundaries to the requirements of communication (11.5.3). In addition
to flexible word meanings, other linguistic devices such as lexical differentiation or
quantification allow for relativization of what would otherwise be heavy black-andwhite thinking (11.5.4).
No doubt the main merit of PT is its focus on the conceptual level. The other two
mainstreams of semantics, the structuralist tradition and formal semantics (to be
introduced in the final chapter), avoid addressing meaning itself. The structuralist
tradition approaches meaning indirectly via the investigation of meaning relations.
(Recall the claim of structuralism that the meaning of an expression is the sum of its
relations to other expressions within the system.) Formal semantics tries to capture
meaning equally indirectly by investigating reference and truth conditions.
Addressing the conceptual level itself raises another fundamental problem, the role
of linguistic meaning within our overall cognitive system (11.6). Are the concepts
that are assumed to constitute the meanings of words identical with the concepts
that define our everyday categories? Given that language is a general means of
communication within a cultural community, a distinction must be drawn between
the knowledge at our personal disposal and that part of it which we may assume to
be (by and large) shared by the other users of the language. The latter was dubbed
‘cultural knowledge’ as opposed to ‘personal knowledge’. A closer consideration of
regular compounds led to the conclusion that cultural knowledge must be further
distinguished from semantic knowledge. Semantic concepts are much leaner than
the corresponding cultural concepts. Only this assumption allows an account of basic
phenomena such as simple meaning relations and the stability of word meanings.
Within the given limits, it is not possible to present all of the major cognitive
approaches to semantics in a way that would do justice to the whole field. One area
should be mentioned where modern developments of cognitive semantics are very
successful: the investigation of polysemy and, closely related, semantic change.
In particular, it was shown that metaphor and metonymy are general cognitive
mechanisms that play a central role in meaning variation and historical meaning
shifts. In very many cases, meaning variants of the same word are linked by a
metaphorical or metonymical relation (or both).
We focused on the earlier claims of PT because this allowed us to address several
issues of central importance. Meanwhile ‘Cognitive Semantics’ is an established field
of research and theory. But it must be stated that it is still far from a full-fledged
semantic theory. Central fields of semantics have never been elaborated within the
cognitive framework. These include meaning relations and composition. In part this
is a consequence of particular assumptions. The assumption of widespread graded
membership would render an account of the composition of sentence meaning
very difficult. Similarly, the identification of semantic knowledge with cultural
300
Understanding semantics
knowledge impedes the proper treatment of meaning relations (11.6.2). But none
of these assumptions are necessary components of a cognitive semantic theory. In
the next chapter, a different theory of cognitive representation promises to be able
to overcome some of the basic shortcomings of the older approaches discussed here.
EXERCISES
1. Explain the difference and relation between concepts and categories.
2. Explain the difference between mushroom, MUSHROOM and ›mushroom‹.
3. What kind of entity is a prototype?
4. What is the role of prototypes in PT?
5. What is problematic with the notion of ‘similarity to the prototype’?
6. What is meant by ‘polarization’ and why does it pose a serious problem to
prototype semantics?
7. Perform an experiment asking several people a question like the one at the
beginning of 11.3.1. Discuss the result.
8. Consider the following terms. Using the criteria mentioned in 11.3.2, try to assign
the terms to the basic level, to a more general (higher) or to a more specific level.
television set, radio, toaster, washing machine, portable TV, cell phone, iPhone,
household appliance, PC, notebook (electronic), electric appliance
9. Ask three people spontaneously to write down 30 terms for clothing in the order
in which they come to mind.
a. Try to assess which kinds of clothing are prototypical.
b. Try to arrange the terms in a taxonomy. Which ones are basic level terms?
10. Discuss the distinction between personal knowledge, cultural knowledge and
semantic knowledge using the word money as an example. Try to find features
that (a) are part of your personal knowledge about money, but not of the cultural
knowledge and (b) are part of (present) cultural knowledge but not necessarily
part of the meaning of the word money.
FURTHER READING
For a discussion of PT see Aitchison (2012, chs 5, 6); Ungerer and Schmid (2006,
ch. 1); and Lakoff (1987, ch. 2) for its history. On the hierarchical organization of
the category system see Ungerer and Schmid (2006, ch. 3). On the role of metaphor
and metonymy in cognitive semantics, see Ungerer and Schmid (2006, ch. 4), Palmer
(1996, ch. 8), Foley (1997, ch. 9) and the case study on the category ANGER in Lakoff
(1987: 380–415).
12
Frames
In this book, it is assumed that meanings are concepts. We have talked a lot about
the content of these concepts and the relations among them. We have learnt about
general aspects of categorization and concept formation. But up to now, no proposal
has been presented as to what these concepts are like: What is their structure? What
are their conceptual components? This will be finally done in this chapter, though the
answer to these questions is, necessarily, preliminary and incomplete. There is one
theory of concepts in cognitive psychology which has come up with a concrete model
1
of concepts: Lawrence W. Barsalou’s theory of cognitive frames. The theory claims
that all concepts in human cognition are ‘frames’ and presents various empirical
evidence and theoretical arguments in favour of this claim. Barsalou’s theory of
cognitive representations is superior to the binary feature approach (BFA) as well as
to prototype theory. Unlike BFA, it is much less restricted; unlike prototype theory, it
is explicit about the internal structure of concepts. It is very attractive for linguists,
as it offers solutions to many problems of decompositional semantics; it also offers
the possibility of tackling semantic composition. The application of frame theory
in semantics and in linguistics in general is still under development, but there is
enough to be said about this promising theory of cognitive representations in order
to include it here.
12.1 BARSALOU FRAMES
12.1.1 Chunks of knowledge
The notion of ‘frame’ belongs to a family of similar notions, including ‘schema’,
‘scenario’ and ‘script’, that came up in the 1970s in the fields of artificial intelligence
and cognitive science. These were notions for the ready-made complex ‘chunks of
knowledge’ which we apply in everyday life. For instance, there is the schema (or
frame) ‘university’; it consists of the typical ‘ingredients’ of a university, like its
campus, its faculties, its teachers, lectures, students, courses and study programmes,
examinations, etc.; or the ‘family’ frame that defines the members of a family,
the relationships between them, their responsibilities, etc. There is the ‘script’ or
‘scenario’ of a children’s birthday party or of dining at a restaurant, shopping in a
supermarket, taking an exam or getting married. Such a framework is defined by
constitutive ingredients such as persons in specific roles, engaged in particular
1 The theory was introduced in Barsalou (1992a, 1992b).
302
Understanding semantics
interactions, pursuing certain aims, etc.; their actions will involve certain objects
and locations, they may be undertaken in a particular order, etc. Just think about
eating in a cafeteria: we have the kind of location with its fittings and furniture that
would make up a cafeteria, the staff and customers that will be there, the kind of
food and beverages you may find, the procedure applied to get your meal, pay for it,
take a place, eat and drink and dispose of your rubbish. Chunks of knowledge like
the cafeteria schema form our general world knowledge. They are tacitly employed
as background knowledge when we mention visiting a cafeteria or other activities.
There is no doubt that our cognitive systems host a vast number of such chunks of
knowledge which are interwoven which each other and allow us to activate large
networks of knowledge in a very short time.
In the literature on artificial intelligence or cognitive linguistics, no precise
2
formal definitions are provided for a general format of such chunks of knowledge.
Barsalou was the first to come up with a model of considerable precision and with
the claim that this is the actual format in which our minds organize concepts to
represent the categories of things and phenomena in the world. Barsalou’s is not
primarily a theory of semantic representations. Much of his empirical work was
spent on concept formation in particular contexts and experimental settings. He
emphasizes that concept formation is flexible and able to adapt to the needs and
circumstances of a given situation. Concepts can be arbitrarily enriched or thinned
out or otherwise modified. This general flexibility does not apply, though, to stored
lexical meanings, which is what we will apply frame theory to. Lexical meanings
form stable cognitive entries only subject to the general process of life-long
language learning.
Barsalou frames are cognitive representations of single objects, either individual
objects or objects as members of a category. Compared to these frames, the chunks
of knowledge envisaged in artificial intelligence are macrostructures that integrate a
greater number of objects, events, actions, etc. in a much more complex whole.
12.1.2 The passport frame
The general notion of a frame (henceforth ‘frame’ is to be understood as ‘Barsalou
frame’) can be illustrated by means of an example we all know: the information
that is provided in a passport for the identification of its bearer. Passports issued by
a certain country use a fixed paradigm of description for the holder. The example
we will give is based on the European Union passport (as of 2012). It contains the
following data on its bearer: name, given names, nationality, date of birth, place
of birth, sex, residence, height and colour of eyes; these are specified in words. In
addition, there are two non-verbal pieces of information: a token of the bearer’s
signature (this is primarily non-verbal because it’s the mere written shape that
matters, readable or not), and a photograph of the bearer’s face. In the context of
frames, these parameters of description are called attributes; we will use small
2 For example, Kövecses (2006: 64) gives the following working definition of frames: ‘A frame is a
structured mental representation of a conceptual category.’
Frames
303
capitals for attributes. The attributes take values, e.g. the value Angelika for the
attribute GIVEN NAMES or the value ‘female’ for the attribute SEX. The values can be of
various kinds, depending on the attribute. It may or may not be possible to describe
the values in words (a signature or a face cannot be described in words, although it
can of course be referred to). Figure 12.1 gives the data in a fictitious passport in a
so-called frame matrix.
Figure 12.1
Frame matrix for an EU passport bearer (1)
bearer
.NAME
.GIVEN NAMES
.NATIONALITY
.DATE OF BIRTH
.PLACE OF BIRTH
.SEX
.HEIGHT
.COLOUR OF EYES
.RESIDENCE
.SIGNATURE
.[FACE]
: Postowski
: Angelika
: German
: 03.08.1971
: Bottrop
: female
: 178 cm
: blue
: Köln
: (a signature)
: (a photograph)
The bearer appearing as the only entry in the first column is the object that the
whole frame describes; all entries in the second column relate to this item. The
second column of the matrix lists the attributes, the third column the values they
take. Each line represents one piece of information about the bearer; for example
‘bearer.NAME: Postowski’ which is to be read as ‘the name of the bearer is Postowski’,
or ‘bearer.COLOUR OF EYES: blue’. The attribute FACE is given in square brackets because
it does not appear explicitly in the passport.
The matrix contains eleven attributes of the bearer and their respective values,
as far as they can be verbalized. Four of the attributes take a complex formulation:
DATE OF BIRTH, PLACE OF BIRTH, COLOUR OF EYES and SIGNATURE OF BEARER. The last one
makes explicit the possessor of the attribute which is left implicit in all the other
cases. The other three entries can be decomposed into two steps of attributing. The
term ‘colour of eyes’ refers to the colour of the eyes of the bearer; this is, in a first
step, the value of the attribute COLOUR of the eyes of the bearer; in a second step, the
eyes of the bearer can be analysed as the value of the attribute EYES of the bearer.
Thus, the attribute COLOUR OF EYES is the result of a chain of two attributes; it is the
attribute COLOUR of the value of the attribute EYES of the bearer. This decomposition
is given in (1a). The attributes PLACE OF BIRTH and DATE OF BIRTH can be analysed as
involving the attribute BIRTH of the bearer. They can be analogously decomposed as
in (1b, c):
304
(1)
Understanding semantics
a.
bearer.COLOUR OF EYES: blue
= bearer.EYES: [bearer’s eyes].COLOUR: blue
b.
bearer.DATE OF BIRTH: 03.08.1979
= bearer.birth:[bearer’s birth].date: 03.08.1979
c.
bearer.PLACE OF BIRTH: Bottrop
= bearer.birth:[bearer’s birth].place: Bottrop
The attributes BIRTH and EYES open two small subframes for their respective values:
one represents the birth of the bearer with the attributes DATE and PLACE, the other, the
eyes of the bearer with the only attribute COLOUR. These subframes can be unfolded
in the frame matrix as in Fig. 12.2. Square brackets indicate implicit attributes and
values.
Figure 12.2
Frame matrix for an EU passport bearer (2)
bearer
.NAME
.GIVEN NAMES
.NATIONALITY
.[BIRTH]
: Postowski
: Angelika
: German
: [bearer's birth]
.SEX
.HEIGHT
.[EYES]
.RESIDENCE
.SIGNATURE
.[FACE]
: female
: 178 cm
: [bearer's eyes] .COLOUR :blue
: Köln
: (a signature)
: (a photograph)
.DATE
.PLACE
: 03.08.1971
: Bottrop
The example illustrates a very important property of frames: they are in principle
recursive, i.e. frames may contain embedded subframes which add information on
the value of an attribute by adding to it its own attributes and their values.
The basic unit of frame description is the ascription of an attribute and its value
to either the referent of the whole frame, or to the value of an attribute in the frame:
a piece of information of the form ‘_____.ATTRIBUTE: value’. Frames can not only
be represented by matrices, but also by graphs of nodes and arrows. This mode of
representation is more space-consuming, but it is easier to comprehend. In a frame
graph, the object of representation and the values of attributes are represented as
nodes. The node representing the object of description of the whole frame is encircled
with a double line; it will be called the ‘referent node’. Attributes are represented by
arrows pointing to the value; the arrows carry a label that names the attribute. Thus,
the basic unit in a frame graph takes the form shown in Fig. 12.3. The node to which
the attribute is applied is represented by a broken line circle; we will refer to it as the
‘possessor’ of the attribute.
Frames
305
Figure 12.3
Attribute-value element of a frame graph
(possessor)
value
ATTRIBUTE
Figure 12.4 displays the graph equivalent of the frame matrix in Fig. 12.2.
Figure 12.4
Frame graph for an EU passport bearer
NAME
GIVEN NAMES
NATIONALITY
DATE
[BIRTH]
PLACE
bearer
SEX
HEIGHT
[EYES]
COLOUR
RESIDENCE
SIGNATURE
[FACE]
A passport contains not only information about the bearer, but also about the
document itself. It specifies, among other details, a unique passport number, a type
( = ‘P’), dates of issue and of expiry, the issuing authority – and of course the bearer.
The bearer is an attribute of the passport, and thus the whole bearer frame can be
regarded as embedded in a superordinate passport frame. This is depicted in the
frame matrix in Fig. 12.5. The frame box containing ‘[bearer]’ indicates the frame
matrix in Fig. 12.2.
Understanding semantics
306
Figure 12.5
Frame matrix for an EU passport
passport .NUMBER
.TYPE
. DATE OF ISSUE
. DATE OF EXPIRY
. AUTHORITY
.[BEARER]
: xxxxxxxxx
:P
: 16.03.2012
: 15.03.2022
: Stadt Köln
: bearer
….
….
….
As we saw with the examples of DATE OF BIRTH and COLOUR OF EYES, attributes can
form chains; the general structure is as in (3a), graphically as in (3b).
(2)
a.
possessor.ATTRIBUTE1:value1.ATTRIBUTE2:value2.ATTRIBUTE3:value3 …
poss
b.
value1
ATTR1
ATTR2
value2
ATTR3
value3
…
The passport frame is an example of what is called ‘institutional categorization’, i.e.
a socially established standardized way of categorization. There are countless other
examples essentially in frame format, e.g. telephone registries or library catalogues,
where each entry amounts to a frame format description with the same attributes.
We actually used a frame in 3.1 above when we defined the notion of the lexeme. It
can be transformed into a frame matrix such as in Fig. 12.6.
Figure 12.6
Frame matrix for the lexeme ‘teacher’
lexeme
.SOUND FORM
.WRITTEN FORM
.MEANING
.CATEGORY
: ['ti:t]
: 'teacher'
: ›teacher‹
: noun
In this lexical frame, the entry for the value of the attribute MEANING would have to
be filled by the frame representing the complex concept for a teacher.
12.1.3 The basic structure of a frame
A frame is the description of a potential referent. This can be a particular object,
as in the passport example, or an arbitrary member of the category described – if
the frame is not specific enough to single out a particular case. In the applications
Frames
307
considered in the following, frames will always be of the latter kind. The potential
referent of the frame is represented by the referent node in a frame graph or by the
only element in the first column of a frame matrix. The frame contains attributes;
each attribute assigns a value to the potential referent; further attributes may assign
values to these values and so on. The frame structure is subject to a set of uniqueness
conditions; these are the essential conditions that define a frame. For different types
of concepts there may be additional conditions imposed on their structure:
DEFINITION 1 Frame
A frame is a conceptual network of attribute-value assignments that fulfils the
following uniqueness conditions:
UR Unique frame referent
There is a unique element that represents the potential referent of the
frame. Every element in the frame is connected to the frame referent by a
chain of attributes.
UA Unique attributes
For each element in the frame, an attribute is assigned no more than once.
UV Unique values
For each element in the frame, and each attribute applied to it, the
attribute takes a unique value.
If you take a look at the examples, you see that UR is fulfilled: there is a particular
element representing the referent and all other elements are connected to it. In a
frame matrix, the central element is represented by the only entry in the first column;
in a frame graph, the central element is the node marked as a double-lined node,
from which every other node can be reached.
For each element in the frame, the same attribute is never applied more than once
(UA). Also, the attributes are understood as taking a unique value (UV). The value
may be complex; e.g. the value of GIVEN NAMES may be a series of two or more given
names; but then the whole constitutes one value of the attribute. The value of an
attribute need not be specified with ultimate precision – it may even be left open in
certain cases; for example, in a frame for the concept ›person‹, one might provide for
the attribute SEX without fixing its value. However, the attribute must be such that for
a given possessor it returns exactly one value.
12.1.4 Attributes are functional concepts
It follows – and this point is of utmost importance – that ATTRIBUTES ARE FUNCTIONAL
Barsalou is correct in assuming that frames are the format of cognitive
representations, it follows that our mental descriptions of objects and categories are
entirely in terms of functional concepts and the values they take. Functional concepts
correspond to functions in the mathematical sense in that they return a unique value
CONCEPTS. If
308
Understanding semantics
for any given argument. The only difference between attributes and mathematical
functions is that the arguments and values of attributes are not restricted to
mathematical entities such as numbers, sets, functions, etc; arguments and values
of attributes can be of any type whatsoever. Attributes are subject to something like
selectional restrictions: they can be applied only to a certain domain of arguments.
For example, the attribute SEX can only be applied to certain organisms, HEIGHT only
to objects with a vertical extension, NAME only to nameable objects, and so on. In
addition, attributes take certain types of things as values. For example, SEX takes the
values ‘female’ or ‘male’, HEIGHT a measure of length, NAME some verbal expression
subject to complicated social conventions of naming.
We encountered functional concepts when we dealt with functional nouns in
4.4.2.1: the meanings of functional nouns are functional concepts. If you look at
the attribute terms in the passport example, you will realize that, in the readings
used in the passport, these are functional nouns: they are relational in that they
need a possessor argument for determining their referents and they refer to exactly
one referent for a given possessor. The text of the passport contains the functional
nouns name, nationality, birth, date, place, sex, height, colour, residence and signature;
except for signature, they come without a possessor specification because they are
tacitly understood as relating to the bearer of the passport. The noun name in the
combination given names is used as a relational noun; obviously, the plural admits
for more than one referent. However, in the plural, referring to the totality of given
names, it provides a functional concept. Similarly, the relational concept ›eye‹ yields
a functional concept ›eyes‹ when turned into the plural.
Since attributes are the elementary element of frame composition, frames for
attributes themselves are extremely simple. A frame graph for an attribute contains
a single arrow, labelled for that attribute. It originates from a node that represents its
possessor, or argument. Since this argument is to be filled, we represent it not with
a circle, but with an empty rectangle. The attribute links the argument to its value.
The value forms the referent of the frame. This corresponds to the fact that for an
expression such as price of oil, price expresses a functional concept to be applied to
the referent of oil; the whole expression refers to the value of the ‘price’ function for
oil, i.e. to the price of oil. Fig. 12.7 displays the general frame graph for an attribute
concept.
Figure 12.7
Frame for an attribute
ATTRIBUTE
value
The chaining of attributes mentioned above, repeated in (3a), corresponds to the
chaining of the underlying functions (3b); the mechanism of chaining mathematical
functions is known as ‘function composition’ in mathematics.
Frames
(3)
a.
possessor.ATTRIBUTE1:value1.ATTRIBUTE2:value2.ATTRIBUTE3:value3
b.
f3 (f2 (f1(x)))
309
The chaining mechanism of functions and attributes is replicated in language by the
possibility of forming possessive chain constructions as in (4a) and, less elegantly, in
(4b). The order of elements in (4a) corresponds to the order in (3b), the order of the
Saxon genitive chain, to (3a):
(4)
a.
the colour of the eyes of the bearer of the passport
b.
the passport’s bearer’s eyes’ colour.
The passport frame has terms for each attribute (except for FACE) that it specifies,
and these terms are functional nouns or complex expressions with functional
concept meanings, in the readings relevant here. While this is necessary for frames
of institutional categorization, we cannot expect that there will be words to label all
the attributes that figure in all the frames we supposedly employ in our cognitive
systems. On the contrary, it is to be expected that for most, if not almost all, attributes
we do not have a functional noun in our lexicon. As we stated in 11.1, cognitive
representations are not in terms of words.
12.1.5 Types of attributes
Possible attributes in frames can be classified into four basic types according to the
types of values they assign. All four classes are instantiated in the passport frame:
(i) constitutive part attributes, (ii) correlate attributes, (iii) property attributes and
(iv) event attributes.
Part attributes. Frames for the representation of complex things like cats or pianos
have attributes for their constitutive parts. They contain a mereology of the object
(recall 8.4.4). For example, a frame for a person will contain a subframe for the
body, which in turn will contain two embedded subframes for the arms, each with a
sub-subframe for the hand containing the attributes PALM, THUMB, INDEX FINGER, etc.,
each of them with their own mereologies. Due to UA, a frame must not contain the
same kind of attribute more than once for one possessor. Thus, in the mereological
subframe for the body, there must not be two attributes ARM, but different attributes
such as LEFT ARM and RIGHT ARM. The values of the part attributes are the respective
parts of the whole (the value of the attribute HEAD in a cat frame is the head of the
cat). The constitutive parts of a whole do not exist independently; if the whole is
given, its parts are too. In the passport frame, there are two part attributes, the
attributes EYES and FACE of the bearer.
Correlate attributes. Correlate attributes specify things of independent existence
to which the referent of the concept is uniquely related. For pianos this may be
attributes such as PRODUCER or OWNER; for cats it might be MOTHER or OWNER. The
passport example contains six correlate attributes: BEARER, NUMBER and AUTHORITY for
the passport; NAME, GIVEN NAMES and RESIDENCE for the bearer.
310
Understanding semantics
Property attributes. This group is instantiated in the passport frame by the
attributes TYPE of the passport, NATIONALITY, SEX and HEIGHT of the bearer and COLOUR
of the bearer’s eyes. The values of property attributes are of an abstract nature. They
may be located on a scale, such as in the case of HEIGHT, LENGTH, WIDTH, SIZE, WEIGHT,
AMOUNT, QUANTITY, TEMPERATURE, PRICE, WORTH, VALUE, DURATION, EXTENSION, etc. or not,
as with the attributes NATIONALITY, SEX, COLOUR, SHAPE, CHARACTER, NATURE, MEANING,
FORM, etc.
Event attributes. This is a somewhat imprecise class of attributes that link the
possessor to events and activities. In the passport example, we encountered the
implicit attribute BIRTH of the bearer; there is also the implicit attribute ISSUE of the
passport which is involved in the attributes DATE OF ISSUE and also in AUTHORITY, as this
is meant to denote the issuing authority; finally there is the implicit attribute EXPIRY.
The values of these attributes are events uniquely connected to their ‘possessors’:
there is only one birth per person and one issue and expiry per passport. The
attributes of the values of event attributes (e.g. of the birth of the bearer and the
issuing of the passport) are thematic roles of the events: LOCATION (in PLACE OF BIRTH),
TIME (in DATE OF BIRTH/ISSUE/EXPIRY) and AGENT (in [ISSUING] AUTHORITY). Later we will
see that there is another group of event-related attributes: those specifying manners
of interaction with the referent of the frame or parts of it. As we saw with basic
categories in 11.3, manners of interaction are an important aspect of categorization.
12.1.6 Constraints
Each attribute in a frame takes values from a certain set of alternatives. These
alternatives are constrained. First, an attribute returns values of a certain type, e.g.
a date, a name of a person, a measurement in centimetres, etc. Then, among the
logically possible values, it may be the case that only some are eligible. In the case of
the passport frame attributes, the attributes underlie constraints such as these:
∑
HEIGHT: possible heights of
∑
DATE OF BIRTH: dates in the past that result in a possible age of
∑
GIVEN NAMES: names admissible as given names for persons
∑
NAME: a registered family name
∑
COLOUR OF EYES: a colour possible for human eyes
∑
FACE: a photographic picture that looks like a human face
persons
the bearer
In some cases, the choice is constrained to a fixed and closed set of alternative values,
e.g. for the attributes SEX and NATIONALITY of the bearer and TYPE of the passport.
Further constraints on the values of arguments arise from the fact that the values
of attributes may depend on, or correlate with, the values of others. For example,
sex and nationality of the bearer constrain the possible values of the attribute GIVEN
NAMES, at least in some countries (a male German bearer cannot have ‘Angelika’ as
an entry for his given name). Sex and age will constrain to a certain extent what the
Frames
311
face of a person may look like; the name and given names correspond to a certain,
if often modest, extent with the shape of the signature, and so on. These constraints
apply to the passport frame in general; their violation would result in an inconsistent
or at least implausible description. In the rest of this chapter we will leave constraints
more or less aside. This part of frame theory is not as well developed as the theory
of attributes and values.
We will now turn to various applications of Barsalou frames in semantics, first
looking at verbs (12.2) and then at nouns (12.3).
12.2 VERBS AND FRAMES
12.2.1 Verbal argument frames
The argument structure of a verb is a simple straightforward frame; the referent
is the referential argument of the verb; the attributes are the thematic roles of
the arguments. In fact, such argument frames were the first examples of frames
discussed in linguistics in Fillmore (1968) and are an important precursor to the
more general present notion of the frame. Fillmore observed that each argument of a
verb is of a different general role (he called them ‘cases’) and that each of the general
roles can occur only once in a sentence. Given that any argument is only specified
once in a sentence, argument frames satisfy the three uniqueness conditions in
Definition 1: there is a central element (UR), each role can occur only once (UA) and
the roles correspond to exactly one argument (UV). Figure 12.8 shows the argument
frame for a ditransitive verb like give or send. In the figure, the argument nodes are
indicated by rectangles rather than circles. Rectangles are used for open argument
elements in a frame (except for referential arguments).
Figure 12.8
Frame graph of the argument structure of ditransitive verbs
AGENT
THEME
RECIPIENT
From the perspective of frame analysis it is not only natural, but in fact necessary
to assume that verbs have a referential event argument (recall the discussion in
5.3.2): it is the possessor of the argument attributes.
The condition UA is responsible for a constraint on the argument structures that
is otherwise unexplained: a verb cannot have two equal semantic roles – even if the
event expressed involves participants that play equal roles in this kind of situation.
312
Understanding semantics
For most two-place or three-place verbs, the roles of the arguments are clearly
different; for example with ‘hit’; the AGENT and the PATIENT are clearly involved in
the event in very different ways. However, there are reciprocal (8.3.6) verbs such as
struggle, agree or correspond where the interacting participants factually play the
same role. Yet they have to be assigned different semantic roles and linked differently,
e.g. as subject and with-PP; alternatively they can be both packed into a conjunction
and linked as subject (recall Levin’s marry class of verbs discussed in 6.1.3). In this
conceptual variant, the predication applies to only one, though complex, argument.
It must be cautioned that an argument frame is still a long way off representing
the full meaning of the verb. Hundreds if not thousands of verbs may have the same
argument frame. The argument frame contains little information about the event
itself, the kind of process or action it constitutes. In addition, these ‘flat’ argument
frames, where all arguments are directly assigned as attributes to the referent, fail
to reflect possible conceptual dependencies among the arguments. For example, a
verb like go in the construction ‘go from A to B’ would have a similar structure: three
argument attributes, in this case labelled AGENT, SOURCE and GOAL. While these three
attributes are represented on a par, they are in fact conceptually dependent on each
other: the SOURCE is the LOCATION OF AGENT at the beginning of the event and the GOAL
is the LOCATION OF AGENT at the end. How this can be modelled in frames is not yet
settled. More generally, the problem of modelling the situation structure of an event
(recall 6.2), e.g. initial and resultant condition, is not yet solved. Nevertheless, verb
argument frames are very useful, as we will see.
12.2.2 Deverbal nouns
The meaning of deverbal nouns, i.e. nouns derived from verbs, can to a great extent
be understood as the result of simple operations on the argument frame of the
underlying verb. Figure 12.9 shows three nouns derived from the verb walk. The verb
concept has an AGENT attribute and a PATH attribute; the latter is syntactically not
obligatory. It can be partly or completely specified by source, goal and/or path
specifications like from the station, to the post office, through the park, respectively. The
first variant, as in I had a nice walk this morning, refers to the event itself. The two
original verb arguments are two potential relational arguments of the noun walk1.
Therefore they are depicted as rectangles. The argument frame of the noun is
Figure 12.9
Frame graph for three nouns derived from the verb walk
(act of walking)
(route for walking)
walk1
walk2
walk
AGENT
PATH
walker
walk
walk
AGENT
PURPOSE
PATH
PATH
AGENT
AGENT
PATH
ACTIVITY
Frames
313
identical with the argument frame of the verb walk (in one of its alternates). The
principal difference between a verb concept and a noun concept – whatever it is – is
not captured by the mere case frames.
The second variant walk2 denotes a route for walking. Its referent is the path that
is walked along. Thus the referential node is shifted from the event to the path. The
frame contains not only a link from the ‘walk’ node to the referential node, but also
a reverse link PURPOSE. This is owed to the meaning of walk2: it is a path FOR walking.
The attributes AGENT and PATH of the event node are not deleted. They belong to the
concept of walk2: a walk is the path of walking and there is no walking without
walkers. However, neither the event node nor the agent node is a relational argument
of the noun walk2, since the reference of ‘path’ can be identified without determining
somebody engaged in an act of walking on the path.
The concept ›walker‹ is formed from the verb concept by shifting the referential
node to the AGENT of walking. Again, there is an inverse link from the ‘walker’ node
back to the event node. The corresponding attribute can be added and labelled
‘ACTIVITY’: a ‘walker’ is someone engaged in an activity, their activity (a functional
3
concept) is walking.
In this way, frames can be used to analyse the meanings of certain types of word
formation. We will see in the next section that frame theory can not only model the
meanings of certain types of deverbal noun, but also helps us to understand the
semantics of nominal compounds.
12.3 NOUNS AND FRAMES
Nouns have few if any relational arguments. Thus, the argument frames of nouns do
not offer much access to noun concepts. It may well be the case that the concepts for
complex things like humans and animals or cars and universities are, in themselves,
very complex. In general, it is very hard to get at the internal components of noun
meanings.
12.3.1 Metonymy
One way of getting at the components of noun meanings is to study metonymy.
In 3.4.3.1, metonymy was introduced as shifting the reference of an expression to
something that BELONGS to the original kind of referent. For example, metonymy
applied to the concept ›university‹ may shift reference to the university campus,
the university administration, the university staff, the university’s student body,
the university’s course system, and so on. From the point of view of frame theory,
these things belonging to the original referent are just the values of corresponding
attributes: the campus is the value of the attribute CAMPUS of the university, and so on.
Therefore we can infer from the possible metonymical shifts that the original concept
3 The walker might be one who is engaged in walking in a particular situation, or habitually. This
kind of difference cannot be modelled in the simple frames used here.
314
Understanding semantics
contains certain attributes of the referent. In the case of the concept ›university‹ these
include the attributes CAMPUS, ADMINISTRATION, STAFF, STUDENTS BODY and COURSE SYSTEM.
Using frames, we can characterize metonymy much more precisely: reference is
shifted to the value of one of the original referent’s attributes. This is, however, only
a necessary, not a sufficient condition. Metonymy is not possible for every attribute
of universities. For example, every university has a year of foundation, say 1869.
But one cannot use the term university for referring metonymically to the year of
its foundation, like in Boston University was a year of considerable unrest. In order to
understand the restriction, let us first have a look at the university example. Figure
12.10 illustrates the effect of applying the metonymical shift from the institution to
the campus to the concept ›university‹.
Figure 12.10
Metonymical shift ›university‹ to ›campus‹
INSTITUTION
CAMPUS
CAMPUS
LOCATION
LOCATION
The original frame on the left contains an attribute CAMPUS of the referent node,
and many other nodes, only vaguely indicated. The value of the CAMPUS attribute,
i.e. the campus of the university, has its own attributes, among them an attribute
LOCATION which is responsible for the possibility to form sentences like the university
is on a hill. With the metonymical shift, the value of CAMPUS becomes the referent
node. The new referent can be linked back to the original referent by an attribute
which takes the university itself as its value. Such a functional concept does exist,
although we do not have a specific functional noun for it; in the figure, I have
simply dubbed the attribute INSTITUTION. It is a proper functional concept because,
according to our notion of a campus, a campus will host exactly one university (or
similar institution), so there is a one-to-one correspondence between campuses and
institutions with a campus.
The example shows that there is another necessary condition for metonymic
shifts: the referent node can only be shifted to an element of the frame that is linked
back to the referent node by an appropriate attribute. There must be a one-to-one
correspondence between the original referents and referents resulting from the
metonymic shift. This is in accordance with a general condition for metonymy.
Frames
315
In metonymic use, the literal referent ‘stands for’ the metonymic referent: the
university stands for its campus – and vice versa: if the campus is a location
called, say, Ormond Hill, the name of the location can be metonymically used for
the university with its campus there. This is impossible if there is more than one
campus for a university, or if a certain area hosts more than one university. In fact,
if these conditions are not given, metonymy fails. For example, Tokyo University has
two separate campuses, one at Komaba and one at Hongo. Thus, one cannot say, for
example, (5a), because this would apply only to the Komaba campus. Conversely, (5a)
would not refer to Tokyo University as a whole.
(5)
a.
Tokyo University lies close to Shibuya.
b.
Hongo has entrance examinations tomorrow.
If you take a second look at the types of metonymy that were mentioned in 3.4.3 and
listed in (10) in 3.5.2, you will realize that this condition is fulfilled in all cases. For
example, the pars-pro-toto type metonymies (cf. redneck, asshole) shift reference to
a part of the original referent; parts belong to a unique whole. The analogue holds
for types like ‘carrier for content’ (write a paper), ‘container for content’ (drink a
bottle), ‘residence for government’ (Beijing reacted immediately), ‘clothing for wearer’
(blue helmets), etc. In the case of fluid or celebrity, an instance of something having
a property stands for the carrier: this instance has a unique carrier, and the carrier
displays this unique instance.
Cognitive linguists like Lakoff consider metonymy (along with metaphor) a
fundamental conceptual mechanism of cognition (cf. Lakoff 1987). Indeed, the
mechanism of shifting the referent node in a frame and linking it back to the original
referent node is not restricted to metonymical shifts. The examples of deverbal
nouns analysed in 12.2.2 are based on the same conceptual mechanism.
12.3.2 Affordances
We saw in 11.3 on levels of categorization that one crucial aspect of categorization is
the manner in which we interact with members of a category. This holds in particular
for categories of artefacts. Artefacts exist because people invented and produced
them, and – except in the case of art – they did so in order to use them for certain
practical purposes. The purpose for which an artefact is designed constrains its
other characteristics such as its shape and size and the material it is made of. But, at
the same time, a given purpose can be achieved in very different ways; just think of
the practically unlimited ways in which chairs may be designed. What unites these
exemplars of the general category of chairs is that they are made for sitting on them
and therefore more or less sit-on-able. (Note that not everything sit-on-able is a chair,
e.g. a sofa, a bike or a camel.) We may therefore assume that the types of activities
for which an artefact is designed are included in the conceptual representation
4
of the corresponding category: the frames for concepts like ›cup‹, ›bowl‹, ›vase‹,
4 Recall the discussion of the Labov experiments in 11.4.2.
316
Understanding semantics
›hammer‹, ›car‹, ›saxophone‹ or ›T-shirt‹ have an attribute that specifies certain
ways of interacting with this kind of object. Taking up a term from philosophy
and generalizing it for our purposes, we will call these interactional properties of
objects ‘affordances’: chairs have the affordance that they are for sitting on; a cup
has the affordance of being for drinking from; a saxophone is for playing music on;
a car for sitting in and driving from A to B. In the frame graphs, I will use the label
‘FOR’ for the affordance attribute. The affordance attribute links the referent to an
activity; the activity carries its argument frame. One of the arguments is identified
with the referent of the main frame; the corresponding role link is the inverse of the
affordance link. Thus, an affordance attribute not only provides a link to an activity
but also specifies the role that the frame referent plays in the activity. Figure 12.11
shows the frame for ›cup‹ with the drinking affordance. All other attributes of the
cup, indicated by the little arrows pointing off the referent nodes, are not elaborated.
The cup fills the container role in the drinking activity.
Figure 12.11
Frame for ›cup‹ with drinking affordance
drink
FOR
AGENT
THEME
CONT.
cup
The THEME and AGENT attributes provide conceptual links from the frame referent to a
beverage and a drinking person. These links can be employed for the interpretation of
possessive constructions such as my cup or cup of tea, which can be taken as ›cup speaker
drinks from‹ and ›cup containing tea to drink‹. (Recall the discussion of the interpretation
of possessive constructions with sortal nouns in 4.1.4: one needs some relation between
the possessor and the referent for interpreting possessive constructions.)
There are concepts and categories at the superordinate level of categorization
which are essentially reduced to a certain affordance specification: for example,
›food‹, ›garment‹, ›footwear‹, ›vehicle‹, ›vessel‹, ›aircraft‹, ›musical instrument‹, etc.
12.3.3 Nominal compounds
Affordance attributes are a key to certain types of nominal compounds. Consider
the case of coffee cup. In one of its meaning variants, coffee denotes a beverage. As
such it has the affordance of playing the THEME role in a drinking activity. Since both
Frames
317
parts of the compound coffee cup have a drinking-affordance, the two frames can be
connected by unifying the drinking nodes in the two frames. Figure 12.12 shows the
two frames of ›coffee‹ and ›cup‹ with their drinking affordances and indicates the
unification of the two ‘drink’ nodes.
Figure 12.12
Unification of frames for ›coffee‹ and ›cup‹
drink
drink
FOR
FOR
CONT.
THEME
coffee
cup
The unification has the effect of integrating the ‘coffee’ node and the ‘cup’ node
into the same ›drink‹ frame (see Fig. 12.13). The referent node of the ›cup‹ frame
becomes the referent node of the total frame, since cup is the head of the compound.
The ‘coffee’ node loses its referent status (otherwise the frame would violate UR).
This is in accordance with the fact that the compound coffee cup, unlike the syntactic
combination cup of coffee, does not refer to any coffee – a coffee cup is a coffee
cup even when it is empty, or filled with water or another substance. According to
condition UR, every node in a frame is to be connected to the referent node by a
chain of attributes. The ›coffee cup‹ frame complies with UR due to the FOR link from
the ‘cup’ node to the ‘drink’ node, and the THEME link from there to the ‘coffee’ node.
Figure 12.13
Frame for ›coffee cup‹
drink
coffee
FOR
FOR
THEME
CONT.
cup
318
Understanding semantics
The example is representative of a very frequent conceptual pattern of
compounding. It is not necessary that the linking activity frame is an affordance
for both parts of the compound. For book shop we need not assume that one of the
affordances of a book is to be sold in a shop. The ›shop‹ frame contains a selling
(or buying) affordance in which the shop takes the role of the location of the event.
For the regular interpretation of book shop we will unify the THEME argument of
the ›selling‹ frame with the referent node of the ›book‹ frame. This is admissible in
terms of the selectional restrictions of the THEME argument; but they require less than
the affordance of being for sale. I would like to call this type of compound a frame
compound. There are three more types of compound that can be straightforwardly
described within the frame approach.
Value compounds. This type is represented by plastic bag, company flat or park
bench. The modifier directly specifies the value of an attribute: the MATERIAL of the
bag is plastic; the OWNER of the flat is the company; the LOCATION of the bench is a park.
Thus the referent node of the modifier concept is unified with the value node of one
of the attributes in the head frame (cf. Fig. 12.14), thereby losing its referent status.
Figure 12.14
Unification of frames for ›plastic bag‹
plastic
MATERIAL
bag
Argument compounds. If the head noun is relational or functional, its frame
provides for a possessor argument or other relational arguments. There are
compounds which are interpreted as the modifier specifying an argument of the
head frame, e.g. air pressure ›pressure of (the) air‹, oil price ›price of oil‹ or chicken
leg ›leg of a chicken‹. Figure 12.15 shows the way in which the frame for air pressure
comes about. Again, the modifier node is stripped of its referent status.
Figure 12.15
Unification of frames for ›air pressure‹
air
PRESSURE
pressure
Synthetic compounds. So-called synthetic compounds are cases such as piano
player or bus driver. The head is a deverbal noun which comes with the verb frame
Frames
319
integrated into its frame (cf. the frame for walker in Fig. 12.9). The modifier specifies
another argument of the event; the unification mechanism is the same as with frame
compounds (cf. Fig. 12.12).
Reflection on the analysis of regular compounds. The examples illustrate that
the frame approach to word meaning is able to account for central aspects of the
semantics of regular compounds.
∑ The basic mechanism is unification. By means of unification the conceptual
content of the modifier is added to the frame of the head and connected to the
referent.
∑ The referent of the head becomes the referent of the compound.
∑ The meaning of the head is contained in the meaning of the compound.
∑ Therefore, the compound is a hyponym of the head (cf. 8.2.1).
∑ Unification leads to a specific relation between the meaning of the modifier and
the meaning of the head (cf. 8.2.2).
12.4 FRAMES AND COMPOSITION
12.4.1 Composing a predication
Frame theory also allows us to model elementary steps of composition. The basic
mechanism is unification. Predicate expressions carry open argument nodes as the
values of the role attributes. These open nodes are unified with the referential nodes
of the frames for the arguments. This will be illustrated with the first example from
chapter 5:
(6)
Johnny sent money to a dubious company.
The frame for the verb send is an argument frame with the three attributes AGENT,
and RECIPIENT, like the frame in Fig. 12.8. For ›company‹ we will just use
an unanalysed frame, i.e. a referent node with unspecified attributes. The frame
for the proper name Johnny will be just the referent node with one attribute NAME
that assigns the value Johnny. The adjective dubious requires a frame without a
referential node (because adjectives do not have a referential argument). We will
represent it as an open argument node with one attribute, in this case RELIABILITY,
assigning the value ‘low’. Figure 12.16 shows the ingredient frames for the sentence
and the unifications that bring everything together into one frame for the complex
sentence predication.
THEME
320
Understanding semantics
Figure 12.16
Unification of frames for ›Johnny sent money to a dubious company‹
send
AGENT
THEME
RECIPIENT
money
Johnny
company
NAME
RELIABILITY
low
Johnny
The grammatical structure of the sentence indicates which nodes are to be unified
in order to link all five frames to each other. The frame for the adjective is linked
by unification of its argument node with the referent node of the ›company‹ frame
because the adjective dubious is syntactically combined with the noun company.
According to the linking rules of English, the subject of the sentence specifies the
agent of the sending; the direct object specifies its theme and the to-PP the recipient.
This leads to the unification of the referent nodes of the frames for ›Johnny‹, ›money‹
and ›company‹ with the respective argument nodes of the ›send‹ frame. Figure 12.17
shows the result.
Figure 12.17
Frame for ›Johnny sent money to a dubious company‹
send
AGENT
THEME
RECIPIENT
NAME
Johnny
Johnny
money
company
RELIABILITY
low
The frame for the whole sentence violates the UR condition on frames. UR only
holds for lexical frames. In the frame of a sentence, the referents of the referring
Frames
321
elements are all marked. Composition connects all conceptual information into
one coherent whole. Thus, every piece of information is in the context of the rest
of the sentence. It is this frame in which the meaning variants of the lexemes and
grammatical forms occurring in the sentence are selected, and eventually shifted, to
make everything fit (recall the discussion of the Principle of Consistent Interpretation
in 3.4). In the case of our example, the combination of dubious with company rules
out the meaning variant ›hesitating, doubting‹ because the company is very unlikely
to be in that state. Company is taken in the ›commercial business‹ sense, required by
its role as the recipient of sending; the meaning ›being with someone‹ (as in I enjoyed
5
his company) is ruled out in the setting of sending.
12.4.2 Open problems
The frames in Fig. 12.16 and Fig. 12.17 represent only the predicate-argument
structure of the proposition, and this is even without the tense argument (recall
6.3.5, 6.4). Tense and aspect are not represented; this is connected to the fact that the
situation structure is not represented in the verb frames. Similarly, the frames we
used do not capture the determination of noun phrases: in the example, we would
get the same frame if money or a dubious company were replaced by the money or the
dubious company. The frames capture only the core of the verbal and the nominal
onions (6.6.1, 4.5.5). All of the functional elements of the sentence are missing. This
does not mean that representation of these elements is principally impossible in
the frame approach; it just represents the present stage of applying frame theory in
semantics.
12.5 FRAMES AND COGNITION
12.5.1 Frames vs features
As a theory of concept and category representation, the frame approach is superior to
the feature approach in three important respects. First, the frame approach employs
explicit attributes while these are left implicit in the feature approach. For example,
the three features that make up the description of ›girl‹ in Table 9.3, [+HUMAN], [–
ADULT], [–MALE], implicitly specify the values of three attributes:
(7)
[+HUMAN]
[–ADULT]
[–male]
____.SPECIES: human
____.AGE: young
____.SEX: female
Thus, the features used in BFA are in fact fixed combinations of an attribute and one
out of two possible values. It goes without saying that semantic analysis should be
explicit about the notions actually employed. Barsalou cites empirical evidence that
5 The meaning descriptions relate to the Oxford English Dictionary, http://oxforddictionaries.com/,
accessed 18 September 2012.
322
Understanding semantics
shows that humans actually use attributes for categorization rather than just feature
lists (cf. Barsalou 1992b: 25–29).
Second, BFA restricts the description to binary features. This is psychologically
unmotivated. In fact there are very few attributes that admit only two values;
SEX would be one. Given that visual perception plays a very important role in
categorization, there is no reason whatsoever to exclude attributes such as COLOUR or
SHAPE, which can take an unlimited number of values. Due to its restriction to binary
features, BFA is unable to deal with these aspects of objects. There is no feasible
way of describing the shape of a pig or the colour of its skin by a combination of
binary features. At best, values of non-binary attributes are just turned into so many
different features, like [HUMAN], [ROUND], [GREEN], etc.
Third, BFA descriptions lack recursion; no information can be included about
the values of the attributes of the referent. Recall the problem we encountered with
a feature description of bikini in 9.3.2. If one wanted to include into the meaning
description of bikini that the wearer is female, there is no way to do so by using the
feature [–MALE] (one could only use an ad hoc feature [–WORN BY MALE PERSONS]).
With the frame approach, this property of bikinis could be easily represented as:
bikini.FOR:wear.WEARER:[wearer].SEX:female. The lack of recursion is an immediate
consequence of the restriction to binary features: there is no information which one
could reasonably add about the + or – value of a binary feature.
12.5.2 Frames vs prototypes
Barsalou emphasizes that prototypes can be integrated into frame theory (Barsalou
1992a: 161ff.). Prototypical properties can be integrated into frames as default values
for certain attributes; for example, there might be a default entry for the SIZE value in
the ›bird‹ frame. Also, one might add, certain prototype effects can be captured by
constraints. Again relating to birds, the fact that most, but not all kinds of bird are
able to fly can be captured by a constraint that links the ability to fly to the relative
size of the wings of the bird and to its weight. Using a frame model for concept
representation would also offer the possibility of coping with another question left
unanswered by prototype theory: PT bases category membership on the degree
of similarity to the prototype, but it does not say which aspects of the objects
considered matter for the similarity judgement: shape? size? colour? affordance?
(recall 11.2.9). Frame theory would be able to define the criterion of similarity in
terms of the attributes involved.
12.6 CONCLUSION
The application of Barsalou’s frame concept to semantics is a recent development. It
offers possibilities for analysing word meanings in terms of attributes and the values
they take. Superior to both the binary feature approach and prototype theory, it is the
first general model of concepts that promises applicability to the decomposition of all
parts of speech as well as to the questions of semantic composition. It is also the first
Frames
323
theory of representation which is, at the same time, formally more precise and explicit
than its competitors and based on experimental cognitive research. As a framework
for doing conceptual semantics, the theory is still in its early stages, but it is very
promising, both as a perspective for the classical problems of semantics – lexical
semantics (including the semantics of word formation), grammatical meaning and
compositional semantics – and as a perspective for cognitive linguistics in general.
As far as compositional semantics is concerned, there is another framework
incomparably further developed than what we have sketched above: formal
semantics, to which an in-a-nutshell introduction will be given in the next chapter.
Formal semantics, however, is an approach which does not aim at the cognitive level
of meaning; in fact, the cognitive level is not included at all in that framework.
EXERCISES
1. Characterize the crucial properties of attributes in Barsalou frames.
2. Identify attributes of all the four types distinguished in 12.1.5 that might figure
in the frame for a shoe.
3. Translate the frames in Figs 12.9 (all three of them), 12.10 (both), and 12.13 into
frame matrices. Proceed as follows:
a. Disregard the attributes rendered in grey.
b. Disregard the attributes indicated by unlabelled small arrows without value
nodes.
c. Insert variables x, y, z for the values of argument nodes (rectangles)
d. Use entries of the form ‘(attribute)’ for the specification of unspecified
attribute values, for example: … .AGENT: (agent).
4. Determine the readings of the following deverbal nouns in terms of the verb
argument they refer to, and draw the respective frame graphs like those in
Fig. 12.9:
a. drink
b. building (two readings)
c. baker
5. Characterize metonymy in terms of frames. What are the crucial conditions for
this operation?
6. Try to construct a frame for the metonymical expression blue helmet.
7. Describe the notion of ‘affordance’.
8. Categorize the following compounds according to the four types mentioned
in 12.3.3.
a. bus ticket
b. bus driver c. school bus
d. bus stop e. Greyhound bus
Describe their meanings in terms of a frame analysis. Feel free to choose
appropriate names for the attributes involved.
324
Understanding semantics
9. Explain why frames are superior to binary features as a format for representing
meanings.
FURTHER READING
Ungerer and Schmid (2006) contains a chapter about ‘frames’ and ‘scripts’ and their
applications in cognitive semantics; the notion of a frame used there is similar, but
not the same as the one used here. Kövecses (2006, chs 5–9) applies a very broad
notion of frames to phenomena including metonymy and metaphor. Barsalou
(1992b) gives an introduction to the notion of frames (as used here) and their
applications; Barsalou (1992a, ch. 7) discusses frames in the general context of an
introduction to cognitive psychology. On frames vs feature lists and the relation of
frames and prototype theory see Barsalou (1992a, 7.2).
There is a huge online resource of frames related to English words: the FrameNet
site at www.icsi.berkeley.edu/framenet, initiated and supervised by Charles Fillmore.
It contains a network of thousands of frames, in terms of core and non-core attributes
and their relations to other frames. The network is permanently being extended and
refined; it is a very valuable resource of lexical semantics – browse it!
13
Formal semantics
This chapter deals with composition. More precisely, it offers an elementary
introduction into formal semantics (alternatively: model-theoretic, truth-conditional,
referential, logical or possible-world semantics). In formal semantics, meaning is a
matter of reference and truth conditions. It is the most technical and difficult variety
of semantics, very mathematical, but it is the main framework in which sentence
semantics has been developed.
13.1 JAPANESE NUMERALS: A SIMPLE EXAMPLE OF A COMPOSITIONAL
ANALYSIS
13.1.1 The system of numerals
Japanese has two sets of numerals, a native set of expressions for the numbers 1 to
10, 20 and a couple of isolated greater numbers such as 100 and 1000. The numeral
system proper is of Chinese origin (as is the case in many languages neighbouring
China). It is chosen here because it is perfectly regular, unlike the system of number
names in English or those of any other European language. The basis of the system
for the first ninety-nine numerals consists of just ten lexical units, the expressions
for the first ten numbers:
(1)
ichi
1
ni
2
san
3
yon
4
go
5
roku
6
nana
7
hachi kyū
8
9
jū
10
When we attach the word for 10 jū to the words for the digits 2 to 9 we obtain the
terms for the tens higher than 10:
(2)
jū
10 nijū
20 sanjū 30 yonjū 40
…
kyūjū 90
Thus ‘thirty’ is expressed as ‘three ten’. Terms for the numbers 11, …, 19, 21, … 29
and so on are formed by attaching the words for the digits 1, …, 9 to the terms for
the tens, including jū. The number 11 is expressed as ‘ten one’ and 21 as ‘two ten one’.
The complete system for 1 to 99 is given in (3). The overall system continues in the
same regular and economic fashion, but we will confine the analysis to the numerals
for 1 to 99:
326
Understanding semantics
jū
(3)
10
ni
jū
20
…
kyū jū
90
ichi
1 jū ichi
11
ni
jū
ichi
21
…
kyū jū ichi
91
ni
2 jū ni
12
ni
jū
ni
22
…
kyū jū ni
92
san
3 jū san
13
ni
jū
san
23
…
kyū jū san
93
…
…
kyū
9 jū kyū
…
19
ni
jū
kyū
…
29
…
kyū jū kyū
99
13.1.2 Formal description of the system
The system in (3) is extremely economical: with as few as ten basic elements, the
numerals for 1 to 10, a total of ninety-nine numerals is given or can be formed.
The meanings of the complex numerals derive unambiguously from the meanings
of their components and their order. For example, the meaning of nijūsan can be
calculated from the components ni 2, jū 10 and san 3 as 2 × 10 + 3, i.e. 23. Therefore,
the complex numerals constitute a good example of expressions with compositional
meaning. Recall that, according to the Principle of Compositionality, the meaning of
a complex expression is determined by the lexical meanings of its components, their
grammatical meaning (irrelevant here) and the way in which they are combined.
There are two types of composite expressions: the terms for the tens in (2) and the
terms for the numbers between the tens in (3). Both types are formed by the same
morphological operation, called concatenation. Concatenation (from Latin ‘chain
together’) forms one expression out of two by attaching the second to the end of the
first. While both types of composites are formed by concatenation, the interpretation
of the combinations differs. The word sanjū ‘three ten’ for 30 means ›three TIMES ten‹,
but jūsan ‘ten three’ for 13 means ›ten PLUS three‹. If the first part denotes a smaller
number than the second, concatenation is interpreted as multiplication. Otherwise
concatenation stands for addition. Now, if a numeral is composed of three parts,
it must have been built up in two steps, since concatenation only allows for the
combining of two expressions. Therefore, in a first step two basic elements have to be
combined to form a two-part numeral and this in turn combined with a third part.
These two steps can be taken in either order. For example, sanjūroku could, for its
mere form, be derived in two ways:
(4)
a.
(3 10) 6:
First concatenate san 3 with jū 10 to form sanjū, then
concatenate sanjū with roku 6.
b.
3 (10 6)
First concatenate jū 10 with roku 6 to form jūroku, then
concatenate san 3 with jūroku.
When we follow (4a), we obtain the interpretation 30 for sanjū (smaller number 3
times greater number 10) and 36 for sanjūroku (greater number 30 plus smaller
number 6). But with (4b) we obtain the interpretation 16 for jūroku and would then
Formal semantics
327
have to multiply this with 3, obtaining 48. Only the first derivation is correct: first
form the ten word and then attach the digit word.
We are now in a position to give a compositional description of the system. A
1
compositional description always consists of four components.
FB
Formation Base
A list of the basic expressions.
IB
Interpretation Base
A specification of the interpretations of the basic expressions.
FR
Formation Rules
A specification of the formation rules for complex expressions.
IR
Interpretation Rules
A specification of the corresponding interpretation rules for complex
expressions.
2
FB was defined in (1).
FB
Basic expressions are ichi, ni, san, yon, go, roku, nana, hachi, kyū, jū
FR consists of two rules:
F1
form ‘UV’
where
and
U = ni, san, yon, go, roku, nana, hachi or kyū
V = jū
F2
form ‘ZX’
where
and
Z = jū, nijū, sanjū, …, kyūjū
X = ichi, ni, san, yon, go, roku, nana, hachi or kyū
We will use square brackets for denoting interpretations: [A] is ‘the interpretation of
A’. IB assigns interpretations to the ten basic expressions:
IB
[ichi] = 1, [ni] = 2, [san] = 3, … , [kyū] = 9, [jū] = 10
1 Compared to the general scheme of semantic composition given in Fig. 1.1, the intermediate level
of grammatical meaning is missing here. It will be neglected throughout this chapter.
2 We use the more general term interpretation instead of meaning. Interpretations are whatever the
expressions are assigned in such formal systems. How far these interpretations can be considered
meanings in the proper sense will be discussed later.
328
Understanding semantics
IR consists of two interpretation rules, I1 and I2, that correspond to F1 and F2,
respectively. They determine the way in which the interpretation of a complex
expression derives if it is built according to these rules.
I1
If UV is formed according to rule F1,
then [UV] = [U] × [V]
I2
If ZX is formed according to rule F2,
then [ZX] = [Z] + [X]
We can now ‘compose’ the meanings of the complex numerals. In (5), this is done for
nanajūroku 76. On the basis of the given rules, the expression can only be derived
by first applying F1 and then F2: nana and jū combine to nanajū 70, then nanajū is
concatenated with roku 6. If we first derived jūroku 16 by F2, we could not concatenate
this with nana, as neither formation rule allows for jūroku as the second part.
(5)
f.
formation of nanajūroku
Step 1: F1: nana, jū → nanajū
Step 2: F2: nanajū, roku → nanajūroku
i.
interpretation of nanajūroku
Step 1: I1: [nanajū]
=
=
=
Step 2: I2: [nanajūroku] =
=
=
[nana] × [jū]
7 × 10
70
[nanajū] + [roku]
70 + 6
76
(IB)
(above, IB)
Figure 13.1 shows the parallel steps in deriving the form and the interpretation of
nanajūroku.
Figure 13.1
Formation and compositional interpretation of nanajūroku
冦
nana, jū
Interpretations
7 , 10
冦
Expressions
×
nanajū, roku
70 , 6
冦
I1
冦
F1
+
F2
nanajūroku
76
I2
Formal semantics
329
13.1.3 The general scheme of meaning composition
The formal treatment of the numeral system illustrates the general scheme of
compositional semantics as given in Fig. 13.2.
Figure 13.2
The general scheme of semantic composition
LEXICON
FB
basic expressions
F1,
IB
interpretations of
basic expressions
F2, …
complex expressions
GRAMMAR
I1,
I2, …
interpretations of
complex expressions
SEMANTICS
The essence of composition is the stepwise parallel formation of a complex expression
and its interpretation. The basis of the system consists of the components FB and IB.
These constitute the lexicon: a list of the lexical items with specified interpretations
(= lexical meanings). Unlike in our example, the lexical items are usually categorized
as, for instance, nouns, intransitive verbs, prepositions, etc. Such grammatical
categories are sets of expressions with the same combinatorial properties. They are
needed for the formulation of the formation rules, because these rules usually apply
only to certain categories of expression. For example, the rules F1 and F2 require four
different input categories for which we used the variables U, V, X and Z. U consists of
the unit words for 2 to 9 because these numerals share the combinatorial property
that they can form the first part of a complex ten word (for 20 to 90). The output of
a formation rule can be categorized, too: F1 generates expressions in the category of
tens, which in turn is an input category for F2. Both F1 and F2 generate expressions
that can be attached to the hundreds, not covered by our little system.
The components FB and FR form the grammar of the system, i.e. the part that
determines which expressions, basic or complex, are regular expressions and to
which categories the regular expressions belong.
The interpretation of the basic expressions, component IB, and the set IR of
interpretation rules I1, I2, … constitute the semantics of the system, i.e. the
apparatus that provides each regular expression with an interpretation. The basic
expressions receive interpretations directly by IB. The interpretations of complex
regular expressions are derived step by step from the interpretations of the basic
expressions they contain. Each time a formation rule applies to derive a new
expression, the corresponding interpretation rule applies to derive its interpretation.
The broken-line arrow in the bottom part of the scheme represents the resulting
assignment of interpretations to complex expressions.
330
Understanding semantics
13.2 A SMALL FRAGMENT OF ENGLISH
In formal semantics, the description of composition is usually broken down into
two steps. The first consists of a ‘translation’ of natural language sentences into an
appropriate formal language, usually some variant of predicate logic. This step itself
is compositional in the sense that the translations are systematically derived step by
step in the same sort of parallelism as semantic composition. In a second step, these
logical formulae receive a standard compositional interpretation. While the two-step
procedure may seem complicated at first sight, it is in fact of great advantage. The
translation into a logical language is the essential step of the analysis. Since logical
formulae are, to the trained eye, semantically completely transparent, the translation
of a sentence reveals its semantic structure as far as it can be expressed in the logical
language. The second step, a compositional interpretation of the logical language, is
standard and trivial, once the general rules are defined; actually it is usually omitted.
The logical formulae are therefore practically considered as meaning representations.
It is technically possible to skip the translation into logic and supply the natural
language sentences directly with the interpretations the translation formulae would
receive. In fact some variants of formal semantics proceed in this way.
Semantic theory is still far from being able to provide this kind of account for the
major part of any language. The analysis is therefore always restricted to a ‘fragment’,
a limited set of representative lexemes and a limited set of formation rules, for
which the corresponding interpretation rules are set up. Compositional analysis of
natural language is difficult because syntactic combinations tend to allow for several
semantic interpretations. In what follows we will treat a small fragment of English
containing just a handful of basic expressions. For the sentences of the fragment, we
will derive translations into predicate logic formulae similar to those used in 5.5.
13.2.1 The grammar of the fragment
The lexicon of the fragment is minimal, just designed to contain representatives
of the major word classes and to allow for minimal variation. In order to keep
the intermediate level of free grammatical forms out of the system, the forms are
simply fixed: all nouns are singular, all adjectives in their positive form and all verbs
in simple present tense, 3rd person singular, indicative, active. NPs are restricted
3
to proper names and singular indefinite count NPs. The basic expressions are
categorized in the traditional way.
FB Basic expressions are the members of the following categories:
NP
noun phrases
Mary, John
3 It must also be mentioned that we disregarded quantification (recall 4.5). Quantification is a
central issue of formal semantics; many introductions into formal semantics start from it. For
the purpose of this chapter, quantification can be left aside; its treatment is not necessary for
explaining the basic notions and methods of formal semantics.
Formal semantics
331
FB contd
N
nouns
person, fox
A
adjectives
young, busy, wicked
D
articles
a
VP
verb phrases
squints, smokes
TV
transitive verbs
knows, dislikes
CV
copula verb
is
4
Formation rules. In a system with categories, a formation rule specifies three things:
(i) the categories of the input expressions; (ii) the form of the output expression; and
(iii) its category. According to the first rule, a sentence may consist of an NP followed
by a VP. This is the only rule for forming a sentence. It supplies the VP with a subject
NP, i.e. a specification of its first argument:
F1
NP VP → S
The notation of the rule represents the general format used here. On the left side of the
arrow, the input categories are specified. The order in which they are arranged indicates
the form in which they must be combined. Rule F1 takes as input an NP and a VP and
forms a string consisting of the NP followed by the VP. On the right side of the arrow, the
output category is specified. F1 reads as follows: take an NP and let it be followed by a
VP; the result is of category S (sentence). Members of S are always complex expressions.
If we had only this formation rule, it would allow us to form four sentences:
(6)
NP
VP
result: S
Mary
squints
Mary squints
Mary
smokes
Mary smokes
John
squints
John squints
John
smokes
John smokes
The system contains two basic NPs, Mary and John. In addition to these, complex NPs
can be formed by combining nouns (N) with the indefinite article (D).
F2
D N → NP
4 All VPs of the fragment are to be taken in their habitual reading. Smokes means as much as is a
smoker, squints is to be taken as has eyes that look in different directions. The restriction enables us
to use the verbs in the simple present tense and assume present time reference.
332
Understanding semantics
F2, so far, allows the formation of the complex NPs a fox and a person. F2 enlarges
NP, an input category of F1. So we are able now to derive four more sentences: a fox
smokes, a person squints, etc.
With F3, we can combine adjectives with nouns. The output category is N, i.e. the
category of expressions that yield an NP when combined with the indefinite article.
AN→N
F3
When we apply this rule once, we can form six AN combinations, among them young
fox and wicked person. But since the output of rule F3 is again of category N, the
rule can be applied repeatedly. It allows for the formation of young wicked person
out of the derived N element wicked person and the adjective young and so on.
Consequently, an N can be preceded by any number of adjectives, for example young
wicked young young young wicked fox. This consequence of the rule, though it looks
strange, is harmless. There is nothing ungrammatical to such formations. F3 also
indirectly enlarges the number of NPs and sentences, since all N elements provide
input to rule F2 and the output of F2 is input to F1. With F1, F2, F3 we can form
sentences such as a young busy person smokes.
Let us now turn to the rules for forming complex VPs. The two transitive verbs
and the copula is can be combined with an NP to form a VP. There are two different
formation rules F4 and F5 because, as we will see, the interpretation of the two types
of VPs differs. The last rule allows the formation of a VP by combining the copula
with an adjective.
F4
TV
NP
→
VP
F5
CV
NP
→
VP
F6
CV
A
→
VP
Now the grammar of the system is complete. Using F4, F5 and F6, we can form
sentences like these:
(7)
a.
John is busy
F6: CV is, A busy → VP is busy
F1: NP John, VP is busy → S John is busy
b.
Mary is a busy person
F3: A busy, N person → N busy person
F2: D a, N busy person → NP a busy person
F5: CV is, NP a busy person → VP is a busy person
F1: NP Mary, VP is a busy person → S Mary is a busy person
Formal semantics
c.
333
a young fox dislikes John
F3: A young, N fox → N young fox
F2: D a, N young fox → NP a young fox
F4: TV dislikes, NP John → VP dislikes John
F1: NP a young fox, VP dislikes John → S a young fox dislikes John
13.2.2 The predicate logic language PL-F: its grammar
For the translation of the fragment, we define an appropriate predicate-logic
language, PL-F. All PL languages make use of the same repertoire of formation
rules, but they differ in the choice of the basic expressions, i.e. predicate constants,
5
individual constants and individual variables.
PL-F contains just the basic expressions we need: individual constants for Mary
and John, one-place predicate constants for the adjectives, nouns and intransitive
verbs and two-place predicate constants for the transitive verbs. In addition we
will need two variables for individuals. Both, individual constants and individual
variables, belong to the category T of [individual] ‘terms’. In addition, individual
variables form a category of their own, V (variables). When specifying FB for PL-F,
bold face is used for the constants that correspond to lexical items of the natural
language fragment. As PL-F does not obey the rules of English syntax, the ending -s
is omitted for the predicate constants that correspond to verbs.
FB
Basic expressions of PL-F are the members of the following categories:
T
individual constants, terms
m
j
V, T
individual variables, terms
x y
P1
one-place predicate constants
fox person
young busy wicked
squint smoke
P2
two-place predicate constants
know dislike
The formation rules for PL-F are only a subset of the usual PL system. F1 and F2
allow the formation of so-called prime formulae by attaching the required number of
argument terms to predicate constants (recall the description of PL formulae in 5.5):
F1
P1(T) → S
F2
P2(T, T) → S
5 Expressions in PL are either basic or complex; basic expressions are either variables or constants.
In first-order logic as used here there are individual constants and individual variables; basic
predicate expressions are all constants.
334
Understanding semantics
S, again, is the category of sentences (or formulae). The two rules allow the formation
of a limited number of formulae (60 in total), among them:
(8)
a.
person(m)
young(x)
b.
dislike(j,y)
know(x,m)
squint(j)
In addition to this type of prime formulae we can form identity statements by
connecting two terms with the ‘equals’ sign, yielding formulae such as ‘x = m’:
T=T→S
F3
Semantically, the equals sign is just another two-place predicate constant. It could be
introduced as such and treated syntactically as the others. We would then write ‘=
(x,m)’ instead of ‘x = m’. The usual way of writing identity statements is commonly
preferred in PL. But, as you see, the price of doing so is an extra syntactic rule.
Next we conjoin two formulae by logical conjunction. More connectives are used in
richer PL languages, such as ∨ for ‘and/or’ and → for ‘if … then …’. But we can do
without them here. We also do not include negation.
S∧ S→S
F4
F4 can be applied repeatedly. It allows the formation of complex formulae:
(9)
a.
young(j) ∧ wicked(j)
b.
fox(x) ∧ fox(y)
c.
fox(x) ∧ young(x) ∧ dislike(x,y) ∧ y = m
(9a) says that John is young and wicked, (9b) is a way of expressing that both x and
y are foxes, while the longish formula (9c) tells us the following: x is a fox and young
and dislikes y and y is Mary. In other words: x, a young fox, dislikes Mary.
The last formation rule introduces the existential quantifier ∃. This operator,
combined with a variable, is used as a prefix to a formula. The formula is enclosed in
parentheses in order to mark the part of the total expression to which the quantifier
applies, its so-called scope.
F5
∃V(S)→S
The existential quantifier is to be read and interpreted as ‘there is at least one V
such that S’. Usually, the ‘at least’ part is omitted in reading (but nevertheless to be
understood). The rule gives us formulae like these:
Formal semantics
335
(10) a. ∃x (young(x))
there is an x that is young
b. ∃x (young(x) ∧ person(x)) there is an x that is young and a person
= there is an x that is a young person
c. ∃y (know(x,y) ∧ x = m)
there is a y such that x knows y and x is Mary
= there is a y that Mary knows
Both F4 and F5 can be applied repeatedly and in combination with each other. The
formation potential of the system is considerable, but we will confine our examples
to such cases as will actually occur as translations of English fragment sentences, for
example:
(11) ∃x (x = m ∧ ∃y (dislike(x,y) ∧ fox(y) ) )
The formula reads as follows: ‘there is an x such that x is Mary and there is a y
such that x dislikes y and y is a fox’. (11) is somewhat complex, so we will give
its derivation in full detail. First the prime formulae are generated (steps 1 to 3).
Then they are packed according to the structure of the total formula. Following the
bracketing, we first conjoin the second and the third subformula (step 4) and then
apply ∃y in step 5. Next we conjoin ‘x = m’ to the ∃y-subformula (step 6). ∃x is
applied in the last step.
(12) step 1
F3
x=m
step 2
F2
dislike(x,y)
step 3
F1
fox(y)
step 4
F4
dislike(x,y) ∧ fox(y)
step 5
F5
∃y(dislike(x,y) ∧ fox(y))
step 6
F4
x = m ∧ ∃y(dislike(x,y) ∧ fox(y))
step 7
F5
∃x(x = m ∧ ∃y(dislike(x,y) ∧ fox(y)))
The formation rules of PL allow for two different uses of variables. First, by F5, they can
occur in the scope of an existential quantifier with that variable. Such occurrences are
called bound by the quantifier. Alternatively, a variable may occur in a formula without
being bound by a quantifier. It is then called free. To see the difference, consider the
stepwise derivation of the formula in (12). In step 1, x is free and it remains free until
it is bound by ∃x in step 7. The variable y is introduced in step 2 as a free variable. It is
eventually bound in step 5 by ∃y. Note that in step 5 ∃y only binds y.
It is important to note that free variables are interpreted on a par with individual
constants. They stand for a particular individual, functioning roughly like ‘the x’.
Existentially bound variables, however, achieve the reading of ‘some x’. Therefore,
steps 4 and 5 are interpreted as follows:
(13) a.
b.
dislike(x,y) ∧ fox(y)
the x dislikes the y and the y is a fox
∃y(dislike(x,y) ∧ fox(y)) there is some y that the x dislikes and y is a fox
i.e. there is some fox that the x dislikes
336
Understanding semantics
Note that the bound variable can be eliminated when the formula is expressed in
ordinary language (there is some fox …). The free one cannot, because it essentially
works like a pronoun.
13.2.3 Translating the fragment into predicate logic
13.2.3.1 Challenges for a compositional account
The system for the English fragment is quite small, yet it already poses a reasonable
set of compositional challenges. To begin with, the translation of the basic expressions
must be carefully arranged in order for the rest of the system to work out properly.
Then there are a number of problems concerning the system of composition rules.
Referential and predicative NPs. The system allows for both referential and
predicative NPs. If an NP is used according to F1 or F4, it is a verb complement with
a referent, i.e. referential. If F5 is applied, the NP is part of a copula VP, e.g. is a young
fox, is Mary, etc., it is then predicative and has no referent of its own (5.4.3). The rules
of the system have to do justice to both kinds of NPs.
Attributive and predicative adjectives. A parallel problem arises with adjectives.
How can they be analysed in a uniform way that covers their attributive use (F3) as
well as their predicative use (F6)?
Proper names and indefinite NPs. There are NPs of two different kinds, proper
names and indefinite NPs. In this chapter, we follow the usual approach taken in
formal semantics of treating proper names like individual terms and indefinite NPs
as predicators. For example, a fox dislikes John would be rendered as (14):
(14)
fox(x) ∧ dislike(x,j)
The proper name NP John directly provides the second argument of the verb
dislike. By contrast, the indefinite NP a fox contributes a further predication, fox(x),
about the first argument of the verb. The translation system has to take care of the
difference. At the same time, however, it should manage to treat both types of NPs in
a uniform way: for each of the three formation rules F1, F4 and F5, there should be
only one translation rule.
Argument sharing. A more general problem is a proper account of argument
sharing. The translation rules have to produce an appropriate account of which
predicates share which arguments. We have to aim at solutions such as fox(x) ∧
dislike(x,j) for a fox dislikes John. Here, the use of the same variable x as an argument
term of the subject noun fox and the verb dislike shows that they share one argument.
Thus, the subject specifies the first argument of the verb.
13.2.3.2 The translation system
The translation of the fragment into PL-F follows the scheme of semantic composition.
The notion of interpretation is to be replaced by translation. The general schema is
given in Fig. 13.3.
Formal semantics
337
Figure 13.3
The general scheme of the translation procedure
FB
basic expressions
F1,
translations of
basic expressions
TB
F2, …
T1,
T2, …
translations of
complex expressions
complex expressions
GRAMMAR
TRANSLATION
The translation base TB fixes the translations of the basic expressions. Those given
here will not be regular expressions of PL-F inasmuch as they contain empty slots
for argument terms. In the course of applying the translation rules, these empty slots
are properly filled with individual terms. In the end, each fragment sentence receives
a regular PL-F formula as its translation. The use of empty slots helps us to keep
track of the translation steps. There are two basic expressions of the fragment that
do not receive a translation: the indefinite article and the copula. These will be taken
care of in the translation rules concerning the categories D and CV. An underlined
expression stands for the translation of the expression. (John is the translation of
John, Mary is a young person is the translation of Mary is a young person.)
13.2.3.3 The translation base
Except for the copula and the indefinite article, we treat all basic expressions
including proper names as predicate expressions.
TB
NP
N
A
VP
TV
John:
_=j
Mary:
_=m
fox:
fox( _ )
person:
person( _ )
young:
young( _ )
busy:
busy( _ )
wicked:
wicked( _ )
squints:
squint( _ )
smokes:
smoke( _ )
dislikes:
dislike( _ , _ )
knows:
know( _ , _ )
338
Understanding semantics
You may wonder why the proper names John and Mary are not translated into oneplace predicate constants, rendering the translations John( _ ) and Mary( _ ). The
reason is that proper names correspond to particular individuals. Translating Mary
as ‘_ = m’ takes account of both aspects of proper names. On the one hand, the
proper name is treated as a predicate constant, since the empty slot in ‘_ = m’ has to
be filled with an argument term. On the other, the resulting predication only holds
for one single individual, namely Mary. Other one-place predicate constants such as
the translations of ordinary nouns, verbs and adjectives, may be true of any number
of individuals, zero, one or more. For a predicate constant such as person there is no
particular individual p such that ‘x is a person’ would be tantamount to ‘x = p’.
13.2.3.4 The translation rules
The simplest rules are those handling the indefinite article and the copula. As a
matter of fact, both expressions are treated as semantically zero:
T2
aN
=N
T5
is NP = NP
T6
is A
=A
A noun, e.g. fox, constitutes a one-place predicate constant that is true of all foxes.
An indefinite NP such as a fox provides the same predication. If an indefinite NP is
combined with the copula, the result is a one-place verb phrase, another predicate
6
constant with the same meaning. The NP a fox and the VP is a fox differ only in
one respect: the NP, if used referentially, predicates about its own referent, while
the VP predicates about the referent of the subject of the sentence. Such differences
are, however, eliminated in predicate logic as all one-place predicate constants are
grammatically treated in a uniform way. In predicate logic, differences such as those
between verbs, adjectives and nouns do not exist.
Rule TR6 is motivated similarly: an adjective like busy provides the same predicate
as the copula expression is busy. The three rules lead to the results in (15):
(15) a. a fox
= fox
b. is a fox
= a fox
c. is Mary
= Mary
d. is wicked = wicked
= fox( _ )
= fox = fox( _ )
= _ =m
(same for a person)
(same for is a person)
(same for is John)
= wicked( _ ) (same for is young, is busy)
6 Strictly speaking, the meanings of a fox and is a fox are not the same, as only the latter carries
present tense. The difference is not relevant in the fragment, because tense is neglected in our
treatment.
Formal semantics
339
We interpret the combination of A and N as logical conjunction, which is adequate
for the adjectives in the fragment, because these adjectives are predicating adjectives
(see 5.4.2.3 for cases of non-predicating adjectives).
T3
AN=A∧ N
Together with T2 and T5 we get for example:
busy person
= busy ∧ person = busy( _ ) ∧ person( _ ) T3
b.
a busy person
= busy person
c.
is a busy person = a busy person = busy( _ ) ∧ person( _ ) T5, b
(16) a.
= busy( _ ) ∧ person( _ ) T2, a
The rules of combining a verb with an argument, T1 and T4, introduce an
existential quantifier for the argument.
T1
NP VP = ∃x( |NP ∧ VP| _ = x )
The notation ‘|…| _ = x’ stands for the expression that we obtain when we insert
the variable x into the last empty slot of each predicate constant within |…|. Thus,
for example, |young( _ )| _ = x yields young(x) and |know( _ , _ )| _ = y results
in know( _ ,y). The procedure will be referred to as variable insertion. All of the
predicate constants treated so far are one-place. Thus, the last argument slot is the
only one. Let us apply T1 to the sentence Mary squints.
(17) Mary squints
= ∃x( |Mary ∧ squints| _ = x )
T1
= ∃x( | _ = m ∧ squint( _ )| _ = x )
TB
= ∃x( x = m ∧ squint(x) )
x insertion
Note that x insertion can only be performed when the resolution of the translation is
down to the basic translations with the empty slots. Next, we combine the NP Mary
with the VP is a busy person from (16):
(18) Mary is a busy person =
= ∃x( |Mary ∧ is a busy person| _ = x )
T1
= ∃x( | _ = m ∧ busy( _ ) ∧ person( _ )| _ = x )
TB, (16)
= ∃x( x = m ∧ busy(x) ∧ person(x) )
x insertion
The translation rule for the combination of a transitive verb with its object NP is
similar to T1. Instead of x, the variable y is used for the object referent. Rule T4 is
340
Understanding semantics
applied before T1 (because the VP must be formed before it can be combined with
the subject NP). Therefore the provision that the variable be inserted into the last
empty slot of each argument term ensures that the object argument slot is filled first.
TV NP = ∃y( |TV ∧ NP| _ = y )
T4
One example, a fox dislikes John, may suffice to show how this rule works in
combination with T1. The formation of the sentence is given in (19f), its translation
in (19t).
(19) f.
(19) t.
F4:
TV dislikes, NP John → VP dislikes John
F2:
D a, N fox → NP a fox
F1:
NP a fox, VP dislikes John → S a fox dislikes John
a fox dislikes John =
=
∃x( |a fox ∧ dislikes John| _ = x )
T1
=
∃x( |fox ∧ dislikes John| _ = x )
T2
=
∃x( |fox ∧ ∃y(|dislikes ∧ John| _ = y) | _ = x )
T4
=
∃x( |fox(_) ∧ ∃y(|dislikes(_,_) ∧ _ = j| _ = y) | _ = x )
TB
Now that we are down to the translations of the basic expressions with empty slots
for arguments, we perform the variable insertions, first the inner insertions of y and
then the outer insertions of x.
(19) (contd.) = ∃x( |fox(_) ∧ ∃y(dislikes(_,y) ∧ y = j) | _ = x )
= ∃x( fox(x) ∧ ∃y(dislikes(x,y) ∧ y = j) )
y insertion
x insertion
13.2.3.5 Discussion of the analysis
How does the system cope with the problems addressed at the end of 13.2.1? To
begin with, TB assigns the proper types of translations, i.e. one-place or two-place
predicate constants, to the basic expressions of the fragment. The four problems are
then taken care of as follows:
Referential and predicative NPs. T1 and T4 introduce referents only for referential
NPs, i.e. NPs in subject or object position. For each referent there is a variable bound
by an existential quantifier. This is adequate since reference necessarily involves
existence of the referent. Rule T5 for predicative NPs does not introduce an existential
quantifier but leaves the argument slot(s) of the NP unoccupied.
Attributive and predicative adjectives. The system treats both uses of adjectives
alike. T3 for attributive and T6 for predicative adjectives both add the adjective
meaning as a predicate conjunct to the meaning of N or to the zero meaning of is.
Formal semantics
341
Proper names and indefinite NPs are treated in the same way. They uniformly add
a predicate conjunct to the verb predication or to the zero meaning of the copula.
Argument sharing. The system properly accounts for argument sharing among
predicate constants. Two predicate terms that share an argument (A and N, NP and
VP, TV and NP) translate into predicate expressions with an empty slot and are
connected by conjunction; then the same variable is inserted into the empty slots of
conjoined predicate expressions. For the sentences in (18) and (19), Fig. 13.4 displays
argument assignments for the predicate constants and referent assignments for the
NPs (recall the analyses in Fig. 5.2 and Fig. 5.7).
Figure 13.4
Referents and arguments of two sentences
sentence
NPs
[Mary]NP is [a busyA personN]NP
(no referent)
referents
x
translation
∃x(x=m ∧ busy(x) ∧ person(x))
sentence
NPs
[a fox N]NP dislikesTV [John]NP
referents
translation
x
y
∃x(fox(x) ∧ ∃y(dislike(x,y) ∧ y=j ))
The translation system yields PL representations that essentially correspond
to the intuitive representations we used in 5.5. Superficially, the translations are
different. For example, a fox dislikes John would have been analysed there as (20a)
instead of our actual result (20b):
(20) a.
b.
fox(x) ∧ dislike(x,y) ∧ John(y)
∃x( fox(x) ∧ ∃y(dislike(x,y) ∧ y = j ) )
The difference between the two formulae is due to two properties of the translation
system. First, proper names result in identity statements like ‘y = j’ rather than
predications like ‘John(y)’. The two treatments are logically equivalent (if we assume
342
Understanding semantics
that the individual term ‘j’ denotes the only individual for which the predicate
constant John is true, which we did indeed do here). We can thus replace ‘John(y)’ by
‘y = j’. This yields (21) out of (20a):
(21)
fox(x) ∧ dislike(x,y) ∧ y = j
The result lacks the existential quantifiers binding x and y. We will see in the next
section that we actually need them if we want to achieve the intended reading.
(20b) can be replaced by a simpler logical equivalent: observing the conjunct
‘y=j’, we can replace ‘y’ by ‘j’ and eliminate the quantifier ‘∃y’ and the identity
conjunct:
(20) c.
∃x(fox(x) ∧ dislike (x,j)
Let me add a few critical remarks about this system. First, while the system
illustrates the general way in which formal semantics operates, it involves the
unusual employment of empty slots and argument insertion. This allowed me to
avoid additional formalism (such as λ operators, for those who know). But due to
this, only sentences receive regular PL-F expressions as translations: translations
of everything below sentence level contain empty argument slots, which are not
provided for by the syntax of PL-F. Second, the reader must be cautioned that the
system takes advantage of considerable restrictions of the lexical input. For example,
not all types of one-place adjectives can be treated in that way, as was pointed out
in 5.4.2.3. None of the rules applies to all adjectives, all one-place nouns, all NPs, all
intransitive or transitive verbs.
13.3 MODEL-THEORETIC SEMANTICS
We will now undertake the next major step, the introduction of a formal semantics for
the PL language and thereby indirectly for the fragment. A formal semantics for a PL
language serves one purpose: the assignment of a truth value to every formula. The
interpretation base of such a system is a so-called model. It fixes the interpretations
of all individual terms and predicate constants. The truth values of formulae derive
from these fixings by means of general rules valid for all PL languages. Thereby,
a general definition of the truth conditions of PL formulae is achieved. For each
language, an infinity of different models can be defined. We will first fix a model for
PL-F and then introduce the general interpretation rules.
13.3.1 A model for PL-F
A model for a PL language consists of two components, a so-called universe
(also ‘universe of discourse’) and an interpretation base. The universe is needed
for the interpretation of quantifiers and predicate constants. The basic idea of
interpreting an existential quantification is this: there is a given domain of cases;
Formal semantics
343
these cases are checked for the condition formulated in the scope of the quantifier;
if there is at least one case in the domain for which the scope formula is true, then
the existential statement is true, otherwise it is false. The universe of the model
provides this domain of cases. It is the ‘domain of quantification’ and plays the
same role as the domain of quantification of referential quantificational NPs in
natural language (recall 4.5). If no such domain were fixed in advance, it would
be impossible to assess that an existential quantification is false. Even if all cases
checked so far were false, there would be others yet to be checked and therefore
possibly true.
The universe indiscriminately provides the domain of cases for all quantifying
formulae. What are the ‘cases’? They are the potential arguments of the predicates
employed in the language. For example, if we are to determine the truth value of the
formula ‘∃x(fox(x) & dislike(x,y) ∧ y = j)’, the cases to be checked are the entities
the variables x and y can stand for, and these are the possible arguments of the
predicates denoted by fox and dislike. It is common practice to use the general term
individuals for all those abstract or concrete things that might be arguments of the
predicates of the PL language. Depending on the model, i.e. on the application of the
PL language, the individuals may be persons, or numbers, or points in time or sets
or whatever.
IB1 The universe U of our model of PL-F is the set consisting of the
following four individuals:
– Ken: a squinting dog
– Fritz: a young, wicked fox
– Mary: a person, neither young nor busy nor wicked nor squinting
nor a smoker
– John: a young, busy, wicked smoker
Having fixed the universe, we can assign interpretations to the basic expressions. Each
individual term receives one of the individuals in the universe as its interpretation.
The model must fix interpretations for the variables too because they can occur free
in formulae such as ‘fox(x)’ or ‘x = y’. If no interpretations for variables were fixed, it
would not be possible to determine the truth values of such formulae.
IB2 [j] = John
[m] = Mary
[x] = Ken
[y] = Fritz
The interpretations of predicate constants are entities that assign a truth value to
7
every individual in the universe of the model. As a convenient notational convention,
7 Note the different use of small capitals here; the interpretations of PL predicate constants are
neither categories nor frame attributes.
344
Understanding semantics
from here on small capitals are used for the interpretations of predicate constants,
FOX for [fox], KNOW for [know] and so on. FOX assigns 1 to Fritz and 0 to Ken, Mary
and John, because in our model Fritz, unlike Ken, Mary and John, is a fox. The table
in IB3 displays the assignments of truth values to individuals for PERSON, FOX, YOUNG,
BUSY, WICKED, SQUINT and SMOKE.
IB3
Ken
Fritz
Mary
John
[person] =
PERSON
0
0
1
1
[fox]
=
FOX
0
1
0
0
[young]
=
YOUNG
0
1
0
1
[busy]
=
BUSY
0
0
0
1
[wicked] =
WICKED
0
1
0
1
[squint]
=
SQUINT
1
0
0
0
[smoke]
=
SMOKE
0
0
0
1
Two individuals, Mary and John, are persons, one is a fox. Two individuals each are
young and wicked (Fritz and John). YOUNG and WICKED assign the same truth value in
every case, as do BUSY and SMOKE. Apart from specifying Ken as neither being a person
nor a fox, the model leaves open what species Ken is (he must have two eyes, though,
because he squints).
FOX, YOUNG, WICKED, etc. assign truth values to individuals; they are thus functions
in the mathematical sense. The particular type of function that assigns truth
values to its arguments is called a ‘predicate’ in mathematics. Since we have been
using the term predicate for a certain type of concept, we will call predicates in the
mathematical sense ‘m-predicates’.
DEFINITION 1 M-predicate
An m-predicate over a set A is a function that assigns a truth value to every
element of A.
M-predicates resemble predicates-as-concepts – let us call them ‘c-predicates’
for short – in that both yield truth values when applied to their arguments. But
while c-predicates are part of our cognitive systems, m-predicates are abstract
set-theoretical constructs, namely arbitrary pairings of arguments and truth values.
For example, the function YOUNG is the set of the four pairs Ken→0, Fritz→1,
Mary→0 and John→1. That is why m-predicates can be defined in the form of
tables that list the truth values for all arguments as we did in IB3. If two m-predicates
(or functions in general) assign equal values to all arguments, they are identical.
Formal semantics
345
Hence, in our model, YOUNG = WICKED and BUSY = SMOKE. The corresponding mental
predicates ›young‹ and ›wicked‹, or ›busy‹ and ›smoke(r)‹, are of course by no means
identical. They just happen, in this model, to be true of the same individuals. In Fig.
13.5, the seven m-predicates of our model are depicted as what they just are: sets of
argument-value pairings.
Figure 13.5
The five one-place m-predicates of the model
1
[person]
2
[fox]
3
[fat]
[wicked]
4
[busy]
[smoke]
5
[squint]
KenÆ 0
KenÆ 0
KenÆ 0
KenÆ 0
KenÆ 1
FritzÆ 0
FritzÆ 1
FritzÆ 1
FritzÆ 0
FritzÆ 0
MaryÆ 1
MaryÆ 0
MaryÆ 0
MaryÆ 0
MaryÆ 0
JohnÆ 1
JohnÆ 0
JohnÆ 1
JohnÆ 1
JohnÆ 0
There are one-place and multi-place m-predicates. The interpretations of the twoplace predicate constants know and dislike are two-place m-predicates; they assign
a truth value to pairs of individuals. Their definition requires a different form of
table. In IB4, a1 stands for the subject argument of the corresponding transitive verb,
a2 for the direct-object argument.
Fritz
Mary
John
Ken
Fritz
Mary
John
0
1
1
0
Ken
1
1
1
1
Fritz
1
0
0
1
Fritz
1
1
0
1
Mary
1
0
0
1
Mary
1
0
1
1
John
1
1
1
0
John
1
1
1
1
a2
a1
a2
Ken
Ken
IB4
a1
DISLIKE
KNOW
According to IB4, Ken dislikes Fritz and Mary, Fritz dislikes Ken and John, and so on.
In connection with m-predicates we use the common mathematical notation
‘f(x) = y’:
(22)
FOX(Fritz) = 1
for:
FOX assigns 1 to Fritz
DISLIKE(Ken, Fritz) = 1
for:
DISLIKE assigns 1 to Ken and Fritz
(in this order, i.e. Ken dislikes Fritz)
346
Understanding semantics
With the definition of the universe and interpretations for all basic expressions of
PL-F, the model – one possible model – is complete. Since the same interpretation
rules apply to all PL languages and all models, these rules are not part of a particular
model.
13.3.2 Interpretation rules for PL-F
Since all complex expressions in PL-F are formulae, we can give the interpretation
rules in a uniform format:
[S] = 1
iff
(conditions)
The conditions on the right side of iff are conditions for the sentence S being true,
i.e. its truth conditions. The first three interpretation rules take care of the prime
formulae. T and T’ are used when two terms must be distinguished:
I1
[P1(T)] = 1
iff
[P1]([T])=1
I2
[P2(T,T’)] = 1
iff
[P2]([T],[T’]) = 1
I3
[T = T’] = 1
iff
[T] = [T’]
In (23), rule I1 is applied to ‘fox(j)’:
(23) a.
[fox(j)] = 1
iff [fox]([j]) = 1
by I1
This step takes us down from the interpretation of the formula to the interpretations
of the predicate constant and its argument term. [fox] is the m-predicate FOX, [j] is
the individual John. In the next step, these are inserted:
(23) b.
[fox(j)] = 1
iff
FOX(John) = 1
by IB
We now have to check the truth value of the m-predicate FOX for the individual John
in our model. It is 0 (see IB3). Thus, the statement on the right side is false, and
consequently the one on the left side is too. In other words: [fox(j)] = 0.
An application of I2 is given in (24). (25) illustrates the interpretation of identity
statements.
(24)
[dislike(y,m)] = 1 iff [dislike]([y],[m]) = 1
iff
DISLIKE(Fritz,Mary) = 1
Since DISLIKE(Fritz, Mary) = 1, [dislike(y,m)] = 1.
by I2
by IB2, IB3
Formal semantics
[m = x] = 1
(25)
iff [m] = [x]
by I3
iff Mary = Ken
by IB2
347
Since Mary is not Ken, [m = x] = 0.
Two rules are left. The interpretation of conjunction is straightforward. It states that
a conjunction is true iff both conjuncts are true (cf. the definitions in 7.4).
I4
[S ∧ S’] = 1
iff
[S] = 1 and [S’] = 1
(26) [wicked(x) ∧ know(x,y)] = 1
iff [wicked(x)] = 1
and
[know(x,y)] = 1
by I4
iff [wicked]([x]) = 1 and
[know(x,y)] = 1
by I1
iff [wicked]([x]) = 1 and
[know]([x],[y]) = 1 by I2
iff
KNOW(Ken,Fritz) = 1
WICKED(Ken) = 1
and
by IB2,3,4
Since WICKED(Ken) = 0 and KNOW(Ken,Fritz) = 1, the conjunction is false:
[wicked(x) ∧ know(x,y)] = 0
Rule I5 for existential quantification is in need of some explanation. Grasping the
sense of such formulae is one thing, giving a precise interpretation rule, another. Let
us consider an example first:
(27) a.
∃x(fox(x) ∧ wicked(x))
In a given model, the formula is true if there is at least one individual in the universe
that is both a fox and wicked. How can this be put on terms with the general design
of the interpretation rules? What we have to do in order to check whether (27a) is
true is check the truth value of the formula fox(x) ∧ wicked(x), for different possible
interpretations of x. Is it true if we assume that [x] = John? Is it true for [x] = Mary?
And so on for each individual in our universe. If the formula is true for at least
one individual, then the existential quantification is true. This yields the following
formulation of I5, where V represents the bound variable:
I5
[∃ V(S)] = 1
iff
there is at least one individual u in U such that,
for [V] = u, [S] = 1
Applied to (27), the new rule yields the following derivation:
348
Understanding semantics
(28) [∃x(fox(x) ∧ wicked(x))] = 1
iff
iff
iff
iff
there is at least one u in U such that, for [x] = u,
[fox(x) ∧ wicked(x)] = 1
by I5
there is at least one u in U such that, for [x] = u,
[fox(x)] = 1 and [wicked(x)] = 1
by I4
there is at least one u in U such that, for [x] = u,
[fox]([x]) = 1 and [wicked]([x]) = 1
by I1
there is at least one u in U such that, for [x] = u,
FOX([x]) = 1 and WICKED([x]) = 1
iff
by IB3
there is at least one u in U such that
FOX(u) = 1 and WICKED(u) = 1
Since Fritz is such an individual, the formula is true.
We can now see why employment of an existential quantifier makes a big difference.
Compare (27) with the ‘naked’ (27b):
(27) a.
b.
∃x(fox(x) ∧ wicked(x))
fox(x) ∧ wicked(x)
The quantifier-free formula in (27b) is just about a single case, namely about the
individual we happened to fix as the interpretation of the variable x in case it is
used as a free variable. In our model, (27b) thus happens to be false because [x]
is defined as Ken, and Ken is neither a fox nor wicked. If (27b) were true, it would
be a coincidence. The use of a quantifier with a certain variable cancels the value
assignment for the bound variable. This is why the distinction between free and
bound variables is so important: free variables are simply terms denoting a certain
individual specified in the model as the value of the variable; bound variables,
however, are ‘real’ variables that stand for all individuals in the universe.
Now that the interpretation system for PL-F is complete, let us have a look at
how it works and what it buys us. It yields the assignment of an interpretation (in
the technical sense) to every basic or complex expression of the language. Thus, in
mathematical terms, an interpretation system is a complex definition of the function
[ ] which assigns each expression its interpretation in the model. Different categories
of expressions receive different types of interpretations: we assign individuals to
individual terms, m-predicates to predicate constants and truth values to sentences.
The interpretation base and the rules are designed in a way that the value of a formula
is always a truth value. For example, the interpretation rule I1 for the formula type
P1(T) produces a truth value by the following mechanism: it takes the individual [T]
and the m-predicate [P1] and lets the m-predicate work on the individual to produce
a truth value. This, then, is the resulting ‘interpretation’ of the formula, its truth value
Formal semantics
349
in the model. Table 13.1 displays the general picture for PL-F and how it works out
for such a formula.
Table 13.1
Categories of expressions and types of values
Category
individual term
predicate constant
formula/sentence
Expression
m
busy
busy(m)
Ken
→ 0
Fritz → 0
Interpretation
Mary
Mary → 0
0
John → 1
Type of interpretation
individual
m-predicate
truth value
13.3.3 Application to the translations of fragment sentences
With the interpretation system for PL-F, a first formal semantics for our fragment
of English is complete. The semantics consists of two steps. First, sentences from
the fragment are translated into PL-F. In a second step, their truth values in the
given model are determined in the interpretation system for PL-F. We will illustrate
the process for one sentence of the fragment, a fox dislikes John. The example is
somewhat complex because we have to deal with double quantification.
(29) a.
a fox dislikes John
b.
translation (cf. (19)):
c.
truth value:
∃x(fox(x) ∧ ∃y(dislike(x,y) ∧ y = j))
[∃x(fox(x) ∧ ∃y(dislike(x,y) ∧ y = j))] = 1
iff
iff
iff
there is at least one u in U such that, for [x] = u,
[fox(x) ∧ ∃y(dislike(x,y) ∧ y = j)] = 1
by I5
there is at least one u in U such that, for [x] = u,
[fox(x)] = 1 and [∃y(dislike(x,y) ∧ y = j)] = 1
by I4
there is at least one u in U such that, for [x] = u,
8
[fox(x)] = 1 and there is at least one v in U such that,
for [y] = v, [dislike(x,y) ∧ y = j] = 1
by I5
8 Each time when I4 is applied, a new variable must be chosen, hence v instead of u.
350
Understanding semantics
iff
there is at least one u in U such that, for [x] = u,
[fox]([x]) = 1 and there is at least one v in U such that, for
[y] = v, [dislike]([x],[y]) = 1 and [y] = [j]
by I1,2,3,4
iff
there is at least one u in U such that
FOX(u) = 1 and there is at least one v in U such that
DISLIKE(u,v) = 1 and v = John
iff
by IB2,4
there is at least one u in U such that
FOX(u) = 1 and DISLIKE(u,John) = 1
Since the condition is fulfilled – Fritz dislikes John, see IB4 – the sentence a
fox dislikes John is true in the model.
What is the benefit of all this formal apparatus? After all, the result looks utterly
trivial. Is not the result of the formal analysis of the sentence a fox dislikes John just a
somewhat more cumbersome way of saying the same thing? Yes, it is. But the crucial
point is: it must be that way. If the resulting truth conditions were not recognizably
those of the original sentence, the analysis would be plainly wrong. After all, semantic
analysis is to reproduce our semantic intuitions, and these are with respect to the
sentence a fox dislikes John just that the sentence says ‘a fox dislikes John’.
The main achievement of the analysis is a description of HOW THE TRUTH CONDITIONS
COME ABOUT. In the course of deriving the truth conditions for the sentences of the
fragment, we had to settle a number of semantic questions that are far from trivial.
What are the meanings of the basic expressions and what are their respective
contributions to the truth conditions of a sentence? How do their meanings combine
with those of the other parts of a sentence? How can expressions of different
meanings and forms but of the same syntactic category, e.g. proper names and
indefinite NPs, be analysed in a uniform way? The answers to these questions are
nothing less than a description of central composition rules – the main objective of
sentence semantics.
13.3.4 Model-theoretic semantics
The interpretation system for PL-F is an example of model-theoretic semantics.
Model theory is a branch of mathematics concerned with the interpretation of
formal languages such as PL. It provides general definitions of possible models and
the general rules of deriving truth conditions for the formulae of such languages on
the basis of a given model. A model-theoretic semantics is a semantics in a technical
sense in that it assigns non-linguistic entities to linguistic expressions, entities such
as referents, m-predicates and truth values which are at least somehow related to
meaning in the usual sense. When natural language sentences are translated into a
formal language and then given a model-theoretic interpretation (see Fig. 13.6 for
the general scheme), this kind of ‘semantics’ is applied to, or supplied for, natural
language. ‘Meanings’ in model-theoretic semantics are whatever interpretations the
function [ ] assigns to natural language expressions. Let us take a look at what kind
of semantics we have achieved so far.
Formal semantics
351
Figure 13.6
Two-step interpretation assignment in formal semantics
NATURAL LANGUAGE
FRAGMENT
sentences
translation system
FORMAL LANGUAGE
formulae
model-theoretic
semantics
interpretations […]
The first thing to observe is that the expressions receive interpretations only for
a given model. What is such a model? A model fixes the reference of all referring
expressions and the facts relevant for determining the truth value of every sentence.
For a natural language, a model would just be a context of utterance. There are as
many possible models as there are assignments of referents and constellations of
facts in CoUs. Hence the model-theoretic interpretation (i.e. ‘meaning’) of a proper
name is its referent in the CoU given by the model and the interpretation of a
sentence is its truth value in the model.
Clearly, the truth value of a sentence in one particular CoU is not its meaning. If
it were, all sentences true in that CoU would have the same meaning, and all false
sentences alike. Likewise, the referent of an NP cannot be equated with its meaning.
Nor does an m-predicate that assigns a truth value to every individual in a given
CoU constitute the meaning of a predicate constant. Recall how the interpretations
of young and wicked coincided in our model. Therefore, these interpretations cannot
be considered the meanings proper.
What we need in order to account for the meanings of natural language expressions
is more: a description of the potential of the expressions with respect to ALL possible
CoUs. This would, after all, gain a description of GENERAL truth conditions of predicate
expressions and sentences. In the next section of this chapter, this is accomplished by
extending model-theoretic semantics to ‘possible-world semantics’.
352
Understanding semantics
13.4 POSSIBLE-WORLD SEMANTICS
13.4.1 Possible worlds
Let us assume we are dealing not with a small fragment of English but have managed
to set up a full-fledged formation system of English, a complete lexicon and grammar.
Let us further assume that we have a translation system that yields translations of
arbitrary English sentences into an appropriate logical language LE (for Logical
English). LE has constants corresponding to all lexical items of English, in particular
predicate constants for all verbs, adjectives and nouns. What would it mean to give
a model for LE? Recall that a model is the basis for assigning truth values to all
formulae of the logical language. Since these comprise the translations of all English
sentences, the model must provide the relevant information for assigning TRUE or
FALSE to every possible English sentence. Now, sentences are not simply true or false
(except for logically true or false sentences, 7.2). Rather their truth values depend on,
and change with, the CoU. For example, the sentences in (30) keep changing their
truth values through time; at some times one is true and the other false, at others
both are true or both false.
(30) a.
b.
Mary is tired.
Mary was tired.
The time dependence is due to the fact that the VPs carry tense. That is why the time
of utterance is one component of the CoU (1.1.2). In addition, the truth values of
the sentences also depend on who is referred to by the name Mary, again a feature
of the CoU. And even if time and person reference is fixed, the truth values depend
on the particular facts given in the CoU, on what Mary has been doing and how
much and well she has slept. That is why the given facts must also be considered a
component of the CoU. When we defined the model for PL-F, we implicitly fixed all
these components of a possible CoU.
We assigned referents to the proper names; we assumed certain facts as given,
such as Fritz being young and John being a young, wicked, busy smoker; we also
implicitly fixed a reference time and interpreted all predications as relating to this
time of utterance.
What we have been calling a CoU is called a possible world in formal semantics.
There are possible worlds that correspond to reality in sharing all their facts with
what we consider our real world, while other worlds represent alternative ways of
how things might be. There is not only one ‘real’ world but many, since even if the
9
facts are real, different CoUs could be fixed within reality. Alternative, non-real
9 In some variants of formal semantics, what we call a possible world is divided into two
components. One called context comprises the parameters of the situation of utterance, i.e. it takes
care of fixing reference, while a world consists of those facts that are independent of the CoU.
In such frameworks, there is only one ‘real’ world, which, however, can be paired with different
‘contexts’.
Formal semantics
353
worlds will not be completely different, but differ only in some details. For example,
Mary may be tired in a real world w1, but might just have had a nap in some other
world w2 which is otherwise not so different from w1. This is the kind of world one
would refer to in a statement like (31):
(31) If Mary had had a nap, she would not be so tired.
The verb forms in the if-clause and the main clause indicate a shift of reference from
the real world into an alternative world. In this world Mary is said to be not as tired
as in the real one.
In actual applications of possible-world semantics (PWS), i.e. the branch of formal
semantics that employs this notion, concrete possible worlds are never actually
spelled out. This would, of course, be impossible. We are not omniscient. Possible
worlds are simply assumed as given, and even the ingredients that make up a possible
world such as time of utterance, speaker, addressee, location and circumstances are
only made explicit to the extent that is required by the semantic data to be analysed.
Possible worlds are theoretical constructs for what the truth value of a sentence may
depend on. This construction enables us to describe how reference and truth values
can be derived IF we are given the necessary information in a particular CoU.
13.4.2 Intensions
With the notion of possible worlds, we are able to assign general interpretations to
linguistic expressions. These interpretations are functions that assign the expression
not just its referent, truth value, etc., but a referent, truth value, etc., FOR EVERY POSSIBLE
WORLD.
Consider an arbitrary sentence, such as the above Mary is tired. In every possible
10
world, the sentence is either true or false. Let W be the set of all possible worlds.
Then a function can be defined that assigns to every possible world w in W the truth
value of Mary is tired in that world. In PWS the usual notation for the function is
double square brackets 冀 冁. 冀Mary is tired冁 is the function that assigns the sentence a
truth value for every possible world; the truth value in a particular world w is 冀Mary
is tired冁w .
DEFINITION 2
For arbitrary sentences S, 冀S冁 is a function that assigns to each possible world
w in W the truth value of S in that world, i.e. 冀S冁w.
10 Strictly speaking, the sentence is only true or false in those possible worlds where its presuppositions
are fulfilled. We will ignore the issue of presuppositions here.
354
Understanding semantics
This function is considered the meaning, and called the proposition, of the sentence
in PWS. You may think of the proposition of a sentence S as an infinite table that lists
a truth value for every possible world, as illustrated in (32):
(32)
Proposition of S, 冀S冁
world w1
truth value of S
1
w2
w3
w4
w5
w6
w7
w8
…
1
0
1
0
0
0
1
…
Formally, for every sentence S its proposition is an m-predicate over the set of
possible worlds. It is true of those worlds in which S is true and false of all other
11
worlds.
Let us next consider the PWS interpretations of natural language individual
terms, for example the name Mary, taken as a proper name, or the pronoun I. For
these, interpretations are defined that yield a (potential) referent for every possible
world. We will assume that part of what makes up a particular possible world w is
a ‘universe’ Uw of its own. Thus, an individual term in each world w is assigned an
individual in that universe.
DEFINITION 3
For arbitrary individual terms T, 冀T冁 is a function that assigns to each possible
world w in W 冀T冁w, the individual in Uw that is the referent of T if T is used
referentially.
For the pronoun I, 冀I冁w is the speaker who utters the sentence in world w. For Mary,
the interpretation assigns to each possible world the person referred to as ‘Mary’ in
that world. If a possible world were to have more than one such person Mary, we
would have to distinguish different names Mary1, Mary2, etc. in the lexicon and assign
them different interpretations. The interpretations for individual terms in PWS are
called individual concepts (although, again, these entities are not really concepts,
but rather merely mathematical functions). Again, you may think of an individual
concept as an infinite table like that in (32) that lists an individual for every possible
world.
11 It must be noted that a proposition in the sense of formal semantics is not a proposition in the
sense introduced in chapter 2. In perfect analogy to the distinction between m-predicates and
c-predicates, one might talk of m-propositions and c-propositions. C-propositions are mental
descriptions, i.e. concepts, of situations, while m-propositions are set-theoretical constructs,
namely assignments of truth values to possible worlds. As we saw in 5.8, c-propositions can
be considered c-predicates about situations potentially referred to. An m-proposition is the
m-predicate determined by this c-predicate about situations. Having stated this point, I will
continue to talk of ‘propositions’ instead of ‘m-propositions’ in the given context of possible-world
semantics.
Formal semantics
355
In PWS, functions of the type 冀…冁, e.g. propositions and individual concepts, are
generally called intensions. The intension of an expression assigns it its extension
for, or in, every possible world. Thus the intension of a sentence S is a proposition
and the extension of S in a particular world is its truth value there; the intension of
an individual term is an individual concept, its extension in w is an individual in U w.
DEFINITION 4 Intension, extension
For arbitrary expressions E, the intension of E, 冀E冁, is a function that assigns
to every possible world w in W an entity in w of the appropriate type. 冀E冁w is
the extension of E in the world w.
For a predicate constant, its extension in a world w is an m-predicate over the
respective universe Uw. We will therefore define the intensions and extensions of
predicate constants as follows.
DEFINITION 5
For arbitrary predicate constants P, 冀P冁 is a function that assigns to each
possible world w in W an m-predicate over Uw as its extension 冀P冁w.
Intensions of one-place predicate constants are called properties, those of multiplace predicate constants, relations. Table 13.2 gives a survey of the three types of
intensions and extensions introduced.
Table 13.2
Intensions and extensions
Expression
E
Its intension
冀E冁
Its extension in world w
冀E冁w
sentence
proposition
truth value
individual term
individual concept
individual
predicate term
(1-place)
property
function that assigns truth values
to individuals
predicate term
(2-place)
relation
function that assigns truth values
to pairs of individuals
13.4.3 Intensional models
In application to formal languages, a possible world corresponds to a model: it fixes
reference and truth values. Hence, the interpretation of an expression E in a particular
356
Understanding semantics
model corresponds to the extension of E in the corresponding possible world w: [E]
(in that model) is 冀E冁w. We will therefore call the type of models introduced in 13.3.1
extensional models.
In an intensional model for a formal language, the basic expressions receive
intensions as interpretations. Individual terms and predicate constants receive
interpretations of the following forms (for individual terms T and one-place
predicate constants P1):
(33) a.
b.
冀T冁
=
a function that assigns to each possible world w in W an
individual 冀T冁w in Uw.
冀P1冁
=
a function that assigns to each possible world w in W a oneplace m-predicate 冀P1冁w.
What we then need are intensional interpretation rules for deriving the intensions
of complex expressions. This too is straightforward. Recall the extensional rule for
prime formulae of the form ‘P1(T)’ (we attach the subscript e for extensional):
I1e
[P1(T)] = 1 iff [P1]([T]) = 1
The INtensional interpretation rule must ensure that I1e remains valid at the
extensional level. We can ensure this if we simply define I1i (i for intensional) as
follows.
I1i
for each possible world w in W 冀P1(T)冁w = 1 iff 冀P1冁w (冀T冁w) = 1
Since 冀A冁w is just [A] in the model corresponding to the world w, I1i means that the
former extensional rule I1e is to be applied for every possible world w.
The use of intensions as interpretations brings us an important step further in
coping with the problems pointed out at the end of 13.3.4. First, the interpretations
are no longer restricted to one particular CoU (possible world). Rather they cover the
range of all possible CoUs. Second, the interpretations no longer massively coincide
for expressions with intuitively different meanings. For example, the adjectives
young and wicked may happen to be true for the same set of individuals in some
possible worlds. But there will always be other worlds where they differ, because it
is POSSIBLE that someone is young and not wicked or wicked and not young. Thus,
there are always possible worlds w and individuals i in Uw such that 冀young冁w(i) ⫽
冀wicked冁w(i). Therefore 冀young冁w and 冀wicked冁w are different and hence also the
intensions 冀young冁 and 冀wicked冁. As we will see, however, intensionalization still does
not guarantee that expressions with different meanings in fact are assigned different
interpretations in PWS.
Formal semantics
357
13.4.4 Logical properties and relations
When an intensional model is supplied for the formal language into which the
sentences of natural language are translated, this yields an intensional interpretation
system for the latter. Such a system provides general truth conditions for all sentences.
For example, for the truth conditions of a fox dislikes John we obtain formally (34)
(see (28) for the derivation):
(34) for every possible world w in W, a fox dislikes John is true iff
let 冀j冁w = John; then there is at least one individual u in Uw such that
冀fox冁w(u) = 1 and 冀dislike冁w(u,John) = 1.
The availability of general truth conditions allows a very important step: the formal
definition of the logical notions introduced in Chapter 7:
DEFINITION 6
a. A sentence is logically true iff it is true in every possible world.
b. Sentence S1 logically entails sentence S2 iff for no possible world S1 is true
and S2 is false.
c. Two expressions E1 and E2 are logically equivalent, iff they have equal
extensions in all possible worlds.
We can therefore check the results of a PWS analysis by testing if the formal
interpretations produce intuitively correct logical properties and relations. For
example, the A + N combination young fox intuitively entails both young and fox:
young foxes are foxes and young foxes are young. This observation is correctly carried
out by our treatment, because it yields the following translations for the respective
test sentences with a free individual variable x:
x is a young fox
fox(x) ∧ young(x)
b.
x is a fox
fox(x)
c.
x is young
young(x)
(35) a.
Given the intensional model sketched in 13.4.3, we can easily verify the entailments.
According to what would be I4i, in every possible world w the formula ‘fox(x) ∧
young(x)’ is true iff ‘fox(x)’ and ‘young(x)’ are both true in w. Thus, whenever (35a)
is true, (35b) and (35c) are true too.
However, the intensional interpretation given so far only captures those logical
relations that originate from the application of the composition rules. The system is
not able to derive entailments based on meaning relations between lexical items, for
example that x is a dog entails x is an animal. There are two ways to handle this kind
of phenomenon.
358
Understanding semantics
First, one can set up a collection of meaning postulates such as dogs are animals,
children of the same parents are siblings, pork is meat from pigs, etc. (cf. 9.3.4). These
are declared axioms of the interpretation system, i.e. sentences that must come
out true in every possible world. The possibilities of defining the overall intension
function 冀 冁 would thereby be considerably constrained. Note that, up to now, we have
only given a general characterization of intensions. In principle, 冀 冁 can be defined
in a way such that, for example, for some world w there are individuals for which
冀duck冁w and 冀human冁w both assign 1; these individuals would then be both ducks
and human beings. If we want to exclude this kind of contradiction, we can take care
of the matter by appropriate meaning postulates.
Alternatively, we can incorporate lexical decomposition into this kind of approach.
In fact, Dowty’s theory of decomposition (9.4.1) was developed within the framework
of PWS. Lexical items to be decomposed are then translated not directly into
corresponding predicate constants (fox → fox) but into decompositional formulae.
For example, the nouns woman and man could be translated as follows:
(36) a.
b.
woman
= human(_) ∧ ¬ male(_)
man
= human(_) ∧ male(_)
We can then immediately derive that the nouns woman and man are logically
contradictory (7.5). This procedure does not constrain the intension function 冀 冁.
Rather it reduces the number of predicate constants in LE for which intensions have
to be defined. But ultimately we would of course want to constrain the interpretations
of the remaining semantic primitives, i.e. the irreducible basic expressions of LE.
13.5 THE SCOPE AND LIMITS OF POSSIBLE-WORLD SEMANTICS
13.5.1 Scope and potential
Above all, PWS is a theory that provides a precise theoretical notion of meaning:
in PWS, the meaning of a lexical or non-lexical expression is its intension in an
intensional model. For sentences, intensions are functions that assign the sentence a
truth value for every possible world, where possible world is a theoretical construct
for the CoU. The semantic description of a sentence in PWS is a definition of its truth
conditions. For all kinds of expressions, the approach provides not only a notion of
meaning (namely, intension) but also a notion of reference: for a given possible world
w, the extension of an expression E can be considered its actual or potential referent
in that world. Thus, PWS can claim to provide theoretical definitions for the following
central semantic notions:
∑ the meaning of a lexical or syntactically complex expression: its intension
∑ the context of utterance: a possible world
∑ the (potential) referent of an expression: its extension in a given world
∑ the truth conditions of a sentence: a description of its intension
Formal semantics
359
As a semantic framework, PWS is primarily aimed at describing semantic composition.
As we saw in the fragment, lexical meanings are essentially treated as simply being
given; by translating them into corresponding constants of the formal language, e.g.
know = know(_,_), they are themselves not analysed. Semantic analysis in these
cases is restricted to the assignment of the appropriate logical type (individual term
or x-place predicate constant). This corresponds to the aim of sentential semantics to
address questions such as: GIVEN the meanings of the adjective young and the noun
fox, what is the meaning of the A + N combination young fox? Neither structuralist
semantics nor cognitive semantics have so far managed to come up with a theory of
composition (but recall the application of frame theory to composition, 12.4). This is
the main asset of PWS. PWS describes composition by translating natural language
expressions into a formal logical language. The two-step procedure yields explicit
meaning representations in the form of semantically transparent and explicit logical
translations. It further allows the description of logical properties and relations.
Thus, as a descriptive apparatus, PWS provides the following:
∑ meaning representations in a logical language
∑ a description of composition rules for sentences and complex expressions in
general
∑ the derivation of logical properties and logical relations such as entailment
In addition, as we saw in the last section, PWS provides a framework for decomposition
and thereby for describing, to a certain extent, word meanings and meaning relations
between lexical items.
13.5.2 Limits
While the basic notion of intension opens up all the possibilities PWS offers, it also
demarcates its limits. First of all, meaning in PWS is defined in terms of reference and
truth conditions. Consequently all the restrictions apply to PWS that were lined out
in 7.6 for the logical approach to meaning in general:
∑ PWS cannot account for non-descriptive meaning, in particular social and
expressive meaning and the meaning of sentence type (cf. 2.2.3).
∑ Even descriptive meaning is only captured indirectly.
In terms of the semiotic triangle, PWS exclusively describes the relation between an
expression and its extensions, i.e. the right side of the triangle (Fig. 13.7).
In fact, it is not that PWS just focuses on the right side of the triangle; rather it
disregards the conceptual level completely. The two-step link from an expression via
its meaning to its extensions or denotation is bypassed, as it were, by the intension,
i.e. by the direct assignment of extensions to the expression. Since to PWS the
intension IS the meaning, there is nothing in the PWS model that corresponds to the
component ‘meaning’ in the semiotic triangle. To PWS it does not matter how the
truth conditions of a sentence come about. What counts is what the truth conditions
objectively are. But can we do without meaning? Can we account for the semantic
360
Understanding semantics
Figure 13.7
Semiotic ‘triangle’ for possible world semantics
expression
intension
descriptive
meaning
extension
data on the basis of direct expression → intension → extension assignments? The
answer is: partly, but not totally.
In PWS, two expressions with the same intension must be considered synonymous,
because synonymy is defined as meaning identity and the intensions are the
meanings. However, as we saw in 7.6.1 and 8.1, logical equivalence does not even
amount to descriptive synonymy (let alone total synonymy, which would also include
non-descriptive meaning). In particular, sentences with identical truth conditions
need not have the same descriptive meaning. One extreme consequence of the PWS
approach is that all logically true sentences receive the same meaning, because they
all have the same intension, namely the one and only function that assigns the value
TRUE to every possible world. Thus, contrary to our semantic intuitions, the following
sentences all have the same meaning in PWS:
(37) a.
Donald is a duck or Donald is not a duck.
b.
All dogs are dogs.
c.
Two plus two equals four.
etc.
The list grows considerably, if meaning postulates are added as axioms. All meaning
postulates would mean the same because they would all come out as logically true
in the resulting interpretation system. Intuitively, however, all these sentences, for
instance the examples in 9.3.5, have quite different meanings.
It is not only logically true sentences whose meanings coincide, contrary to
intuition, in a PWS approach. Likewise, pairs of sentences such as the following
turn out semantically indistinguishable (recall the discussion in 7.6), if the system
contains the relevant meaning postulates, i.e. that the day following Sunday is a
Monday:
Formal semantics
(38) a.
b.
361
Today is Sunday.
Tomorrow is Monday.
What holds for natural language expressions also applies to the meaning
representations in the formal language: they may look like having (and thereby
representing) different meanings, but actually fail to do so. For example, one might
prefer a translation system that produces (39a) rather than (39b) as the translation
of Mary squints, i.e. an analysis that treats proper name NPs as directly inserted into
argument slots:
(39) a.
b.
squint(m)
∃x(x = m ∧ squint(x))
When we interpret the formulae, as we inevitably do, in the same way as we interpret
natural language sentences, the two formulae are semantically different. One is a
simple predication with a one-place predicate applied to one definite argument. The
other is much more complex: an existential quantification applied to a conjunction of
two different predications. But as long as the formulae receive the PWS interpretations
they do, the difference is an illusion: the formulae have the same meaning because
they are logically equivalent. At best, they can be seen as representing the meaning
of Mary squints in different ways. But they do not represent two different meanings.
13.5.3 Possible-world semantics vs mentalist semantics
The shortcomings of PWS with respect to the semantic distinction of logically
equivalent expressions are immediately due to eliminating meanings proper from
the semiotic scheme by equating sentence meaning with truth conditions. To PWS,
meanings are but assignments of extensions to CoUs (possible worlds). Thus, PWS
provides a model for capturing the context dependence of truth and reference. But
it does not provide an answer to the question of how all this works at the cognitive
level. Clearly, we do not have ready-made intensions in our minds that produce the
relevant extensions (truth values, referents, m-predicates) for any context we might
encounter. Intensions are infinite, because there are an infinite number of logically
possible CoUs. Of course, PWS does not claim that we have intensions in our minds.
Rather, the approach does not aim at all at the cognitive level. But the cognitive level
exists and we would like to understand how it works.
From the mentalist perspective on meaning which has been advocated throughout
this book, truth conditions and reference are but an EFFECT of meaning, not meaning
itself. For example, the meanings of (38a) and (38b) are two situation concepts
that hold true of the same situations in the same CoUs, but they do so for different
reasons. They are different concepts. Ultimately, the task of semantics is a description
of the concepts that constitute the meanings of the expressions we use. PWS
contributes to this task in describing what these concepts must yield in terms of
truth and reference.
362
Understanding semantics
Let us compare the situation of a semanticist with that of someone who is to
describe a computer program. Essentially, computer programs are entities that
produce certain outputs when fed with certain inputs. With a PWS-like approach one
would give a complete account of the input-output characteristics of the program,
a table which for each input gives the resulting output, similar to the table given
in (32) for propositions in a PWS approach. An approach in the vein of mentalist
semantics would aim at a description of the program itself, its structure and the way
it works. Given such a description, the input-output characteristics of the program
can be predicted and explained. Clearly, differently written programs may have the
same input-output characteristics. And since this is so, we would not agree that we
know what is essential about a program as long as we just know its input-output
characteristics. Analogously, we can consider sentence meanings as mental software
that when fed with the description of a situation produces the output TRUE or FALSE.
PWS tries to capture the input-output characteristics of the software, but mentalist
semantics would try to describe the software itself, how it works and why it produces
the truth values it does. We do not know what is essential about the meaning of
a sentence as long as we only know its truth conditions. The crucial thing about
meaning is the kind of software it constitutes.
EXERCISES
1. Describe the structure of a compositional interpretation system in your own
words.
2. Introduce categories in the original system for Japanese numerals. (You can
choose arbitrary names for the categories, just be consistent and introduce
different names for different categories.) Assign categories to the basic expressions
noting that an expression may belong to more than one category. Reformulate the
formation and interpretation rules in the style of those for the English fragment.
3. The Japanese word for 100 is hyaku. The word for 200 is nihyaku (lit. ‘two
hundred’), and so on, up to kyūhyaku for 900. (Actually, some terms for the
hundreds are slightly irregular due to phonological effects of concatenation, e.g.
sanhyaku → sanbyaku. These irregularities are to be ignored in the exercise.)
To the terms for the hundreds as first parts the terms for the numbers from 1 to
99 are directly concatenated. For example, the term for 539 is gohyaku sanjūkyū.
Modify the system of rules and categories from exercise 2 to cover all numerals
up to the word for 999.
4. Describe the formation and compositional interpretation of the numeral
nanahyaku nijūichi in the style of (5).
5. a. Give a full account of the formation of the fragment sentence
A wicked fox squints in the way illustrated in (7).
b. Derive, step by step, the translation of the sentence as is done in (17), (18).
Formal semantics
363
c. Derive, step-by-step, the truth value of the translation of the sentence in the
given extensional model of PL-F (cf. (28)).
6. a. Derive the translation of A young person is Mary (the sentence is odd, but do
it for the sake of the exercise).
b. Reduce the formula by quantifier elimination.
7. Consider the PL-F formula wicked(m).
a. Derive the truth conditions of the formula in the given model for PL-F.
b. Find a fragment sentence that has the same truth conditions.
c. Translate the sentence into PL-F.
d. Derive its truth conditions.
8. a. Define an extensional model for PL-F with only three individuals, where Mary
is a young squinting fox, John a wicked dog, x is Mary and y is Sheila, a busy
smoking person. Every individual knows every other individual (including
themselves) and dislikes Mary (including Mary herself). Give complete
definitions of the interpretation of all basic terms of PL-F in that model.
Which m-predicates are equal in this model?
b. Determine the truth values in this model for the formulae in (23), (24), (25),
(27), (28), making use of the derivations given there.
9. Explain the notions possible world, intension and extension and how they are
related.
10. Describe the overall scheme of possible-world semantics. How is meaning
defined in PWS? What is the method of semantic description in PWS?
11. Explain why extensional models are insufficient for a semantic description of
natural language sentences.
12. Explain why meaning cannot be fully described in terms of truth conditions and
reference.
13. What is the difference between m-predicates and predicates?
14. Explain the difference between a proposition in the conceptual sense introduced
in 2.2 and a proposition in the sense of PWS.
15. In which respect is PWS superior to cognitive semantics? In which respect is
cognitive semantics superior to PWS?
FURTHER READING
There are several introductions to formal semantics. Some of the more recent ones
are, in alphabetic order: Cann (1993), Chierchia and McConnell-Ginet (1990), de
Swart (1998), Heim and Kratzer (1998). Bach (1989) offers an informal introduction
to central topics of formal semantics.
References
Aitchison, Jean. 2012. Words in the Mind. Oxford: Blackwell.
Anderson, Stephen R. and Edward L. Keenan. 1985. Deixis. In Shopen, Timothy (ed.),
Language Typology and Syntactic Description. Volume III: Grammatical Categories and the
Lexicon. Cambridge: Cambridge University Press, pp. 259–308.
Andersson, Lars-Gunnar and Peter Trudgill. 1990. Bad Language. Oxford: Blackwell.
Aristotle. 1958. Metaphysics. A revised text with introduction and commentary by W.D. Ross.
Oxford: Clarendon Press.
Bach, Emmon. 1989. Informal Lectures on Formal Semantics. Albany, NY: State University of
New York Press.
Barsalou, Lawrence W. 1992a. Cognitive Psychology. An Overview for Cognitive Sciences.
Hillsdale, NJ: Erlbaum.
Barsalou, Lawrence W. 1992b. Frames, Concepts, and Conceptual Fields. In A. Lehrer and
E.F. Kittay (eds), Frames, Fields, and Contrasts: New Essays in Semantic and Lexical
Organization. Hillsdale, NJ: Lawrence Erlbaum Associates, pp. 21–74.
Berlin, Brent and Paul Kay. 1969. Basic Color Terms. Their Universality and Evolution. Berkeley,
CA and Los Angeles, CA: University of Los Angeles Press.
Bickerton, Derek 1992. Language and Species. Chicago: Chicago University Press.
Bierwisch, Manfred. 1982. Formal and lexical semantics. Linguistische Berichte 80: 3–17. Also
published in Hattori, Shiro and Kazuki Inoue (eds), Proceedings of the XIIIth International
Congress of Linguistics, August 29–September 4, 1982, Tokyo. Tokyo, pp. 122–36.
Bierwisch, Manfred and Ewald Lang (eds). 1983. Dimensional Adjectives: Grammatical
Structure and Conceptual Interpretation. Berlin, New York, Tokyo: Springer.
Brown, Penelope and Stephen Levinson. 1978. Universals in language usage: politeness
phenomena. In E. Goody (ed.), Questions and Politeness: Strategies in Social Interaction.
Cambridge: Cambridge University Press, pp. 56–311.
Brown, R. and A. Gilman. 1960. The pronouns of power and solidarity. In T. Sebeok (ed.),
Style in Language. Cambridge, MA: MIT Press.
Bryson, Bill. 1990. The Mother Tongue. New York. Harper.
Cann, Ronnie. 1993. Formal Semantics: An Introduction. Cambridge: Cambridge University
Press.
Cann, Ronnie. 2011. Sense relations. In Maienborn, von Heusinger and Portner (eds, 2011),
pp. 456–80.
Carlson, Gregory. 2012. Genericity. In Maienborn, von Heusinger and Portner (eds, 2012),
Volume 2, pp. 1153–85.
Chierchia, Gennaro and Sally McConnell-Ginet. 1990. Meaning and Grammar. An Introduction
to Semantics. Cambridge, MA: MIT Press.
Comrie, Bernard. 1976. Aspect. Cambridge: Cambridge University Press.
Comrie, Bernard. 1985. Tense. Cambridge: Cambridge University Press.
References
365
Cooke, Joseph R. 1968. Pronominal Reference in Thai, Burmese, and Vietnamese. Berkeley and
Los Angeles: University of California Press.
Corbett, Greville G. 2000. Number. Cambridge: Cambridge University Press.
Croft, William and D. Alan Cruse. 2004. Cognitive Linguistics. Cambridge: Cambridge
University Press.
Cruse, D. Alan. 1986. Lexical Semantics. Cambridge: Cambridge University Press.
Davis, Anthony R. 2011. Thematic roles. In Maienborn, von Heusinger and Portner (eds,
2011), pp. 399–419.
Dedrick, Don. 1998. Naming the Rainbow. Colour Language, Colour Science, and Culture.
Dordrecht: Kluwer.
Deutscher, Guy. 2010. Through the Language Glass. Why the World Looks Different in Other
Languages. London: Cornerstone.
de Swart, Henriëtte. 1998. Introduction to Natural Language Semantics. Stanford: Centre for
the Study of Language and Information.
Diessel, Holger. 1999. Demonstratives. Form, Function and Grammaticalization. Amsterdam/
Philadelphia: Benjamins.
Diessel, Holger. 2012. Deixis and Demonstratives. In Maienborn, von Heusinger and Portner
(eds, 2012b), Volume 3, pp. 2407–2432.
Dillon, George L. 1977. Introduction to Linguistic Semantics. Englewood Cliffs, NJ: PrenticeHall.
Dowty, David R. 1979. Word Meaning and Montague Grammar. Dordrecht: Reidel.
Drivonikou, G.V., P. Kay, T. Regier, R.B. Ivry, A.L. Gilbert, A. Franklin and I.R.L. Davies. 2007.
Further evidence that Whorfian effects are stronger in the right visual field than in the
left, Proceedings of the National Academy of Sciences 104: 1097–102.
Dryer, Matthew S. and Haspelmath, Martin (eds). 2011. The World Atlas of Language
Structures Online. Munich: Max Planck Digital Library. Available online at http://wals.
info
Engelberg, Stefan. 2011a. Lexical decomposition: foundational issues. In Maienborn, von
Heusinger and Portner (eds, 2011), pp. 124–45.
Engelberg, Stefan. 2011b. Frameworks of lexical decomposition of verbs. In Maienborn, von
Heusinger and Portner (eds, 2011), pp. 358–89.
Fillmore, Charles J. 1968. The Case for Case. In E. Bach, R. T. Harms (eds). Universals in
Linguistic Theory. New York. Holt, Rinehart & Winston, pp. 1–90.
Foley, William A. 1997. Anthropological Linguistics: An Introduction. Oxford: Blackwell.
Galton, Antony. 1984. The Logic of Aspect. Oxford: Clarendon Press.
Givón, Talmy. 1993. English Grammar. A Function-based Introduction. Amsterdam and
Philadelphia, PA: John Benjamins.
Goddard, Cliff. 1998. Semantic Analysis. A Practical Introduction. Oxford and New York:
Oxford University Press.
Gussenhoven, Carlos and Haike Jacobs. 2011. Understanding Phonology. Third edition. New
York, London: Routledge.
Harris, Roy. 1983. F. de Saussure: Course in General Linguistics. Trans. Roy Harris. Oxford:
Gerald Duckworth.
Haspelmath, Martin. 2010. Understanding Morphology. Second edition. New York, London:
Routledge.
Heim, Irene and Angelika Kratzer. 1998. Semantics in Generative Grammar. Oxford: Blackwell.
Huang, Yan. 2007. Pragmatics. Oxford, New York: Oxford University Press
Huddleston, Rodney and Geoffrey K. Pullum 2002. The Cambridge Grammar of the English
Language. Cambridge: Cambridge University Press.
366
Understanding semantics
Hughes, Geoffrey. 1992. Swearing. A Social History of Foul Language, Oaths and Profanity in
English. Oxford: Blackwell.
Isac, Daniela and Charles Reiss. 2008. I-language. An Introduction to Linguistics as a Cognitive
Science. Oxford: Oxford University Press.
Jackendoff, Ray. 1990. Semantic Structures. Cambridge, MA: MIT Press.
Jackendoff, Ray. 2009. Compounding in the parallel architecture and conceptual semantics.
In Rochelle Lieber and Karol Štekauer (eds), The Oxford Handbook of Compounds. Oxford,
New York: Oxford University Press, pp. 105–28.
Jackendoff, Ray. 2011. Conceptual semantics. In Maienborn, von Heusinger and Portner (eds,
2011), pp. 688–709.
Kay, Paul and Chad K. McDaniel. 1978. The linguistic significance of the meaning of basic
color terms. Language 54: 610–46.
Kay, Paul and T. Regier. 2003. Resolving the question of color naming universals. Proceedings
of the National Academy of Sciences of the United States of America 100: 9085-9089.
Published online 10 July 2003, http://www.pubmedcentral.nih.gov/articlerender.fcgi?tool
=pmcentrezandartid=166442#ref1.
Kearns, Kate. 2011. Semantics. Second edition. Basingstoke, New York: Palgrave Macmillan.
Kövecses, Zoltán. 2006. Language, Mind and Culture. Oxford: Oxford University Press.
Kummer, Manfred. 2005. Politeness in Thai. In Richard J. Watts, Sachiko Ide, Konrad Ehlich
(eds), Politeness in Language: Studies in its History, Theory, and Practice. Berlin, New York:
De Gruyter, pp. 325–36.
Kuroda, S.-Y. 1973. Where epistemology, style and grammar meet: a case study from
Japanese. In Kiparsky, Paul and Stephen Anderson (eds), A Festschrift for Morris Halle.
New York: Holt, Rinehart and Winston, pp. 377–91.
Lakoff, George. 1987. Women, Fire, and Dangerous Things. What Categories Reveal about the
Mind. Chicago, IL: The University of Chicago Press.
Lee, Penny. 1996. The Whorf Theory Complex. A Critical Reconstruction. Amsterdam,
Philadelphia, PA: Benjamin.
Levin, Beth. 1993. English Verb Classes and Alternations. A Preliminary Investigation. Chicago,
IL: The University of Chigaco Press.
Levinson, Stephen C. 1983. Pragmatics. Cambridge: Cambridge University Press.
Löbner, Sebastian. 1985. Definites. Journal of Semantics 4: 279–326.
Lucy, John A. 1992. Language Diversity and Thought. Cambridge: Cambridge University Press.
Lyons, John. 1977. Semantics. Cambridge: Cambridge University Press.
Lyons, John. 1995: Linguistic Semantics. An Introduction. Cambridge: Cambridge University
Press.
Maienborn, Claudia, Klaus von Heusinger and Paul Portner. 2011. Meaning in linguistics.
In Maienborn, von Heusinger and Portner (eds, 2011), pp. 1–10.
Maienborn, Claudia, Klaus von Heusinger and Paul Portner (eds). 2011. Semantics: An
International Handbook of Natural Language Meaning. Volume 1. Berlin, New York: De
Gruyter.
Maienborn, Claudia, Klaus von Heusinger and Paul Portner (eds). 2012a. Semantics: An
International Handbook of Natural Language Meaning. Volume 2. Berlin, New York: De
Gruyter.
Maienborn, Claudia, Klaus von Heusinger and Paul Portner (eds). 2012b. Semantics: An
International Handbook of Natural Language Meaning . Volume 3. Berlin, New York: De
Gruyter.
Martin, Samuel E. 1975. A Reference Grammar of Japanese. New Haven, CT, and London: Yale
University Press.
References
367
Matthews, Peter. 2001. A Short History of Structural Linguistics. Cambridge: Cambridge
University Press.
Matthewson, Lisa. 2011. Methods in cross-linguistic semantics. In Maienborn, von Heusinger
and Portner (eds, 2011), pp. 268–85.
The New Oxford Dictionary of English. 1998. Oxford: Clarendon Press.
Olsen, Susan. 2012b. Semantics of compounds. In Maienborn, von Heusinger and Portner
(eds, 2012b), Volume 3, pp. 2120–2150.
Palmer, Frank R. 1994. Grammatical Roles and Relations. Cambridge: Cambridge University
Press.
Palmer, Frank R. 2001. Mood and Modality. Cambridge: Cambridge University Press.
Palmer, Gary B. 1996. Toward a Theory of Cultural Linguistics. Austin, TX: University of Texas
Press.
Partee, Barbara H., Alice ter Meulen and Robert E. Wall. 1993. Mathematical Methods in
Linguistics. Dordrecht: Kluwer Academic Publishers.
Plag, Ingo. 2003. Word-formation in English. Cambridge: Cambridge University Press.
Pustejovsky, James. 1995. The Generative Lexicon. Cambridge, MA: MIT Press.
Radford, Andrew. 1988. Transformational Grammar. Cambridge: Cambridge University Press.
Riemer, Nick. 2010. Introducing Semantics. Cambridge: Cambridge University Press.
Salzmann, Zdenek. 1993. Language, Culture and Society. An Introduction to Linguistic
Anthropology. Boulder, San Francisco and Oxford: Westview Press.
Sapir, Edward. 1921. Language. An Introduction to the Study of Speech. New York: Harcourt,
Brace and World.
Seuren, Pieter A.M. 1991. Präsuppositionen. In: A. von Stechow and D. Wunderlich (eds),
Semantik. Ein internationales Handbuch der zeitgenössischen Forschung/Semantics. An
International Handbook of Contemporary Research. Berlin, New York: De Gruyter, pp.
286–318.
Siewierska, Anna. 2004. Person. Cambridge: Cambridge University Press.
Shibatani, Masayoshi. 1990. The Languages of Japan. Cambridge: Cambridge University Press.
Sohn, Ho-Min. 1999. The Korean Language. Cambridge: Cambridge University Press.
Suzuki, Takao, 1978. Japanese and the Japanese. Words in Culture. Tokyo, New York, San
Francisco: Kodansha.
Tallerman, Maggie. 2011. Understanding Syntax. Third edition. New York, London: Routledge.
Tyler, Andrea and Hiroshi Takahashi. 2011. Metaphors and metonymies. In Maienborn, von
Heusinger and Portner (eds, 2011), pp. 597–621.
Ungerer, Friedrich and Hans-Jörg Schmid. 2006. An Introduction to Cognitive Linguistics.
Second edition. Harlow: Pearson Education.
Van Valin, Robert D. 2001. An Introduction to Syntax. Cambridge: Cambridge University Press.
Van Valin, Robert D. 2005. Exploring the Syntax–Semantics Interface. Cambridge: Cambridge
University Press.
Vendler, Zeno. 1957/67. Verbs and times. Philosophical Review 56: 143–160. Reprinted 1967 in
Vendler, Z., Linguistics in Philosophy. Ithaca: Cornell University Press, pp. 97–121.
Verschueren, Jef. 1999. Understanding Pragmatics. New York, London: Routledge.
Wales, Katie. 1996. Personal Pronouns in Present-day English. Cambridge: Cambridge
University Press.
Watts, Richard J. 2003. Politeness. Key-topics in Sociolinguistics. Cambridge: Cambridge
University Press.
Wierzbicka, Anna. 1996. Semantics. Primes and Universals. Oxford, New York: Oxford
University Press.
Whorf, Benjamin Lee. 1940. Science and linguistics. In Whorf (1956), pp. 207–19.
368
Understanding semantics
Whorf, Benjamin Lee. 1956. Language, Thought and Reality. Selected Writings by Benjamin
Lee Whorf, ed. John B. Carroll. Cambridge, MA: The MIT Press.
Wittgenstein, Ludwig. 1958. Philosophische Untersuchungen/Philosophical Investigations.
Trans. G.E.M. Anscombe. Oxford: Blackwell.
Index
In the index, ‘f ’ refers to the following page, ‘ff ’ to the following two pages; wider ranges of
pages are given explicitly. Consult the bold type page numbers for the pages where the term is
introduced and defined. If you look up a combination of adjective plus noun, e.g. ‘intransitive
verb’, look for a sub-entry of the noun, except for fixed terms like ‘personal pronoun’, ‘social
meaning’ or ‘possible world’.
accommodation 50, 100f
accomplishment term 141–4, 146, 155, 211
achievement term 144ff, see also simple
change term
active voice 124, 136f
activity term 142ff, see also process term
addressee 5, 28ff, 32, 62ff, 67, 70, 94, 113,
192, 253, 353
adjective: as predicate term 117ff;
attributive use 117, 336, 340;
comparative 118; non-predicational
119, 207; predicative use 118, 336, 340;
scalar 210; two-place 118
adjunct 108, 135
adverbial quantification 92f
adverbs/adverbial 33f, 210, 214, 291; deictic
71f; expressive 34f; of frequency 92;
quantifying 93, temporal 72f, 159
affix 9, 226f
affordance 315f
agent (role) 122–6, 128f, 135–44, 165, 214,
227, 230, 238, 310, 312, 316, 319f
aggregate mass nouns 81
agreement 124f
aktionsart 141, see aspectual class
alienability 104, 117, 218
all 74, 85, 89, 90, 171, 210, 291
all-or-nothing contrast 88f, 91f, 98, 290
alternation 136–40; causative/inchoative 139,
237; conative 139; dative 139f; simple
reciprocal 140; understood reciprocal
object 140; understood reflexive object
140; unspecified object 139
ambiguity 41–59, 148, 221; compositional 48f;
contextual 49–54; contextual vs lexical
56ff; of grammatical forms 49; lexical 44ff
anaphor, anaphora 64, 72; definite
descriptions 78
anchoring 15, 62, 74, 162
antecedent 64f, 78f
anticausative 138f
antipassive 137, 139f, 206
antonymy/antonym 209f, 213, 215f, 236, 244
appositions 78
arbitrariness 223
argument 108–11; core ~ 135; implicit
232f, 241, 253; oblique 135; parasitic
110; referential 110f, 113, 115, 116,
151, 241; relational 116f, 119, 312f, see
also possessum; situational 162
argument compound 318
argument frame 311f
argument sharing 120f, 130f, 336, 341
argument structure 123, 135–40, 162, 311f
argument term 108–11
Aristotle 168, 170, 267
arity 108
article 10; articles across languages 83;
definite 5, 19f, 72, 75–9, 258; indefinite
81ff
aspect 150–7, 162; habitual 149f, 154, 160;
imperfective 150, 153f; perfect 154f;
perfective 150–3; prospective 155f; in
English 163; in German 163; in Russian
163; in Japanese 164; interaction with
tense 157ff
370
Understanding semantics
aspectual class 140–50
associative plural 66, 104
at-adverbial 146, 147, 149
attribute: (in a frame) 302–11; (in
grammar) 48; (in logic) 168f; ~ terms
77, 307; ~ types 309f
attributive use 117ff, 131, 336, 340
Austin 6
auxiliary 172
bare mass noun 81, 83, 86
bare plural 81, 83, 86
Barsalou 266, 301, 302, 307, 321f
Barsalou frames 301–23
basic category 276–9, 286, 298, 310
basic expression 327
basic level 276–9
BCT basic colour term see colour term
be about to 155
be going to 155
BECOME 238, 244f
Berlin and Kay 258–61, 268
BFA see Binary Feature Approach
Bickerton 13
Bierwisch 51, 54
big 209f, 236, 244, 287f
Binary Feature Approach 228–36, 245,
268, 298, 301, 321f
binary features 228–36, 321f
body part term 35, 218, 233, 235
both 74, 85, 89, 90
bottom-up process 11, 41, 56
bound morpheme 227
bound variable 335, 348
bra 193, 257f, 278
break 138f, 243f
brother 82, 190, 217, 22, 257
Burmese: personal pronouns 67; social
meaning 31
c-predicate 354
case 124
categorization 266–84
category 19, 24; basic 276–9; ~ boundary
268, 280f, 288, 292; cognitive 266f;
cultural 292–8; grammatical 42, 226,
237, 329, 331f; semantic 292–8
causative voice 137f, 162
CAUSE 238, 240f, 244f
change verb 145ff, 148, 149, 155, 164, 179,
206, 211
Chinese (Mandarin): agreement 124;
articles 83; personal pronouns 67;
social meaning 31; tense 72
citation form 43
CKA see Cultural Knowledge Approach
classeme 234
coda 225
coercion 58, 144, 147, 149, 212
cognate 44
cognitive semantics 269, 284f, 294, 299,
324, 359
coherence 95
collective predication 87
colour term 47, 57, 212, 224, 244, 258–61,
268, 287f, 289
communicative meaning 6f, 12, 15, 16, 28,
30, 38, 41
comparative 8, 27, 42, 117f, 210, 212, 214,
227
complement 96, 108ff, 112, 118ff, 124f,
127f, 130f, 230; clause ~ 98, 112; core
~ 135; oblique ~ 135
complementarity: logical 190f, 194, 196,
212, 219; opposition 212, 216, 230,
232, 236, 245
complex words 8, 12, 226f
composition 7–12, 56, 131, 319ff, 326–9
composition rule 14, 15, 327, 328, 346ff,
356
compositional ambiguity 48f
compositional meaning 7
compositionality: Principle of 10ff, 14
compound 12f; and cultural knowledge
294ff; argument ~ 318; frame ~
317f; meaning 13, 207, 294ff, 316–19;
regular 207, 294ff, 316–19; synthetic
318f; subtypes 207, 316f; value ~ 318
conative 139, 143, 144, 147
concatenation 326
concept 19f, 22, 24f, 36, 47, 52ff, 58, 80f,
108, 140–9, 192f, 219, 239ff, 266f,
296f, 302
Conceptual Semantics 239ff
conclusion (logical) 175
condition (of categorization): necessary
~ 267f, 270, 272f, 279f, 283f, 288; see
also features
Index
conjunction (‘and’) 121, 127, 184, 235, 334;
in logic 184; logical vs natural 185f;
presuppositions 200
connotation 36–9
constituent 10, 225
constraint (in frames) 310f
content word 2, 16, 20–5, 38, 130, 191, 203
context 3ff, 45, 80, 114,125, 128, 155, 169,
249, 287f; sentential 56, 128, 321; social
78; see also context of utterance
context dependence 4–7, 49–56, 62–74, 78ff,
85, 143, 276, 283, 292
context of utterance 5, 15, 19–26, 28, 30, 33f,
47, 49, 55f, 64f, 69, 72, 73–9, 91f, 94ff,
98–102, 110, 111, 170–4, 197f, 200, 282,
291, 351ff, 356, 358, 361
contextual ambiguity 49–59
contextual information 49
contextual knowledge 16
contextual meaning shifts 51–4, 55–9
contextual polarity 198
contingent sentence 173f, 177f, 181ff, 193,
196
contradiction, logical/logically
contradictory 179f, 182f, 190, 196, 209,
230, 358
contrariety, logical/logically contrary 179f,
182f, 194, 212f, 285
contradictory see logically false
contrastive property 226
converses 213f
conversion (in word formation) 12
copula 118, 119
core argument 135
coreferent, coreference 79
CoU see context of utterance
count noun 80f, 83
cue validity 274
culmination 141, 145
cultural see category; concept; knowledge
Cultural Knowledge Approach 294ff
culture 263
Danish: formal personal pronouns 67
de Saussure 223f, 246
declarative see sentence
decomposition 221–45, 322, 357f
definite article 15, 19f, 75–9, 258; evolution
72; function 79; indexicality 78
371
definite descriptions 75, 82; anaphoric use
78; deictic use 78
definite NPs 82; ~ and quantification
85f; mass ~ 87; plural ~ 86f;
presuppositions 96
definiteness 74–80, 115; of demonstratives
72; of personal pronouns 64
deictic 62
deictic centre 62
deictics: adnominal 74; adverbial 74;
pronominal 74
deixis 62–74, 101; person ~ 63–70; place
~ 70, 119; social ~ 40, 104; time ~ 72f,
119, 158; see also tense
demonstratives 70ff; adjectival 71, 72;
adnominal 71, 79, 82; anaphoric use 72;
definiteness 72; distal 71, 72; English
71; German 71; Japanese 70f; meaning
71f; medial 71, 72; pronominal 71, 82;
proximal 71; Spanish 71; systems 71
denotation 23–6, 188–96, 265f, 288, 359f;
actual 24; of colour terms 258–62; of a
content word 24; of mass nouns 80; of
plural nouns 80; of a sentence 25f
derivation 12, 226f, 312
descriptive meaning 20–30, 32, 34–8, 71,
73, 192–6, 204, 222, 266, 359; of content
words 22; of expressions with social
meaning 28f; of expressives 35; of
grammatical forms 23; of sentences 23
determination 74–83; and predication 115
determiner 71, 74, 84f, 89f, 93
deverbal noun 312f
diathesis 136–9, 162
dictionary form 43
differentiation: of categorization 289ff;
conceptual 53f, 58f, 101, 174, 206;
lexical 289
dimensions of meaning 18–39, 192, 203f,
359
directional opposites 210f
disambiguation 45, 49, 50f, 129
discourse roles 63, 67
discrimination 37
disjunction: in logic 184; logical vs natural
186
distal 71
distinguisher 234
distributive predication 87f
372
Understanding semantics
domain of quantification 83f, 96, 101, 343
DoQ see domain of quantification
Dowty 237ff
drink 240; English, German, Japanese 249
dual (number) 66
each 74, 84, 85, 89, 90, 96
eat 114, 139, 141, 143; English, German,
Japanese 249
economy of language 46, 47, 50, 290, 297
egressive 145
English: 2nd person pronouns 67; causative
138; demonstratives 71; perfect 155;
possessive constructions 116f; sibling
terms 190; tense and aspect system 163
entailment, logical 99, 175f, 180–3, 188,
195f, 359; and meaning 195f; mutual
175; ~ reversal 176; transitivity 178;
unilateral 176; vs if – then 177, 181; of
words 188
episodic 91, 149, 153, 156; quantification
92f, 96, 101
equivalence, logical 176, 178f, 183, 357; and
meaning 192–6, 203f, 360f; of words
187f, 203f, 213
euphemism 37, 204
event 113, 153
every 74, 85, 89, 90, 93, 171f
evidential 254ff
exclamation 34
existential quantifier 334, 347
experiencer (role) 123, 251
expression meaning 1–4, 14ff
expressive 33ff; ~ adverbs 34; vs
propositional meaning 34
expressive meaning 33ff, 192, 196, 203f,
359; and truth conditions 192; of an
utterance 38
extension 354f
face-to-face 62
facets of meaning 58, see dimensions of
meaning
factive verb: presuppositions 98
family resemblance 270f, 272, 273
feature 228–36, 268, 270f, 273f, 283, 288,
294ff, 322f; ~ semantics 228–36, 322f
felicity conditions 30
Fillmore 311, 324
focal colour 259f, 262, 268, 287
for-adverbial 144, 146, 147, 149
formal semantics see semantics
formality 16, 28f, 31f, 38, 66f, 69, 192, 222,
251
formation base 327, 330, 332
formation rule 327, 331f, 333f
fountain pen 193
fragment (of English) 330–41, 349f
frame 245, 301–23; basic structure 306f
frame compound 317f
frame referent 304, 307
FrameNet 324
free morpheme 227
free variable 335, 348
Frege 11, 170
French: definite articles 72;demonstratives
71; formal personal pronouns 67, 69;
gender 65, 258; headache construction
251ff; indefinite articles 81; [r] sound
223
function (of objects) 78, 217, 278, 283
function word 2f, 13, 20, 23
functional concept 307ff
functional noun 77, 115f, 308f, 318
fusion 127ff, 131, 241
future tense 161f, 163f
fuzziness 269f, 280–3
gender 42, 64f, 104f, 128, 258; of pronouns
63f, 66, 115, 128
gender languages 64f, 258
gender systems 104
Generalized Quantifier Theory 105
genericity/generic: habitual 149f; NPs 90–3
German: anticausative 139; bitte, danke
29; demonstratives 71; future tense
161; gender 65; headache construction
251f; non-past 160; passive 137;
personal pronouns 63f; progressive 163;
pronouns of address 28f, 31, 63f, 66f,
192, 250; reflexive verbs 140; tense and
aspect system 163; word order 125
give 139f, 240
GO 240f
goal (role) 123
graded membership 271ff, 279ff, 285f, 291,
298f
graded structure 269, 271ff, 279ff, 284, 292
Index
grammatical category 42
grammatical form 8, 42; ambiguity 49
grammatical meaning 8f, 12, 14, 15, 20, 37f,
48f, 135, 323; see also aspect; number;
sentence type; tense
grammatical person see person
(grammatical)
grammatical property 42
Grice 99
group 80
habitual aspect 149f, 154, 160
he, she 64f, 94ff, 234
head: of a compound 204ff, 236, 295, 298,
317ff; of an NP 115, 117
hedge 289
heteronymy 212f
hierarchy: of categories 276–9; lexical 189f,
213, 216f; mereologial 218f; of thematic
roles 135; of verb functions 162
historic present 149, 160, 164
holonym 219
homography 45
homonymy 43ff; and polysemy 46f
homophony 45
Hungarian: formal personal pronouns 67;
gender 66; headache construction 252;
interjections 33; possessive affixes 68,
252; sibling terms 190; verb agreement
124
hyperonym 205
hyponymy 191, 194, 195, 204–7, 213, 216f,
218, 219, 222, 227, 230, 232, 236, 245,
319
I (pronoun) 2, 4f, 15, 32, 34, 62, 64, 68, 115,
242f, 250, 354
idiom 41, 59
ignorance 282
I-language 14
illocutionary act 6f
imperative 15, 27, 29, 112, 285
imperfective aspect 150–4, 156–61, 163f; see
also progressive aspect; habitual aspect
imperfective verb 42, 163
impersonal constructions 137
implicature, conversational 6, 99
in-adverbial 143, 144, 146, 147
inalienability 104, 117, 218
373
inclusive/exclusive we 104
incompatibility, logical 179, 182, 230; of
words 189, 196, 209, 230, 231, 245
indefinite article 19, 74, 75, 81ff, 90, 331,
338
indefinite NP 81ff, 93, 336, 338, 341; and
quantification 85f, 88, 89; generic 90ff
indefinite pronoun 83
indexicality 15, 62, 67, 75, 78, 80, 94, 101
individual (logic) 343f
individual concept (in possible world
semantics) 354f
individual constant 120, 333, 335, 342
individual noun 77f, 115f
individual term 120f
individual variable 120f, 333, 335, 342, 348
indivisibility, presupposition of ~ 98, 151,
198, 290
inference 6, 99f, 101, 177
inflection 68, 112
information: contextual 20, 49, 352f;
semantic 2, 14, 27f, 71, 75, 86, 93, 94f,
107f, 127f, 134, 152f, 174, 177f, 321f; ~
transfer 18ff
informativeness 174, 178, 279
ingressive 145
initial condition 97, 144ff, 312
instrument (role) 123, 124, 135, 139
intension 353ff
interaction, social 6, 7, 18, 28ff, 36, 39, 66,
251, 302; with objects 278, 316
interface 15; of expression meaning and
utterance meaning 15, 56, 57f, 62,
101; of expression meaning with
communicative meaning 15, 30; of
expression meaning with utterance
meaning 15, 56, 101; of language and
thinking 38
interjection 33, 244
interpretation base 327, 329, 343ff, 348
interpretation rule 12f, 327f, 329, 330, 346f,
356; see also composition rule
interrogative sentence 4, 15, 27, 99f, 285f;
meaning 27; presuppositions 99;
proposition 27
intransitive (verb) 42f, 111, 114, 333
Irish: articles 83
irregularity (semantic) 129f
Italian: formal personal pronouns 67
374
Understanding semantics
Jackendoff 13, 127, 239
Japanese: addressing 251, 257; agreement
124; articles 83; brother terms 222;
causative 137f; demonstratives
70f; evidentials 254ff; headache
construction 252–6; homophones
45; numerals 325ff; omission of verb
complements 114; order of verbal
affixes 162; personal pronouns 67;
possessive pronouns 68; progressive
164; self-reference 250; seichō 45;
sibling terms 190; simple change terms
146; social meaning 31, 113; tense and
aspect system 164
Kay 258
knowledge: cultural 292–9; expert ~ 16,
293; personal 292ff; semantic 16, 232,
292–9; semantic vs cultural 296ff; world
~ 16, 138, 177, 232f, 302
Korean: agreement 124; personal pronouns
67; plural 9; social meaning 31, 113
Labov 269, 281, 283
Lakhota: (in)alienable possession 117
Lakoff 315
Law of Contradiction 168f, 170, 185
Law of the Excluded Middle see Principle of
Polarity
levels of meaning 1–7, 15f, 37f, 41
Levin verb classes 139f
lexeme 41–4, 46f, 224, 226f, 306; composite
43
lexical entry 43
lexical field 214–19
lexical meaning 7, 41ff, 49, 50, 221, 231, 245,
267, 302, 326, 329, 359
lexicon 7, 8, 15, 41, 47, 50, 57, 236, 256, 289,
297, 309, 329
light 44
linking 123ff, 139, 227, 241, 320
localism 240
locative 123
locutionary act 6
logic 167–200; and meaning 191–200; and
semantics 196–200; see also predicate
logic; sentential logic
logical exclusion 179
logical independence 183
logical relation 175–83, 187–96, 203,
230, 245, 357ff; between contingent
sentences 182f; between sentences
175–83; between words 187–91;
and meaning 182, 183, 191–200;
pathological cases 180ff; survey 180,
190
logically false sentence 25, 49, 173f, 180ff,
185, 193, 352
logically true sentence 25, 173f, 180ff, 185,
193, 215, 234, 352, 357, 360
many 83, 90, 92, 290f
marker 234
marry 116, 128, 140, 186, 214, 312
mass noun 80ff, 86, 106, 249; bare 81
mass/count distinction 80f
material implication see subjunction
meaning 1–16, 18–39, 191–6, 223, 226ff,
358–62; communicative 6f, 12, 15,
16, 28, 30, 38, 41; compositional 7; of
compounds 13, 207, 294ff, 316–19; of
a content word 20; descriptive 20–30,
32, 34–38, 71, 73, 192–6, 204, 222,
266, 359; grammatical 8f, 14, 15, 49;
lexical 7, 41ff, 49, 50, 221, 231, 245, 267,
302, 326, 329, 359; and logic 191–200,
358–62; of personal pronouns 64ff; of
possessive pronouns 68f; of a sentence
20; unpredictable 13
meaning postulate 234, 358, 360
meaning relation 58, 123, 182, 183, 191–6,
203–19, 221f, 227, 230, 232, 236f, 239,
244f, 265, 297ff, 357ff
meaning shift 5, 14, 15, 49–54, 56, 57ff,
101, 110, 128ff, 144, 147, 149, 160, 174,
299, 314, 321; differentiation 53f, 58,
59, 101, 174, 206; metaphor 52f, 56–59,
101, 129, 194, 206, 208, 299, 315, 324;
metonymy 51f, 53, 56, 57ff, 101, 129,
153, 233f, 313ff; and polysemy 58f; and
presuppositions 101; and selectional
restrictions 129
meaning variant 44–50, 53, 57ff, 127, 203,
204, 207, 216, 249, 299, 316, 321
meaning variation 281
medial (demonstratives) 71
mentalist semantics 14, 38, 285, 361f
mereology 217ff, 278
Index
meronymy 217ff
metaphor 52f, 56–9, 101, 129, 194, 206,
208, 299, 315, 324
metonymical relation 233f
metonymy 51f, 53, 56, 57ff, 101, 129, 153,
233f, 313ff
microsenses 58
modal adjectives 119
modal verb 135, 291
modality 135, 161, 162
modalization 291
model 342–5; extensional 356; intensional
355–8
model-theoretic semantics 342–58
modification 206
monosemy 46
Montague Grammar 237
mood 8, 135, 162
morpheme 226f
motivation (linguistic); of meaning
components 232f
m-predicate 344ff, 348, 350, 351–6, 361
m-proposition 354n
Natural Semantic Metalanguage 242ff, 245
negation 3, 87, 90, 98ff, 134, 145, 171f,
180, 285f; external vs internal 89;
of generic sentences 91f; in logic
184; in logic vs language 185; and
quantification 88ff; radical 199f
negation test for presuppositions 100
no 73, 83, 84, 85, 86, 89, 172, 210
none 86
non-past tense 159ff, 163f
noun 9f, 13, 42f; 115ff, 130, 134f; count ~
80f, 83; deverbal 113, 312f, 315, 318;
functional 77f, 115, 308f; gender 42f,
65, 104f, 258; grammatical meaning
9; individual 77f, 115; inherent
uniqueness 76ff, 116; logical relations
187–91; mass 80ff, 86, 106, 249;
meaning relations 207, 210, 212, 214,
217, 218; one-place 115; partitive
84ff; predicative use 119f, 338;
relational 69, 77, 115ff, 235; sortal 69,
78, 115ff
noun phrase see NP
NP 10; definite 74–80, 82, 86f, 96, 197;
generic 90–3, 96; indefinite 81ff, 85f,
375
90ff, 330, 336, 341; predicative 336,
340; quantifying 83–89, 92f, 197;
referential 336, 340; simple indefinite
83, 86, 90ff
NSC (necessary and sufficient conditions)
see condition (of categorization)
nucleus (syllable) 225
number (grammatical): 9; dual 66; paucal
66; pronouns 63, 65, 66; singular and
plural 80; trial 66; in verb inflection
68, 112, 124
numeral 78, 80, 81, 83, 85, 90, 215, 236,
249; composition 325ff; Japanese
325ff; ordinal 78
object (grammatical): direct 10, 106, 112,
122, 124f, 135ff, 140, 320; indirect 106,
122, 135, 137
oblique see complement
occurrence terms, simple 146ff
one (numeral) 85; (pronoun) 115, 126, 137
one-place, two-place see arity
onion: nominal 90; verbal 162
onset (syllable) 225
open 95, 97, 112, 114, 122f, 126f, 135, 138,
139, 206, 211, 237f, 243
opposition 207–14, 227
orthograpy 43
paradigm 63; in structuralism 225f
pars pro toto 58
part, constitutive 219, 309
participant 122, 135, 162
participants see thematic roles
particular sentence see episodic
partitive 84ff
partitive NPs 84ff
partonomy 219
parts of speech 42
passive voice 8, 123f, 136f, 139, 165, 214;
of causative 138
past tense 8f, 31, 42, 49, 72f, 113, 131, 150,
151, 154, 156–9, 163f, 211
path (role) 123, 135, 239ff, 312f
patient (role) 123, 125, 135ff, 140, 312
paucal (number) 66
perfect aspect 20, 154–9, 163f;
experiential 155; of persistent
situation 155, resultative 154f, 164
376
Understanding semantics
perfective aspect 8, 134, 150–3, 155–61,
163f; vs imperfective 150–3; with
present tense 159f
perfective verb 42, 163
performative verb 149, 160
person (grammatical) 63, 64, 68
person deixis 63–70
person terms 65, 216, 219
personal pronoun 2, 4, 63–8, 82, 95f, 115,
250; definiteness 64; gender 64f, 104f;
languages without 67; meaning 63–6;
number 104f; paradigms 66, person
meaning 64; plural 65f; strategies of
formal address 66f
phoneme 223, 225, 228
phonology 228
place deixis 70, see demonstratives
plural 8f, 42f, 33, 65ff, 74, 80–3, 124, 204,
227; associative 66, 104; collective 87;
definite 88f, distributive 87
polarization 171, 198, 285f, 288f, 291f, 298f
politeness see formality
political correctness 37, 204
polysemy 44, 45ff, 124, 127, 161, 203, 208,
216, 249, 299; and homonymy 46f;
and meaning shifts 58f; principled 58;
systematic 58
possession 69, 116f; (in)alienable 117, 218;
verbs of change of ~ 139, 208f, 235, 240
possessive chain 309
possessive construction 4, 116f, 120, 316
possessive pronoun 2, 4, 68–71, 74, 82, 90,
95f, 116; meaning 4, 67–70; reference
21, 116; social meaning 28, 69
possessive suffix (Hungarian) 68, 252
possessor 68, 76, 77, 116, 303f, 306–11, 318
possessum 69f
possible world 352–8
possible-world semantics 352–62: limits
359ff; vs mentalist semantics 361f;
scope 358f
potential referent 19, 21–4
power button (terms for) 194
PP 48, 55, 108, 112, 118, 120, 134, 262; by-PP
124, 136; of-PP 84, 116; to-PP 106, 320;
with-PP 124, 312
pragmatics 6, 16, 62
pragmatism 282
predicate 108–11; c-predicate 344;
m-predicate 344; of the sentence 10, 62,
169
predicate constant 120, 333
predicate expression 108–11, 115–20, 123,
125–131, 187f, 190, 213, 228, 235f, 241,
286, 298, 319, 337, 341, 351
predicate logic 120ff; semantics 342–9;
syntax 333–6
predicate term 108–11; see also predicate
expression
predication 6, 16f, 62, 106–30, 134, 197ff,
284, 286, 298, 319f; central 134;
collective 87; and determination 115;
distributive 87f, 98; quantified 83–6, 89,
93, 98
predicative: adjective 117ff, 125, 131, 336,
340; noun, NP 82, 119f, 125, 131, 336,
340
prefix 9
premise 175
preposition 2, 42, 124, 214f
prepositional object, adjunct, complement
112, 122, 135
prepositional phrase see PP
present tense 3f, 8, 72f, 124, 149f, 157–61,
163f, 198, 227; of dynamic predications
149; with perfective aspect 159f; with
timeless states 160
presupposition 62, 94–101, 151f, 170f, 185,
196–200, 285; accommodation 101; of
achievement terms 146; ~ carriers 96ff;
of definite NPs 96; existential 96; of
factive verbs 98; of indivisibility 98, 151,
197f, 290; ~ and meaning shifts 101;
and the Principle of Polarity 196–200;
projection problem 200; of quantifers
96; selectional restrictions 97; tests 100;
and truth conditions 98ff; of verbs of
change 97, 145
prime formula 333f
Principle of Compositionality 10ff, 14, 56,
326
Principle of Consistent Interpretation 49,
55ff, 100f, 110, 125, 128, 169, 174, 321
Principle of Polarity 170f, 174, 184f, 197–
200, 285f
principled polysemy 58
Index
process term 142ff, 147, 149, 150, 153, 158,
164
productive 12
progressive aspect 8, 56, 143, 144, 146, 147,
149, 150, 153, 156, 157, 163f; of state
predications 149
projection problem of presuppositions 200
pronominal deictic 74
pronoun 2, 15, 19; impersonal 137;
indefinite 83; of address 28ff, 40, 63,
66f, 250; see also personal pronouns;
possessive pronouns
proper names 51, 82, 83, 95f, 106, 108, 115,
319, 330, 336ff, 341, 151, 354
property 58; ~ attribute 309f; combinatorial
225, 226, 228, 230, 234, 241, 245, 329;
compositional 222, 245; contrastive
226, 234, 235; distinctive 44, 123;
grammatical 42ff, 124; interactional
316; logical 172–5,192, 194, 196f,
202, 357, 359; paradigmatic 226, see
contrastive property; in possible world
semantics 355; of prototypes 271, 322;
syntagmatic 226, 228
proposition 22–8, 29, 34, 38, 112, 116, 130f,
193, 284, 286, 321; in possible-world
semantics 354f
propositional meaning see descriptive
meaning
prospective aspect 155ff, 158, 163f
protolanguage 13
prototype semantics 284–92
prototype theory 267f, 279–84, 291f
prototypes 267–76, 268f, 322; features 274;
similarity 275f
proximal (demonstrative) 71
punctuality 146, 148, 151, 156, 158
Pustejovsky 241, 245
quantification 83–90, 289ff, 330; adverbial
92f; and negation 88ff; domain of ~ 84,
343; episodic 93; generic 92f; universal
see all; each; every
quantified predication 84
quantifier 83ff, 89f, 92f, 171f;
presuppositions 96
quantifier, existential 334, 347ff
quantity specification 81ff, 86, 90, 92
quantized object 143
377
question test 100
questions 285; see also integrative sentence
radical negation 199f
reading 49
reciprocal 186, 214, 312; ~ alternation 139,
140
reference 2, 62, 196, 325, 351, 355, 358, 361;
of adjectives 21; of content words 21f;
and descriptive meaning 21f; of nouns
21; of sentences 22; of tense 21; of verbs
21
referent 2; of a frame 304, 307
referential argument 110; of nouns 110,
117; of verbs 113, 151, 241, 311
relation: in possible world semantics 355;
mereological 233; metonymical 233;
paradigmatic 224f, 230; syntagmatic
224f, 227f; transitive 178, 219; see also
logical relation; meaning relation
relational noun 69, 77f, 115ff, 120, 135, 218,
235
relativism 256ff, 262
Relevance Theory 6
repetition, repetitive 147, 150, 153, 154, 156
representation: meaning ~ 238, 240, 242,
245, 330, 359, 361; mental 266
resultant condition 96, 144–8, 155, 211, 312
rhyme (syllable) 225
roles see thematic role
Rosch 269, 273, 277
running commentary 149, 160
Russian: articles 83; formal personal
pronouns 67; gender 42, 65; headache
construction 252f; no copula 134;
perfective and imperfective verbs 42;
tense and aspect system 163f
Sapir 256f, 262, 292f
Saxon genitive 82, 309
scalar adjectives 210
scenario 301
schema 301
scope: of possible world semantics, 358f; of
a quantifier 334; of semantics 14ff
script 301
Searle 6
selectional restrictions 56, 97, 101, 125–30,
221f, 228, 230, 232, 235, 240, 308
378
Understanding semantics
self-contradiction 49ff, 55, 168, 173n; see
also logically false
self-reference 31, 40, 250f
sell (opposites) 208, 209, 211, 213, 214
semantic irregularity 129f
semantic knowledge 16, 232, 292, 299
semantic prime 242ff
semantic roles see thematic role
semantic uniqueness sees uniqueness
semantics 13–16; cognitive 269, 284f, 294,
299, 324, 359; formal ix, 80, 299, 323,
ch. 13; and logic 196–200; mentalist
14, 38, 285, 361f; model-theoretic 342;
possible-world ~ 351–62; prototype
~ 284–92; structuralist 226ff; of word
formation 15
seme 234
semelfactive 147; see also simple occurrence
terms
semiotic triangle 24, 25, 26, 38, 111, 131,
194, 203, 205, 224, 265, 284, 296, 359
sentence: contingent 174; declarative
4, 15, 26, 27, 30, 99, 100, 170ff, 197,
285f; imperative 15, 27, 38, 112, 285;
interrogative see interrogative sentence;
logically false 25, 49, 173f, 180ff, 185,
193, 352; logically true 25, 173f, 180ff,
185, 193, 215, 234, 352, 357, 360
sentence meaning 7–12, 14ff, 22–8, 47f, 55f,
130f, 196, 298f, 319ff, 336–42, 359–62
sentence type 15, 27, 162; declarative
27; imperative 27; interrogative 27;
meaning 27f, 37, 192, 359
sentential logic 183–7
shared knowledge 177
she, he 64f, 94ff, 234
shift see meaning shift
sign 223f
similarity to ptototype 170ff, 279f, 283f, 322
simple change term 145ff, 148, 149, 153,
164
simple occurrence term 146ff, 149, 150, 153,
155, 164
singular 8f, 20, 74, 80
situation 20; ~ described, expressed 22, 27,
34, 141, 157, 178, 193; ~ referred to 22,
130f, 135, 174, 178
situation structure 140–50, 312
sm 82f, 86
social deixis see social meaning: personal
pronouns
social interaction 6, 7, 18, 28ff, 36, 39, 66,
251, 302
social meaning 16, 28–31, 113, 192, 196,
203f, 250, 359; vs expressive meaning
35f; personal pronouns 66f; of a
sentence 28; and truth conditions 192;
of an utterance 38
some 84, 89, 171f, 290, 335f
sortal noun 69, 76, 78, 115f, 117
sound form 43
source (role) 135, 312
source domain 53
Spanish: anticausative 138; definite article
72; demonstratives 71; formal personal
pronouns 67; gender 65, 258; person
inflection of verbs 68, 112f; terms for
‘bra’ 193
speaker 5, 29, 32, 62ff, 67, 70f, 94, 113, 192,
253, 353
speech act 6, 15, 27, 30, 149
speech act theory 6
speech community 14, 16, 256f, 281, 289,
293
spelling 43
Sperber and Wilson 6
state term 56, 148ff, 151, 153, 156, 163,
237ff
stem 9
structuralism 222–36
subcategory 188, 266, 273f, 296
subject 10, 32, 83, 111f, 122–5, 135f
subjunction 184ff; ~ and entailment 186;
vs if 186
subordination/subordinate 188, 189, 205; vs
hyponymy 195
suffix 9, 226f
sum 80
superlatives 8, 42, 47, 78, 210, 212
superordinate 188
swear word 34f
synchronization 14
syncretism 63
synonymy 191, 203f, 360; partial 204
syntactic structure 12
syntagm 225
syntagmatic property see property:
combinatorial
Index
taboos 257f
tautology see logically true
taxonomy 216f, 276–9
tense 3, 72f, 113, 124, 131, 156–62; absolute
157; in English 163; in German 163;
intraction with aspect 157ff; in Japanese
164; relative 157; in Russian 163f
tense argument 156, 157ff
tertium non datur see Principle of Polarity
Thai: personal pronouns 67; social meaning
31
thanking 29, 36
thematic role 122–5, 135, 241, 310f
theme (role) 123, 125, 135ff, 140
theory of common usage 282f
theta role see thematic role
they 2n, 64, 66, 96; impersonal 137
thou 67
time of utterance 72, 158
Tok Pisin, term for bra 193, 252
top-down process 11, 56
transitive see relation; verb
translation 248–51; into predicate logic
336–40
tree 189
trial (number) 66
truth 4f, 25f, 29, 98ff, 110, 169–73, 192–6,
197f, 285f, 325, 351, 361
truth conditions 25f, 30, 38, 99, 170, 192f,
197f, 325, 342, 346, 351, 357–60; and
presuppositions 98ff, 197f
truth table 184
truth value 25, 170
truth-value gap 197, 285n
Turkish: articles 83; sibling terms 190
Ungerer and Schmid 294ff
uniqueness 75–9; inherent 76ff; pragmatic
75; semantic 76f, 79
uniqueness conditions (frames) 307
uniqueness presupposition 95
universal quantifier see all; each; every
universalism 256ff, 262
universe (of discourse) 342
used to 154
utterance meaning 4ff, 12, 16, 37f, 41, 49f,
55f, 91, 110
vagueness 47f, 282, 286ff, 291f, 299
379
value compounds 318
value of a frame attribute 303, 307
variable 120f; bound 335; free 335
Vendler 150
verb 111–14, 134f; accomplishment 141f,
143f; achievement 144ff; activity 142ff;
causative 206f, 237ff, 243f; of change
211; of change of possession 139;
of change of state 139; ditransitive
112; factive 98; finite 21f, 27, 31f,
63, 113, 124f, 171, 185; of grooming
and bodily care 140; inchoative 139,
145, 206, 237f; of ingesting 139;
intransitive 111; number of arguments
113f; performative 149, 160; process
142ff; punctual 146, 156; reciprocal
312; referential argument 113, 151,
311; simple change 144ff; simple
occurrence 146ff; of social interaction
214; state 148ff, 153; transitive 112;
zero-place 135
verb of change: presuppositions 97
verb phrase see VP
Vietnamese: personal pronouns 67; social
meaning 3
voice (of sounds) 225
voice (of verbs) 122, 136–9; see also active;
passive; conative
VP 10, 42, 55, 331f
wash 143
we 63n, 64ff; associative plural 65; inclusive
and exclusive 66
Whorf 256f, 262
Wierzbicka 242ff, 245
Wittgenstein 271
word class 12, 42, 111, 114, 130, 134, 228,
330
word formation 7f, 12f, 226f, 312f;
semantics 12f, 15, 207, 226f, 294ff,
316–19
word order 27, 112, 124f
work: English, German, Japanese 249
world knowledge 16, 138, 177, 232f, 293, 302
world view 256, 258, 264
you 2, 28f, 35, 62, 65ff, 96, 115, 137, 192,
242f, 250; impersonal 137
you all, you guys 67
Praise for the first edition
‘Understanding Semantics is an original and innovative resource for introductory
courses on linguistic semantics, excelling in particular in conciseness and accessibility
of presentation.’
Linguistics 43-2
‘It is a careful and thorough volume which should be the set text of choice for every
undergraduate course in semantics. The lecturer is presented with a ready-made
course; the student with a key tool for the study of semantics. Both will benefit from
Professor Löbner’s clear presentation of the breadth of ideas and methods of a broad
subject.’
Graeme Davis, University of Northumbria, UK
Praise for the second edition
‘Sebastian Löbner’s Understanding Semantics is a rare treat for both seasoned
researchers in semantics and novices to the field who are just embarking on the
journey towards the ultimate goal in the study of language: the understanding of
what meaning is. Löbner’s direct and engaging style makes even the most difficult
concepts easy to understand and I believe this book should be on every semantics
curriculum as it has been on mine since it first appeared.’
Luna Filipovic, University of East Anglia, UK
‘This excellent and well-written introductory textbook provides students with easy
access to the fascinating world of semantics. It is unique in successfully combining
lexical, cognitive and formal approaches to semantics, and in offering a coherent
picture of elementary concepts as well as of advanced theories.’
Klaus von Heusinger, University of Cologne, Germany
‘Sebastian Löbner deserves great credit for producing an introduction to linguistic
meaning that is both accessible and thorough. The approach, which integrates
the key concepts of formal semantics with a cognitive perspective on meaning, is
engaging and refreshing: readers will come away with an understanding of how
truth-conditional semantics is done, but, more importantly, they will have a clear
sense of how meaning works in natural language – and why this is such a rich and
rewarding field of study.’
Chris Lucas, School of Oriental and African Studies, University of London, UK
‘Without stinting on presenting core areas of sentential logic, predication, and lexical
relations, the new edition of this comprehensive survey of semantics is enhanced
by three new chapters. Students will benefit from the exercises that cover the
intersection of semantics and pragmatics at the lexical, sentential and cognitive
levels and require research within both their own, as well as typologically diverse,
languages.’
Laurel Smith Stvan, University of Texas at Arlington, USA
Download