We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
There are various ways to evaluate the Generative Lexicon (GL). One is to see to what extent it accounts for what we find in text corpora. This has not previously been done, and this chapter presents a first foray. The experiment looks at the ‘nonstandard’ uses of words found in a sample of corpus data: “nonstandard” is defined as not matching a literal reading of any of the word's dictionary definitions. For each nonstandard instance we asked whether it could be analyzed using GL strategies. Most cases could not. The chapter discusses in detail a number of nonstandard uses and presents a model for their interpretation that draws on large quantities of knowledge about how the word has been used in the past. The knowledge is frequently indeterminate between “lexical” and “general,” and is usually triggered by collocations rather than a single word in isolation.
Introduction
The GL claims to be a general theory of the lexicon. Pustejovsky identifies “the creative uses of words in novel contexts” (Pustejovsky, 1995, p. 1) as one of two central issues which GL addresses, where other formal theories have remained silent. He asserts as a principle that “a clear notion of semantic well-formedness will be necessary in order to characterise a theory of possible word meaning“ (ibid, p. 6) and identifies a generative lexicon as a framework in which a core set of word senses is used to generate a larger set, according to a set of generative devices. Most work in the GL tradition has been concerned to identify and formally specify those devices.
What is the meaning of a word? How can the few hundreds of thousands of words we know be used to construct the many millions of utterances we make and understand in a lifetime? It would appear we need more words than we have available to us, if classical wisdom on this subject is to be believed. The subject, of course, is lexical semantics, and classical wisdom can often be wrong. This field has undergone a radical shift in recent years, in large part because of two developments. First, formal frameworks for word meaning have been developed that greatly simplify the description of lexical classes and their properties. Second, we have at our disposal new compositional techniques that allow us to view word meaning as an integral part of the overall process of semantic interpretation. These and other factors have made the issues relating to ‘the meaning of a word’ some of the most central questions being addressed in the field of linguistics today. In fact, some classic issues have resurfaced with new data and arguments, such as the debate over analyticity and semantic knowledge, as well as the evidence of a distinction (or nondistinction) between lexical and world knowledge.
Waismann (1951) argued for what he called the ‘open texturedness’ of terms. Although he was mainly interested in how the notion applies to the nonexhaustive nature of material object statements and the absence of conclusive verification conditions, there is another sense in which this is an interesting property of language and language use; the infinite variability of reference in language is the direct product of the essential incompleteness of terms and their composition.
We consider Pustejovsky's account of the semantic lexicon (Pustejovsky, 1995). We discuss and reject his argument that the complexity of lexical entries is required to account for lexical generativity. Finally, we defend a sort of lexical atomism: though, stricly speaking, we concede that lexical entries are typically complex, still we claim that their complexity does not jeopardize either the thesis that lexical meanning is atomistic or the identification of lexical meaning with denotation.
Introduction
A certain metaphysical thesis about meaning that we will call Informational Role Semantics (IRS) is accepted practically universally in linguistics, philosophy, and the cognitive sciences: the meaning (or content, or “sense”) of a linguistic expression is constituted, at least in part, by at least some of its inferential relations. This idea is hard to state precisely, both because notions like metaphysical constitution are moot and, more importantly, because different versions of IRS take different views on whether there are constituents of meaning other than inferential role, and on which of the inferences an expression occurs in are meaning constitutive. Some of these issues will presently concern us, but for now it will do just to mention such familiar claims as that: it's part and parcel of “dog” meaning ‘dog’ that the inference from “x is a dog” to “x is an animal” is valid; it's part and parcel of “boil” meaning ‘boil’ that the inference from “x boiled y” to “y boiled” is valid; it's part and parcel of “kill” meaning ‘kill’ that the inference from “x killed y” to “y died” is valid; and so on (see Cruse, 1986, Chap. 1 and passim).
The goal of this part of the volume is best explained by the recent interest in evaluating which lexical semantics resources are available for Natural Language Processing (NLP) and whether a methodology can be established for building large-scale semantic knowledge-bases.
The papers address the topic of understanding and structuring word meaning from the particular perspective of building NLP systems. Here, the problem of how to represent word meaning has fairly strict requirements: It affects the choice of particular data structure as well as the specific architectural requirements for a computational lexicon.
All of the contributions in this section present practical questions and dilemmas that are not usually faced in theoretical research. The first issue is one of methodology: Can existing semantic resources (viz. ontologies) be reproduced by identifying a consistent set of criteria? This is an important question, because a positive answer would mean that we have achieved an understanding of how to model conceptual knowledge independently of domains and people's intuitions.
The second issue is whether existing lexical semantics frameworks provide the basis for developing a large-scale resource on systematic grounds. Although related to previous question this problem is also relevant to the notion of “scalability.” In other words, if a particular framework makes certain claims about how meaning should be structured to account for a limited number of linguistic facts, is the suggested structuring sufficiently general to cover a large corpus? Furthermore, were that corpus to become even larger, would the lexicon scale up easily to cover the newer data? More or less explicit answers are found in the different contributions.
This study concerns the causal discourses that express a direct causation. With the help of the extended event structure for causative verbs proposed in Pustejovsky (1995), I will show that they involve an event coreference relation when the result is expressed by a causative verb in its transitive use. Then I will define two types of event coreference: generalization and particularization. Next, I will show that discourses expressing a direct causation with a resultative rhetorical relation involve a generalization relation (which explains their awkward behavior), while those discourses with an explanation rhetorical relation involve a particularization relation (which accounts for their normal behavior). Finally, I will study discourses in which the result is expressed with an unaccusative form of a causative verb. This study leads to question the extended event structure for unaccusatives proposed in Pustejovsky (1995).
Direct Causation and Event Coreference
The Notion of Direct Causation
It is well known that causal relations can be of different kinds. Among them, the direct causal relation is often mentioned in the literature and among others, by Fodor (1970) and Schank (1975). In the line of these works, I define the notion of a direct causation on conceptual grounds as follows: the result is a physical change of state for an object Y, the cause is an action performed by a human agent X, the action is the direct cause of the change of state.
Observations on the creative aspect of language use constrain theories of language in much the way as those on the poverty of stimulus do. The observations as Chomsky discusses them will be explained and their consequences for a semantic theory of the lexicon explored.
Introduction
Chomsky began to mention the creative aspect of language use in the early 1960s; his 1964 Current Issues discusses it. In 1966, Cartesian Linguistics takes it up in detail. At the end of the 1950s he had read extensively in the works of Descartes, Cudworth, Humboldt, and others in the seventeenth to mid-nineteenth centuries “Cartesian linguistics” tradition, and they provided a framework for articulating these ideas. Arguably, though, they were implicit in his review of Skinner and even in The Logical Structure of Linguistic Theory (ca. 1955). The creative aspect observations, along with the poverty of stimulus observations, offer a set of facts with which his and – he holds – any science of language must contend. However, he thinks that the lessons of the creative aspect observations in particular are often ignored, especially in dealing with semantic issues such as truth and reference, meaning and content. He may be right. I review the creativity observations and discuss some of their implications and suggestions for a semantic theory of the lexicon. In their light, I discuss briefly James Pustejovsky's different approach.
In this paper, we present recent extensions of Generative Lexicon theory (Pustejovsky, 1995; Pustejovsky, 1998) in the context of the development of large-scale lexical resources for twelve different European languages: the SIMPLE model.
The development of lexical resources must be guided by an underlying framework for structuring word meaning and generating concepts, which satisfies both onto logical considerations as well as the need to capture linguistic generalizations. The model presented here is a proposal toward this goal.
Introduction
The development of formal frameworks for computational lexical semantics should respond to two needs: capturing the richness of language as revealed in both meaning variation and systematic polysemy, and providing a viable and testable model for building large-scale lexical resources for natural language processing.
In this paper, we address these topics from the dual perspective of theory and applications. The theoretical aspect motivates a generative framework for structuring and generating concepts. The practical aspect focuses on the implementation of the GL-based framework within the EU-sponsored SIMPLE project, which involves the development of harmonized large-scale semantic lexicons (10,000 word senses) for twelve different languages.
Utterers create meanings by using words in context.
Hearers create interpretations.
Patrick Hanks, SPARKLE Workshop, Pisa, January 1999.
This quote from Patrick Hanks reflects very closely the spirit of this volume that tackles the relation between word meaning and human linguistic creativity. We are interested in pursuing the view that words are rich repositories of semantic information that people use to talk about the world in a potentially infinite number of ways. Our goal is to tackle the essence of words insofar as they provide a window onto human cognition and the compositional nature of thought. It is thus imperative that a theory of language addresses how lexical items contribute to the peculiar human ability that goes under the label of “linguistic creativity.”
It is undeniable that words have “meanings” that go above and beyond the scope of linguistic research: They often carry the weight of a person's own experience. We are not aiming at exploring the unexplorable, but in proving that, given an appropriate set of methodological tools, a formal modeling of word meaning and linguistic creativity can be achieved. Linguistic creativity is a “generative” ability to extend the expressive possibilities of language in a potentially infinite number of ways. From the perspective of the lexicon (i.e., word meaning), it is the ability to give new meanings to words beyond their literal use.
As such, the overall task follows the strategy of contemporary generative syntax, which has achieved a basic understanding of how speakers produce and understand novel utterances and has brought simplicity to the great complexity underlying the syntactic structure of sentences.
The contributions in this section are centered around a set of common themes:
developing a theoretical vocabulary sufficiently rich to understand how word meanings compose;
developing and motivating frameworks for lexical semantics with explanatory force;
analyzing cross-linguistic data for achieving linguistically independent models.
Although each contribution approaches the problems from different angles and different data sets, they all highlight the richness of the information carried by words in context. The real challenge, as it emerges from the papers, is whether it is possible to establish a clear boundary between people's words and people's worlds or experiences. The goal is to understand whether there is a level of representation that is independent of specific contextual variations, while accounting for the novel use of words in different contexts.
The authors reach different conclusions: some reject structured representations of lexical information, others show that it is precisely in the structuring of the lexicon that we can achieve an understanding of how meaning changes in context.
The first paper by James Pustejovsky presents recent developments in GL, focusing on the role of qualia structure as a syntax for creating new concepts. The paper addresses fundamental questions on the well-formedness of concepts, the combinatorial possibilities within a generative mental lexicon, and how these principles are motivated by linguistic evidence.
The contribution by Jacques Jayez focuses on the meaning variations of the French verbs “suggerer” (suggest) and “attendre” (wait).
The idea that semantic representations are underspecified, that is more abstract than the specific interpretations obtained in various contexts, is by now current in lexical semantics. However, the way in which underspecified representations give rise to more precise interpretations in particular contexts is not always clear. On one view, context provides missing information, for instance because it contains salient entities that can be referred to. I consider here the symmetric dependency, in which lexical elements impose certain semantic profiles to the contexts they fit in. I show that, although they are highly underspecified, those profiles cannot be reduced to a general semantic frame, unlike what is proposed in Pustejovsky's Generative Lexicon, and that their semantic adaptability reflects the highly abstract and similarity-based character (vagueness) of the predicates that help to define them.
Introduction
Recent work about the relation between lexical items, context, and interpretation has highlighted two notions of underspecification (see van Deemter and Peters, 1996 for various points of view). In some approaches, underspecification amounts to code ambiguities in an efficient way, to avoid carrying a set of alternatives during the interpretation process (Reyle, 1995). In the domain of the lexicon, underspecification sometimes takes the form of information enrichment.
Metaphor, and the distinction between the figurative and the literal uses of language, have puzzled philosophers and linguists at least since Aristotle. The puzzle can be stated in the following, rough, form: How can words in certain configurations mean something different from what they mean in their literal use, prescribed by the rules of the language, and at the same time convey significant insights into what we, in a given context, take as parts of reality? In order to appreciate the force of the question we must separate the metaphorical meanings from the new literal meanings that an individual, or a group, might introduce into a language, such as parenting, or critiquing. Such innovations are extensions of literal language, not metaphors. Metaphors rest on rules of language, but also violate them. They do not describe reality directly. Thus, the true/false dichotomy does not apply to them without qualifications. An adequate theory of metaphor should explain this unique position of metaphoric meaning. To expand on this a little, this essay proposes that a theory of metaphoric meaning should account for the following list of facts or intuitions.
(i) Metaphors give us new meanings and a deepened understanding of the objects of our descriptions or reasoning.
(ii) Metaphors can have aesthetic value.
(iii) At some stage, a subjective element enters into the interpretation of metaphors. This element is creative insofar as it goes beyond what is given by the rules of language but it presupposes and rests on such rules.
I would like to pose a set of fundamental questions regarding the constraints we can place on the structure of our concepts, particularly as revealed through language. I will outline a methodology for the construction of ontological types based on the dual concerns of capturing linguistic generalizations and satisfying metaphysical considerations. I discuss what “kinds of things” there are, as reflected in the models of semantics we adopt for our linguistic theories. I argue that the flat and relatively homogeneous typing models coming out of classic Montague Grammar are grossly inadequate to the task of modeling and describing language and its meaning. I outline aspects of a semantic theory (Generative Lexicon) employing a ranking of types. I distinguish first between natural (simple) types and functional types, and then motivate the use of complex types (dot objects) to model objects with multiple and interdependent denotations. This approach will be called the Principle of Type Ordering. I will explore what the top lattice structures are within this model, and how these constructions relate to more classic issues in syntactic mapping from meaning.
Language and Category Formation
Since the early days of artificial intelligence, researchers have struggled to find a satisfactory definition for category or concept, one which both meets formal demands on soundness and completeness, and practical demands on relevance to real-world tasks of classification. One goal is usually sacrificed in the hope of achieving the other, where the results are muddled with good intentions but poor methodology.
In this paper, we present a novel approach to partial parsing that produces dependency
links between words of a sentence. The partial parser called a lightweight dependency analyzer
uses information encoded in supertags and hence can produce constituency-based as well
as dependency-based analyses. The lightweight dependency analyzer has been used for text
chunking, including noun and verb group chunking. We also present a proposal for a general
framework for parser evaluation that is applicable for evaluating both constituency-based and
dependency-based, partial and complete parsers. The performance results of the lightweight
dependency analyzer on Wall Street Journal and Brown corpus using the proposed evaluation
metrics are discussed.
An algorithmic approach to the semantic interpretation of deverbal nominalizations found
in encyclopedic texts, such as support, publication and control, is described. Interpreting these
nominalizations is crucial because they are quite common in encyclopedic texts, hence a great
deal of information is represented within them. Interpretation involves three tasks: deciding
whether the nominalization is being used in a verbal or non-verbal sense; disambiguating the
nominalized verb when a verbal sense is used; and determining the fillers of the thematic roles
of the verbal concept or predicate of the nominalization. A verbal sense can be recognized
by the presence of modifiers that represent the arguments of the verbal concept. It is these
same modifiers which provide the semantic clues to disambiguate the nominalized verb. In
the absence of explicit modifiers, heuristics are used to discriminate between verbal and
non-verbal senses. A correspondence between verbs and their nominalizations is exploited so
that only a small amount of additional knowledge is needed to handle the nominal form.
These methods are tested in the domain of encyclopedic texts and the results are shown.
In this paper, we propose a new ambiguity representation scheme; Structure Preference
Relation (SPR), which consists of useful quantitative distribution information for ambiguous
structures. Two automatic acquisition algorithms, the first acquired from a treebank, and
the second acquired from raw texts, are introduced, and some experimental results which
prove the availability of the algorithms are also given. Finally, we introduce some SPR
applications in linguistics and natural language processing, such as preference-based parsing
and the discovery of representative ambiguous structures, and propose some future research
directions.
Prepositional Phrases (PP) perform a variety of syntactic functions in a conventional sentence,
and cause severe problems to computer systems that automatically analyse the sentential syntax.
A major issue in this area has been the automatic determination of the syntactic functions
of PPs. Most work published so far makes use of the probabilistic approach, and attach PPs
to either the antecedent noun or verb phrase. Due to the natural limitation of the probabilistic
approach, it is important to evaluate the linguistic behaviour of prepositional phrases and
propose qualitative solutions to the problem. In this article, I first provide a detailed account
of statistics regarding the frequency of use for (i) types of prepositions, (ii) syntactic categories
as realisations of prepositional complements, and (iii) the syntactic functions of prepositional
phrases. Statistics reported here all derive from a representative corpus of contemporary
British English. I then describe a set of rules that has been implemented in order to label
PPs automatically for their syntactic functions. I finally report on the coverage of these rules
empirically observed in an experiment which involved a set of naturally occurring PPs as test
data.
This article evaluates the efficiency of the LiLFeS abstract machine by performing parsing
tasks with the LinGO English resource grammar. The instruction set of the abstract machine
is optimized for efficient processing of definite clause programs and typed feature structures.
LiLFeS also supports various tools required for efficient parsing (e.g. efficient copying, a
built-in CFG parser) and the constructions of standard Prolog (e.g. cut, assertions, negation
as failure). Several parsers and large-scale grammars, including the LinGO grammar, have
been implemented in or ported to LiLFeS. Precise empirical results with the LinGO grammar
are provided to allow comparison with other systems. The experimental results demonstrate
the efficiency of the LiLFeS abstract machine.