9.0 Questions and Answers
(1) Why do you think both linguists and philosophers find the typology of semantic entities interesting?
The goal of this chapter is to discuss which basic semantic entities should be included in formal semantic theory, and on which principles we should include or exclude them. While some framework-internal assumptions in formal semantics are hotly debated, the matter of how many and which basic entities a theory assumes is largely treated as a matter of personal taste or convenience. It will be useful for linguists to have a common standard for introducing and constraining basic entities in their frameworks, so they can better understand and model natural language semantics. Philosophers will find this discussion useful for the same reason, and also to learn more about the possible formalism that can model natural language, and what consequences these conclusions have for metaphysics or the philosophy of science.
(2) What recent developments in linguistics and philosophy do you think are most exciting in thinking about the phenomenon in question?
The issue of which entities are basic has been building momentum for decades, as some semanticists have presented empirical motivations for proliferating types (to include e.g. degrees and events) while others have presented empirical motivations for reducing or combining them (e.g. to situations). But the issue has really gained a toehold in recent work arguing that specific languages can differ in which basic entities they employ; in particular, that languages seem to differ in whether their semantic representation must include degrees (Bochnak Reference Bochnak2015b).
(3) What do you consider to be the key ingredients in adequately analyzing the phenomenon in question?
This chapter presents the most prominent empirical motivations for proliferating types (to include e.g. events) and reducing them (to combine e.g. events and worlds). Effectively, arguments in favor of type-proliferation identify distinct morphology for quantifiers or proforms ranging over different types, and that failing to include a distinct type or entity could predict incorrect truth conditions. Arguments in favor of type-collapsing emphasize languages or phenomena in which morphology seems to be shared across putative type domains, and constructions for which the meaning of one putative type of entity is intrinsically tied to the meaning of another, related putative type of entity. I end by concluding that the arguments for type-proliferation are slightly more compelling in terms of empirical coverage and diachronic considerations.
(4) What do you consider to be the outstanding questions pertaining to the phenomenon in question?
I conclude this chapter by arguing that type-collapsing arguments are each individually compelling but inconsistent as an aggregate. It remains to be seen whether a semantic theory with minimal basic types could satisfy the wide variety of distinct empirical constraints placed on it by type-collapsing arguments across different empirical phenomena.
9.1 Introduction
The goal of this chapter is to discuss which basic semantic entities we should include in our formal semantic ontology, and on which principles we should include them (cf. Bach Reference Bach1986b). The vast majority of formal theories employ individuals as a basic type and primitive; these theories model quantification over, modification of, and reference to individuals in the metalanguage. But many theories include additional types or entities, including possible worlds, but also less common ones like vectors. Some papers have argued that types should be constrained or reduced; others that they should be proliferated. I’ll present some representative arguments on both sides and suggest a path forward in evaluating them against one another.
Standard Montagovian semantics employs individuals as a basic type, but in practice, some theorists embrace many more basic types. Champollion (Reference Champollion2010) uses at least five (differentiating between degrees, numbers and intervals) in just an extensional semantics; Bittner (Reference Bittner2003) uses seven (differentiating between animate and inanimate entities); and Landman (Reference Landman2006) assumes nine basic entities, differentiating between kinds, events, and event-kinds. In contrast, others have taken for granted that the adoption of different types should be constrained by Ockham’s Razor, and have posed and taken up the challenge of eliminating as many as mathematically possible. Thomason (Reference Thomason1980) argued that possible worlds aren’t necessary to model propositional attitudes, and Partee (Reference Partee2009) and Keenan (Reference Keenan and Bianchi2015, Reference Keenan2018) propose doing away with individuals as a basic semantic type.
As a semanticist, I take semantic adequacy to be the primary goal of a formal semantic theory. But a metasemantic concern about modeling the right entities – or at least the right quantity of entities – seems like a plausible secondary goal. This is highlighted by the reductio ad absurdum put forth in Ritchie (Reference Ritchie2016), which depends on the principle that equally semantically adequate treatments should have equal voice in determining the ontological commitments carried by natural language theory. If there are indeterminately many – or even multiple – semantic formalisms that postulate a distinct set of entities, we cannot hope for a principled relationship between semantic theory and ontology.
There are, as far as I can tell, three logically possible positions to take regarding basic entities in a semantic model:Footnote 1
(1) Type Reductionism
Type Reductionalism assumes no basic types, or no nonfunctional types. This position is best instantiated historically by Henkin (Reference Henkin1963), and more recently by Partee (Reference Partee2009) and Keenan (Reference Keenan and Bianchi2015, Reference Keenan2018), whose foundational type is a set. Following Gallin’s (Reference Gallin1975) notation, a Type Reductionalist account is Ty-0, because it has no basic types.
(2) Type Ersatzism
A Type Ersatzist assumes one basic nonfunctional type – usually entities in general – and proposes to treat all natural language phenomena using this type. In Gallin’s notation, a Type Ersatzist account is Ty-1. Church (Reference Church1932) exemplifies Type Ersatzism, as does Carlson (Reference Carlson1977) for times and Klein (Reference Klein1980, Reference Klein1982) for degrees.
(3) Type Proliferationalism
A Type Proliferationalism account is a Ty-n account, for
entities that pass some linguistic tests, e.g. that they have dedicated pronouns. (This test, a modern adaptation of Quine’s Reference Quine1948 generalization ‘to be is to be the value of a variable,’ is explicitly assumed as one of a few tests for entityhood in Schlenker Reference Schlenker2006.)
The difference between Type Ersatzism and Type Proliferationalism is a subtle one, reminiscent of selectional restrictions on verbs and adjectives, illustrated in (1).
(1)
a. *Jane met. number restriction b. *She’s a bachelor. gender restriction c. *The plant tried to grow. animacy restriction d. *Jane bounced the cloud. (solid-)state restriction e. *The piece of paper is fat. (three-)dimension restriction
The predicates are picky about what sorts of arguments they take. Example (1a) is ungrammatical because intransitive meet requires that its subject denote a plurality; (1b) is ungrammatical because the adjective bachelor can only be predicated of males. It’s standardly assumed that these are all predicates of individuals; they can be modeled as denoting partial functions from the subclass of entities that satisfy the relevant selectional restriction. This is something like a Type Ersatzism approach; it treats different kinds of entity within the same type. In contrast, a Type Proliferationalism approach might assign different types to singular and plural entities, male and female entities, etc.
Example (2) illustrates this difference between Type Ersatzism and Type Flexibility for the adjectival modifier very (and some contextually valued standard s); the former involves individuals, type e, and requires a partial function to model the fact that very must modify something gradable (*very student); the latter involves entities other than individuals – degrees of type d – and thus does not require a partial function.
(2)
a. Type Ersatzism: 
b. Type Flexibility: 
I’ll end this section with some ground-clearing. First, I restrict myself to the discussion of basic types – not complex ones – because I assume that the latter come for free, in every formal model, via type formation. This is despite the interesting empirical arguments that some complex types (e.g. quantifiers) are unavailable in some languages (e.g. St’át’imcets, which I take to be a separate discussion, Matthewson Reference Matthewson, Newmeyer and Preston2014), and also despite the fact that the same arguments used to differentiate between basic and complex entities (Link Reference Link, Bäuerle, Schwarze and von Stechow1983) might also be useful in differentiating between two basic entities.
And while considerations of economy and parsimony are certainly relevant when it comes to evaluating theories, these considerations presume empirical equivalence (i.e. Can semantic theories that presume different basic semantic entities account for the same natural language data?). And since it is this empirical equivalence that I will be discussing, I will treat considerations of economy and parsimony as secondary to the discussion at hand.Footnote 2
When I discuss the question of how many basic entities or types there are, I view the issue as regarding the semanticist’s toolbox, rather than a given speaker’s i-language. In particular, if some language were to have what we decide to be irrefutable evidence that the semantic modeling of that language requires e.g. degrees as a basic entity, then this is sufficient evidence that there are degrees in the semantic ontology, but not sufficient evidence that there are degrees in every human being’s i-language. This is compatible with the possibility that some languages do not require degrees in their semantic modeling, or even should not be modeled using degrees, as has been argued for Washo (Bochnak Reference Bochnak2015b) and some other languages.
Finally, there is one diagnostic for basic semantic entities that I will not use: I will not assume that the presence of a referring expression over a particular sort of thing privileges that thing as an entity (cf. Schlenker Reference Schlenker2006). More broadly, I will focus on linguistic tests that don’t involve categories of words whose inventory is potentially infinite; this is the standard lexical vs. functional distinction. It seems as though we can coin a referring expression for anything, really, at least in particular contexts – e.g. Twin Earth (Putnam Reference Putnam1973) – and so it is a trivial test.
9.2 Metasemantic Arguments
Before I turn to empirical arguments made in favor of proliferating or collapsing types, I’ll briefly present two instances of semanticists directly addressing the question of how many basic entities the ontology should include, and discuss why their claims don’t seem to completely settle the issue.
In his 2006 paper, Philippe Schlenker observes that “Reference to individuals, times and worlds is uniformly effected through generalized quantifiers, definite descriptions, and pronouns” (Reference Schlenker2006: 504). He interpreted this observation as evidence that there is a single type,
, ranging over an 〈individual, time, world〉 triple.
As argued above, there is no limit on the sort of entity that can be referred to with a name or definite description, given the innovativeness of natural language. And, as we’ll see below, it seems clear that there are types of entities other than individuals, times and worlds (namely degrees, events, and/or situations) that are associated with quantifiers and proforms as well. So if Schlenker’s criteria are necessary and sufficient for differentiated basic entities – and he provides no arguments that they are, especially as he considers them members of a triple – they lead to a more inclusive ontology than he advocates for.
In her 2003 paper, Maria Bittner adopted a similar perspective. She advocates for a Ty-7 formal semantics, including worlds, times, places, events, states, animate individuals, and inanimate individuals as basic entities. What’s more, she meaningfully differentiates between basic types – like states and events – and derived or complex types, like processes and habits, which she takes to denote partial functions. Bittner’s criterion for a basic type (see also Bittner Reference Bittner2006, Reference Bittner2014) is whether entities of that sort participate in anaphora in a language, and thereby whether a dynamic model of that language requires a separate discourse stack for that type.Footnote 3
However, Bittner herself acknowledges that discourse reference can be to complex types (or, rather, to the basic components of a complex type): “In general, simple episodic discourse involves only simple types of drefs – the seven basic sorts (possibilities
, agentive entities α, non-agentive entities β, events ε, states σ, times τ, and places π or else simple functions (mostly processes ε …)” (Reference Bittner2003: 645). She specifies, of discourse referents to processes, “Formally, these are ε-dependencies, mapping each stage of the process, except the last, to the next stage” (Reference Bittner2003: 641).
So perhaps the ability of a proform or anaphor in a language to refer to an entity is a necessary but not sufficient condition for its being a basic type, as it could be anaphoric on a subcomponent of a complex entity. The result is a very robust and comprehensive crosslinguistic ontology, but one with no clear, reliable way of differentiating between basic and complex or derived types.
In the next section, I present several prominent empirical arguments that natural language semantics needs to include a given entity as basic.
9.3 Type-Proliferation Arguments
Across the formal semantics literature, there have been a number of implicit and explicit arguments for a number of different non-individual entities. I have summarized the most prominent ones in Table 9.1.Footnote 4
Table 9.1. Basic entities: the usual suspects
| Entity | Type | Conventional variables | Origin(s) |
|---|---|---|---|
| Individuals | e | x, y | Montague (Reference Montague and Visentini1970, Reference Montague, Hintikka, Moravcsik and Suppes1973) |
| Possible worlds | s | w | Kripke (Reference Kripke1959) |
| Events | v | e | Davidson (Reference Davidson and Rescher1967) |
| Times | i | t | Partee (Reference Partee1973, Reference Partee1984) |
| Degrees | d | d | Cresswell (Reference Cresswell and Partee1976) |
| Kinds | k | k | Carlson (Reference Carlson1977) |
| Situations | s | s | Barwise (Reference Barwise1981); Kratzer (Reference Kratzer1989) |
| Vectors | v | u, v | Zwarts (Reference Zwarts1997) |
The empirical arguments for adding different types or entities to the semantic ontology have been very similar, and they fall into two broad categories. First, there is morphological evidence: arguments that a given language has functional items that seem dedicated to particular, nonindividual domains. Second, there is semantic evidence in the form of arguments from semantic adequacy: phenomena that seem to be able to be modeled only in a formalism that includes nonindividuals. For several familiar nonindividual entities, I will review some prominent arguments in favor of their inclusion in the semantic ontology.
9.3.1 Possible Worlds
The original Montagovian semantics (Montague Reference Montague and Visentini1970, Reference Montague, Hintikka, Moravcsik and Suppes1973) manipulated only individuals as a basic semantic entity, but included possible worlds as a restriction on the interpretation function. Explicit world variables were quickly added to Montague-inspired formalisms in part based on the empirical arguments from modal semantics in Hintikka (Reference Hintikka1957) and Kripke (Reference Kripke1959): they allowed for a set-theoretic representation of entailment, and a treatment of weak and strong modals as dual quantifiers, in parallel with individual quantifiers.
Effectively, in arguing that possible worlds should be represented as semantic variables, these papers instantiate both of the types of arguments I outlined above. The observation that languages lexicalize (strong and weak) modals like must and might differently from (universal and existential) individual quantifiers like all and some constitutes a morphological argument for adding possible worlds to the semantic ontology. And the claim that relationships between propositions can be best modeled using set theory – and thereby that propositions must be associated with sets – constitutes a semantic adequacy argument.
Interestingly, there appears to be no language that does not have modals, although languages do seem to differ with respect to which elements they unite lexically (i.e. quantificational force; flavor; and evidential base; Matthewson Reference Matthewson2013). This lack of crosslinguistic variation seems in concord with the relatively uncontroversial status possible worlds enjoy in formal semantic theory (in contrast with many other entities in Table 9.1; although see e.g. Kaplan & Montague Reference Kaplan and Montague1960; Fine Reference Fine2012 for arguments that possible worlds shouldn’t be used to model certain phenomena). But the question of how to deal with crosslinguistic variation, in cases where we do find it, will turn out to be a central one.
9.3.2 Times
In a series of papers, Barbara Partee (Reference Partee1973, Reference Partee1984) argued that tense markers should be thought of as temporal proforms rather than operators. While these accounts both include times in their semantic models, they differ in whether they require the modeling of times as a basic entity. A formalism in which tense markers denote temporal operators is compatible with a formalism in which times are restrictions on the interpretation function, an analogue to Montague’s treatment of modals qua possible world operators.
Partee’s arguments that tense markers are temporal proforms amount to arguments of parity. They take for granted that there are individual pronouns, and that individual pronouns are individual pronouns by virtue of their semantic behavior. If tense markers demonstrate identical semantic behavior, then they, too, must be proforms (only over times, not individuals). Stone (Reference Stone1997) made parallel arguments for the words will and would, which he assumed were modal or possible world proforms. The key data are reproduced below.
(3)
nonlinguistic antecedents a. [at a bar] She left me. individual b. [on a road trip] I didn’t turn off the stove. temporal c. [at a stereo store] My neighbors would kill me. modal
(4)
definite antecedents a. Sam is married. He has three children. individual b. Sheila had a party last Friday. Sam got drunk. temporal c. The company would face bankruptcy if the merger goes through. modal
(5)
(6)
(7)
donkey-anaphoric use a. If Pedro owns a donkey, he pets it. individual b. If Mary telephoned on a Friday, it was (always) Peter who answered. temporal c. If a submarine cannot self-destruct if an enemy captures it, the enemy will learn its secrets. modal
These constitute morphological arguments for times and possible worlds as distinct entities, and they have been influential. But it’s worthwhile noting that there are other anaphoric elements that display the same behavior which we do not tend to associate with basic types, including VP ellipsis; propositional anaphors as in Jane believes that too; and adjectival such (Landman Reference Landman2006; King & Lewis Reference King, Lewis and Zalta2018). This is a good illustration of one of the limits of morphological arguments for proliferation, at least for those semanticists who don’t consider verbal properties and propositions to be basic entities. And Sharvit (Reference Sharvit2013) argues that whether tenses are proforms or quantifiers is a crosslinguistic parameter (although, arguably, both must be modeled in a formalism in which times are basic entities).
Many languages have been observed to lack tense marking of any kind: so-called ‘tenseless languages’. Interestingly, these languages differ in how their tenseless clauses can be interpreted, leading to a series of proposals that constitute semantic arguments for times as a distinct entity, as well as crosslinguistic variation with respect to whether or not a language should be modeled using times.
The simple matrix clause in (8), from Lillooet (a Salish language, Matthewson Reference Matthewson2006) lacks any sort of overt tense marker.
(8)
Táytkan hungry.1sgS ‘I am/was hungry.’
In her analysis of such sentences, Matthewson (Reference Matthewson2006) observes that while they are underspecified with respect to a present- and past-oriented interpretation, they can never be future-oriented (i.e. can never mean ‘I will be hungry’). She thus concludes that they must include a (morphologically null) nonfuture marker, effectively a covert tense operator.
However, in other tenseless languages, versions of (8) behave slightly differently, leading other researchers to offer distinct analyses of tenseless sentences (Bittner Reference Bittner2005; Lin Reference Lin2005; Tonhauser Reference Tonhauser2011; Mucha Reference Mucha2013). While none of these analyses assume that times are explicitly introduced in the language, they all assume that it is nevertheless manipulated in the language’s compositional semantics, in some way.
For instance, while Hausa sentences lack overt tense markers, they do include (overt) aspectual operators. And the truth conditions of sentences that include e.g. prospective aspect do involve the manipulation of times. Mucha (Reference Mucha2013: 203) argues that the prospective aspect should be given the analysis in (9), which involves the restriction of an event’s runtime
to precede some temporal argument t.
(9)

for the runtime τ(e) of an event e
We can conclude from this body of work that times – or at least some subtype of strictly ordered entities, or some entity (like events or situations) from which runtimes can be calculated – are necessary for modeling languages universally, despite the fact that not all languages have dedicate time-introducing pronouns. Once again, we have a situation in which there seems to be no crosslinguistic variation with respect to whether languages need to be modeled using times (the semantic arguments), although we do, for the first time, see crosslinguistic variation with whether reference to times is morphologically encoded (the morphological arguments).
I’ll end this section by highlighting the significant empirical and theoretical overlap between times and other temporal entities. While the arguments discussed here explicitly advocate for times, others have used similar arguments to instead advocate for the existence of situations (to be discussed in more detail in Section 9.4.1), events (Section 9.4.2), or states (Parsons Reference Parsons, Achille Varzi and Pianesi2000; Altshuler Reference Altshuler2016; Stojnić & Altshuler Reference Stojnić and Altshuler2021), cf. Katz (Reference Katz, Lang, Maienborn and Fabricius-Hansen2003). Insofaras they advocate for treating these or related phenomena using an entity other than individuals, the arguments in these other works are in some loose sense parallel to the arguments detailed here, but I have of course not done them justice. Section 9.4 discusses reference to events, situations, etc. in type-collapsing arguments.
9.3.3 Degrees
Based on a suggestion in Cresswell (Reference Cresswell and Partee1976), a common and intuitive way of semantically differentiating between nongradable and gradable adjectives is in their arity; specifically, that nongradable adjectives denote individual properties, while gradable adjectives denote relations between degrees and individuals.
(10)
a. 
b. 
Ewan Klein (Reference Klein1980, Reference Klein1982) argued that the phenomena Cresswell focused on – specifically, positive constructions like Jane is tall and comparatives like Jane is taller than Bill – can be dealt with instead in a semantics with only individuals as basic entities. In his analysis, gradable adjective constructions manipulated comparison classes of individuals, partitioned according to metasemantic principles that assure that e.g. a is not both bigger and smaller than b.
However, since Klein’s proposal (and refinements offered in Neeleman et al. Reference Neeleman, van de Koot and Doetjes2004; Bale Reference Bale2011), there have been several adjectival phenomena that, it’s been argued, cannot be properly characterized using a degree-free semantics. Kennedy (Reference Kennedy1999) argued that a Kleinian degree-free semantics can’t differentiate between antonyms like tall and short. He also argued that it can’t account for the semantic contribution of measure phrases (MPs) like two feet, especially in their function as comparative differentials in e.g. Jane is two feet taller than Bill (Schwarzschild Reference Schwarzschild2005). And in a particularly influential paper, Kennedy and McNally (Reference Kennedy and McNally2005) argued that the different types of intervals degrees can form – closed, open, partially closed – are needed to predict the behavior of different subtypes of gradable adjectives, i.e. the difference between relative, absolute, and total/partial adjectives.
Finally, English has different comparative strategies, and it’s been argued that some of them involve the comparison of degrees, while some of them involve the comparision of individuals. This argument, like the others, is a semantic adequacy argument: if there are semantic differences between two comparative strategies that are predicted by modeling one with degrees and one without, then we can conclude that degrees must be in the semantic ontology.
Three comparative strategies are listed below, although there are others, even in English (Stassen Reference Stassen1985).
(11)
a. Jane exceeds Bill in height. ‘exceed’ comparative b. Jane is tall compared to Bill. implicit degree comparative c. Jane is taller than Bill. explicit degree comparative
Explicit degree comparatives, formed with the English -er or more, exhibit subtly different semantic behavior than other comparative strategies (Beck et al. Reference Beck, Oda and Sugisaki2004, Reference Beck, Krasikova, Fleischer, Gergel, Hofstetter, Savelsberg, Vanderelst and Villalta2009; Kennedy Reference Kennedy2005). They (i) don’t require a ‘crisp judgment’ scenario, in which the difference in values is significant; (ii) are nonevaluative when formed with positive relative adjectives like tall; (iii) can be modified by a differential; and (iv) can be formed with absolute adjectives. These differences are illustrated in (12)–(15).
(12)
Context: Jane is 1 mm taller than Bill. a. Jane is taller than Bill. explicit b. #Jane is tall compared to Bill. implicit
(13)
a. Jane is taller than Bill.
Jane is tall.explicit b. Jane is tall compared to Bill.
Jane is tall.implicit
(14)
a. Jane is six inches taller than Bill. explicit b. Jane is (#six inches) tall (#six inches) compared to Bill. implicit
(15)
a. This pole is more bent than that one. explicit b. ?This pole is bent compared to that one. implicit
A broad consensus is that the best way of accounting for these semantic differences is by treating the explicit comparative -er, more as denoting a degree quantifier, or a relation between a set of degrees corresponding to the matrix argument and a set of degrees corresponding to the embedded argument (Beck et al. Reference Beck, Oda and Sugisaki2004, Reference Beck, Krasikova, Fleischer, Gergel, Hofstetter, Savelsberg, Vanderelst and Villalta2009; Kennedy Reference Kennedy2005). For example, a Kleinian account of comparative constructions – in which the embedded argument forms a comparison class restrictor for evaluating the matrix argument – predicts the ‘crisp judgment’ requirement we see in (12b), and cannot account for the acceptability of the explicit comparative in (12a). There is a directly parallel argument for the semantics of various equative strategies (Rett Reference Rett and Hallman2020).
Even more compelling, there seems to be variation with respect to whether a language’s comparison strategies must be modeled using degrees. A series of languages have been identified as ‘degree-free’ languages, meaning they lack any sort of construction that has been argued to require degrees to be adequately semantically modeled: Motu (Beck et al. Reference Beck, Krasikova, Fleischer, Gergel, Hofstetter, Savelsberg, Vanderelst and Villalta2009); Fijian (Pearson Reference Pearson2010); Washo (Bochnak Reference Bochnak and Pasquereau2015a, Reference Bochnak2015b; Beltrama & Bochnak Reference Beltrama and Bochnak2015); Navajo (Bochnak & Bogal-Allbritten Reference Bochnak, Bogal-Allbritten, Bochnak and Matthewson2015);Footnote 5 Walpiri (Bowler Reference Bowler, Grubic and Mucha2016). These languages do not have explicit comparatives (or superlatives), but they also do not have measure phrases or measure phrase constructions (e.g. five feet tall) or degree modifiers like very.
In conclusion, there have been attempts to model English adjectival constructions without the use of degrees as basic semantic entities (Klein Reference Klein1980, Reference Klein1982; Neeleman et al. Reference Neeleman, van de Koot and Doetjes2004). These accounts have been argued to fail on empirical grounds, because they cannot account for the semantic behavior of many constructions, including explicit comparatives. Furthermore, there seem to be languages for which an entire class of constructions are not available, and this class is very naturally characterized as constructions that must be modeled using degrees.
9.3.4 Interim Summary
There are, roughly, two types of arguments made in favor of adding basic entities to the semantic ontology: morphological and semantic ones. The morphological arguments pertain to the language’s inventory of functional morphemes. They assume (generally implicitly, as in Partee Reference Partee1973, Reference Partee1984), that a language differentiates between entity x and entity y if: (i) it lexicalizes different proforms for x and y; (ii) it lexicalizes different modifiers for x and y; and (iii) it lexicalizes different quantifiers for x and y. These arguments have all been extended – implicitly or explicitly – to possible worlds, times, and degrees.
The other arguments for type-proliferation pertain to semantic adequacy: in the case of times, there was the argument that, even in the absence of a dedicated temporal proform or quantifier in Hausa, aspectual markers need to reference times in order to impose strict enough truth conditions. And in the case of degrees, there was the argument that a semantic model that does not include degrees can adequately model implicit comparatives in English, but not explicit comparatives.
In the case of degrees – but not, as far as I know, any other putative basic entities – there has also been a robust research project arguing that languages can differ with respect to whether they must be modeled using degrees. Evidence for this distinction comes in the different inventory of constructions or morphological strategies for expressing certain situations, so they constitute a sort of comparative morphological argument.
These criteria – imposed by the morphological and semantic arguments – are arguably not necessary conditions, as we can imagine a language with a relatively impoverished lexicon and accidental homophony or syncretism in e.g. its proform inventory, but not its quantifier inventory. This sort of thing happens all the time, and I’ll return to discuss it in more detail in Section 9.5.
They are also arguably not sufficient conditions for basic entities, which is a very unsatisfying situation. Wh-phrases could be construed as either proforms or quantifiers, depending on one’s semantic theory, but they do not perfectly track what is generally thought of as being a plausible basic entitity. On the one hand, English and several other languages co-opt the same wh-phrase to range over degrees and manners (how), prompting some to argue that these are the same sort of entity (see Section 9.4.3). On the other hand, English has a dedicated wh-phrase where for locations and vectors (it has a locative and a directional interpretation) but two distinct ones, why and how come, for reasons.
However, it’s worthwhile noting the consistency of these morphological and semantic arguments across different, unrelated theoretical projects and empirical phenomena. The idea that e.g. distinct proforms belie distinct basic entities seems to be a common and thereby intuitive assumption.
Finally, the introduction of the possibility that languages can differ with respect to their basic entity inventory predicts the existence of a robust and theoretically attractive universal typology, reminiscent of Greenberg’s (Reference Greenberg and Greenberg1963) Universals (e.g. “If a language is exclusively suffixing, it is postpositional; if it is exclusively prefixing, it is prepositional”). The claim that arises from this discussion of type-proliferation arguments is that if a language has morphemes and constructions that must be semantically modeled using degrees, it also has morphemes and constructions that must be modeled with individuals, possible worlds, and times, but not vice versa. However, there is arguably more work to do in replicating the detailed study of degrees in other domains, including times.
9.4 Type-Collapsing Arguments
In contrast to those who have advocated for adding basic entities to the semantic ontology are those who have argued that basic entities should be collapsed or eliminated. These arguments, too, come in two forms: those from morphological cues and those from semantic convergence.
9.4.1 Situations
In a several prominent papers written in the 1980s, it was claimed that the putative basic entities possible worlds, times, and locations, were collectively too blunt to model natural language semantics (Barwise Reference Barwise1981; Barwise & Perry Reference Barwise and Perry1983; Kratzer Reference Kratzer1989). These papers highlighted a number of constructions that seemed to restrict all of these things, dependently. Since these modal/temporal/spatial restrictions covary in a predictable way, they argued, a semantics that modeled them independently can’t be restrictive enough.
There are many sentences that can illustrate this point, I focus on one (from Cresswell Reference Cresswell1990; Kratzer Reference Kratzer and Zalta2007):
If, whenever it snowed, it had snowed much more than it actually did, the town plow would have removed the snow for us.
The point is that the putative anaphor would can’t range just over possible snowing events (cf. Cresswell Reference Cresswell1990; Stone Reference Stone1997), but it must range over possible snowing events indexed to a particular location (the town in question) and a particular time (the reference time, prior to the time of utterance). As Kratzer (Reference Kratzer and Zalta2007) explains,
we have to be able to consider for each actual snowfall s a set of counterfactual alternatives and compare the amount of snow in each of them to the actual amount of snow in s. This means that we have to be able to ‘go back’ to the actual snowfall situations after considering corresponding counterfactual situations.
The proposed solution involves situations: spatio-temporally specified partial worlds, or particulars. They are a basic entity that effectively collapses possible worlds, times, and locations into one. Incidentally, Kratzer (Reference Kratzer and Zalta2007) argues that situations can be used to define Davidsonian events, so the use of situations actually obviates four distinct putative basic types:
(17)


It’s been assumed that situations have corresponding proforms, although it’s not clear that these proforms have a single common morphology in English.Footnote 6 On the one hand, assuming situations as basic entities – instead of worlds, times, and locations – seems to explain the synonymy of when and if clauses (cf. Rothstein Reference Rothstein2009), as in (18).
(18)
a. Mary opens the door when(ever) the bell rings. b. Mary opens the door if the bell rings.
But if when and if do both range over situations in this assumed way, it would be nice to have an explanation of why English has more than one situation proform (or quantifier, as the case may be).Footnote 7 Such arguments are reminiscent of those in Lewis (Reference Lewis and Keenan1975) about domain-unspecified quantification, which highlights the challenge of empirically differentiating between a theory that assumes a single basic type with dedicated quantifiers or one that assumes a variety of different types with unspecified quantifiers (i.e. a Type Ersatz theory and a Type Flexible theory).
9.4.2 Individuals and Events (and Degrees)
A distinct but simultaneous semantic tradition has blurred the lines between individuals and events, beginning with Bach (Reference Bach1986a), who argued that individuals and events have the same mereology.
Bach was inspired by Link’s (Reference Link, Bäuerle, Schwarze and von Stechow1983) observation that plural count nouns and mass nouns behave similarly, and his consequent conclusion that both should be analyzed as denoting join semi-lattices (individuals whose plurality is modeled in their internal Boolean structure). Bach extended the empirical parallels to events, arguing that Link’s ‘cumulativity of reference’ property extends to atelic events – which are like mass nouns and plural count nouns – but not to telic events, which are like singular count nouns.
The empirical parallels between individuals and events have been strengthened further by the observation that some sentences are ambiguous between individual and event readings (Krifka Reference Krifka, Bartsch, van Benthem and van Emde Boas1989, Reference Krifka1990, Reference Krifka, Sag and Scabolcsi1992; Lasersohn, Reference Lasersohn1995).
(19)
a. Four thousand ships passed through the lock last year. b. The library lent out 23,000 books in 1987.
In (19a), the numeral four thousand could be counting distinct ships or events of passage; in (19b) the numeral could be counting distinct books or events of lending.
Krifka treats this polysemy as a homomorphism from “concrete entities to abstract entities” (Reference Krifka1990: 194), but assigns individuals (the concrete entities) and events (the abstract ones) different types. In his compositional analysis, the cardinality operator that allows for the semantic composition of the numeral and the NP – associating the nominal property with an argument corresponding to its cardinality – is polysemous between measuring individuals, events, or sets of events.
Some quantifiers seem to be domain-general in just this way, too; most in (20) can range over ships or events of passage.
Most ships passed through the lock at night.
A number of recent articles have examined this polysemy in much more detail, for e.g. French beaucoup (Doetjes Reference Doetjes2007; Nakanishi Reference Nakanishi2007; Burnett Reference Burnett2012).
There seems to be a directly parallel DP polysemy between individuals and degrees (and events). Some examples are illustrated in (21) (from Rett Reference Rett2014).
(21)
a. Four pizzas are vegetarian / is enough. b. Four feet of the plywood are warped / is more than Betty had asked for. c. French fries were eaten by the senators / is not enough, the senators will need protein.
In each example, the DP can denote a plural individual, in which case it is modified by an individual predicate (be vegetarian, be warped, be eaten) and triggers plural agreement on the predicate. But it can also denote a singular degree, in which case it is modified by a degree predicate (be enough, be more than) and triggers singular agreement on the predicate. Note that this is not a polysemy triggered by the numeral or measure phrase (cf. Landman Reference Landman, Coene and d’Hulst2000; Rothstein Reference Rothstein2009), as bare plurals are polysemous in the same way (21c).
In Rett (Reference Rett2014) I observe that the degree interpretations of DPs are restricted to dimensions of measurement that are monotonic on the part–whole structure of the plural individual, in the sense documented by Schwarzschild (Reference Schwarzschild2005). I conclude that the polysemy is the result of a homomorphism from individuals to degrees (not in the other direction, cf. Brasoveanu Reference Brasoveanu, Schardl, Walkow and Abdurrahman2009) that is bound by a restriction that requires the homomorphism be meaning preserving in a particular way. But it is unclear whether these homomorphisms are evidence in favor of Type Ersatzism (i.e. the semantic kinship suggests we should represent them using the same type) or Type Proliferationalism (i.e. the meaning-preservation constraint on the homomorphism suggests we need to represent the homomorphism using different types).
9.4.3 Degrees, Kinds, and Manners
There also seems to be a morphological kinship between degrees, kinds, and manners. Landman and Morzycki (Reference Landman, Morzycki, Antrim, Goodall, Schulte-Nafeh and Samiian2003) and Anderson (Reference Anderson2016) (see also Haspelmath & Buchholz Reference Haspelmath, Buchholz, van der Auwera and Baoill1998; Rett Reference Rett2013) note crosslinguistic evidence for the assimilation of degrees, kinds, and manners. This is shown in (22), in which the same morpheme as ranges over manners in the similative in (22a) and over degrees in the equative in (22b).
(22)
a. Jane danced as Maria danced. manner b. Jane is as tall as Maria is. degree
The polysemy is more expansive in Polish (23), in which the proform tak ranges over manners, kinds, or degrees.Footnote 8
(23)
a. On tańczyl tak. he danced thus ‘He danced like that.’ manner b. Taki pies uciekłwczoraj w nocy. such.masc.sg.nom dog.nom ran.away yesterday in.night ‘Such a dog ran away last night.’ kind c. tak wysoki such tall ‘that tall’ degree
To account for these data, Anderson (Reference Anderson2016) explicitly argues for a particular version of ‘enriched degrees’ (Grosu & Landman Reference Grosu and Landman1998), in which degrees are modeled as kinds of Davidsonian states, and manners are modeled as event-kinds. Formally, the approach is similar to situation semantics, in that it treats words like tak and as as involving abstraction over degree state-kinds.
9.4.4 Degrees and Vectors
Schwarzschild (Reference Schwarz2012) notices another crosslinguistic trend: in Hindi and Navajo, there is polysemy between spatial prepositions and comparative standard markers (equivalent to than in English comparatives). This is illustrated in (24) for Hindi.
(24)
a. anu raaj se lambii hai Anu Raj from tall.fem pres.sg ‘Anu is taller than Raj.’ b. anu us baRe kamre se niklii Anu that.obl big.obl room.obl from come.out.perf.fem ‘Anu came out of that big room.’
Schwarzschild (Reference Schwarz2012) uses data like these to argue for an analysis of an entity he calls a ‘directed segment,’ essentially a two-dimensional interval. Schwarzschild (Reference Schwarzschild and Snider2013) expands on this analysis, drawing on semantic arguments regarding comparison classes (Kennedy Reference Kennedy2007; Bale Reference Bale2011). In the analysis, segments
are shorthand for ordered quadruplets
, with
individuals representing endpoints;
a total ordering; and
a dimension of measurement (e.g. height). It is effectively a relational version of the triples Bartsch and Vennemann (Reference Bartsch and Vennemann1972) interpret degrees to represent. The result is a semantics for a comparative (25) formalized as (25a) and informally summarized in (25b).
(25)
Tom is taller than Susan. a. 
b. ‘There is a rising directed scale segment: it starts with Susan’s measurement on the scale, it ends with Tom’s measurement on the scale, and it is a segment of the height scale.’
9.4.5 Lattice vs. Interval Plurals
Finally, in Rett (Reference Rett2015), I argue that there are two different sorts of plural entities, intervals and lattices. Interval plurals (Schwarzschild & Wilkinson Reference Schwarzschild and Wilkinson2002; Dotlačil & Nouwen Reference Dotlačil and Nouwen2016) have strictly linearly ordered atomic members, e.g. degree scales, temporal intervals, and spatial vectors. Lattice plurals (Link Reference Link, Bäuerle, Schwarze and von Stechow1983) have atomic members that form a (semi-)lattice structure.
I argue that relations between interval plurals – like comparatives, temporal relations like before and after, and spatial prepositions – are interpreted with respect to the same general principle: the matrix argument is related to the most informative closed bound of the embedded argument. In contrast, relations between lattice plurals are interpreted with respect to the maximal plural entity. In addition to explaining parallel semantic behavior between degree, temporal, and spatial relations, it predicts that only interval plural relations have antonyms, and that antonymic constructions are associated with reverse orderings.
9.4.6 Interim Summary
In sum, just like arguments in favor of expanding the ontology, type-collapsing or eliminating arguments tend to take one of two forms. First, there are morphological arguments that we should collapse two or more types together because there is a tendency for (unrelated) languages to use the same functional word or morpheme to range over them. These include degrees, manners, and kinds on the one hand (Anderson Reference Anderson2016), and sets of degrees and vectors on the other (Schwarzschild Reference Schwarz2012). They also include the observation that operators like many and most range over individuals and events (Doetjes Reference Doetjes2007; Nakanishi Reference Nakanishi2007; Burnett Reference Burnett2012), and that operators like when and if range over worlds, times, and locations.
Second, there are semantic arguments in favor of collapsing or eliminating types. Whereas semantic type-proliferation arguments deal with the issue of semantic adequacy – a model needs a different sort of entity to properly characterize a construction or language – type-collapsing arguments deal with semantic convergence. Situation semanticists (Barwise Reference Barwise1981; Barwise & Perry Reference Barwise and Perry1983; Kratzer Reference Kratzer1989) have argued that restricting and quantifying over worlds, times, and locations independently predicts more truth-conditional variation than we see. Bach (Reference Bach1986a) and Krifka (Reference Krifka1990) were the first to observe a widespread polysemy between individuals and events in constructions with numerals and quantifiers; in Rett (Reference Rett2014) I argued that the polysemy extends to degrees as well. And in Rett (Reference Rett2015), I argued that there is a real semantic difference between the way relations are calculated between interval plurals and between lattice plurals.
9.5 Concluding Summary and Discussion
Proponents of both proliferating types and collapsing them seem to care about the same sorts of things, which makes it easy at least in principle to compare the two perspectives. They care about capturing truth conditions (either making the theory sufficiently fine-grained or preventing it from being too fine-grained). And they take seriously the morphological cues of the language (explaining either the use of the same functional morpheme for more than one putative type, or two different functional morphemes for a single putative type).
Luckily, linguistic theory has something to say about the morphological arguments; specifically, it seems to have principled answers to the question: What’s more compelling, differentiated morphology or syncretic morphology? Loosely speaking, studies of grammatical change have suggested that it’s more likely natural language would co-opt function words for different entities than it would innovate different function words for the same entity. First, when languages add morphemes, especially via borrowing, they either tend not to be, or cannot be, functional morphemes (Thomason Reference Thomason2001). So the typical process by which we see lexical growth in a language generally does not extend to functional items, which has been the focus of the morphological arguments discussed here.
Second, functional items like proforms are more likely to converge over time (cf. Norde Reference Norde2009), via processes like analogical change (Hock Reference Hock, Joseph and Janda2005) or syncretism (Baerman et al. Reference Baerman, Brown and Corbett2005). Bußmann (Reference Bußmann1996) characterizes the former as the “diachronic process by which conceptually related linguistic units are made similar (or identical) in form … often regarded as the result of the move towards economy of form” (p. 21). Syncretism is a term for this sort of process when it affects inflectional paradigms, exemplified in Table 9.2.
In terms of semantic arguments, the question of what is more important, a powerful theory or a restrictive one, transcends discussions of semantic ontology and even linguistics. But there is an important difference between the two approaches in the consistency of their conclusions. The arguments for proliferating types seem to lead to consistent conclusions, crosslinguistically, possibly trivally. And they also result in an attractive (albeit incomplete) universal typology: {individuals, times, worlds}
{degrees} (for instance).
The arguments for collapsing types, on the other hand, seem more inconsistent. Consider events: should they be type-assimilated with individuals, as suggested by individual/event polysemy (Bach Reference Bach1986a; Krifka Reference Krifka1990)? Or should they be associated with possible worlds and locations, as the situation semanticists have argued? And consider degrees: are they more like vectors (Schwarzschild Reference Schwarzschild and Snider2013; Rett Reference Rett2015), or are they more like manners and kinds (Anderson Reference Anderson2016)? While these empirical arguments are compelling in isolation, amalgamating them in a single semantic theory seems problematic.
I’ll close by reiterating that none of these arguments appear to be able to differentiate between Type Flexibility – a semantics with numerous basic types – and Type Ersatzism, with a single type that differentiates between e.g. individuals and events at some subtype level. I know of no empirical argument that can do this. Instead, I hope to have presented a wide variety of existing arguments for leaning towards Type Reductionalism and for leaning towards Type Flexibility/Ersatzism, and I have tentatively argued that the latter are more morphologically plausible and semantically consistent.
10.0 Questions and Answers
(1) Why do you think both linguists and philosophers find the question of ontology interesting?
Philosophers have long been interested in questions of ontology, but from a metaphysical point of view.Footnote * In semantics, it is the philosophical versions of ontological primitives that are used in the construction of the nonlinguistic side of the meaning equation, and which inform the structure of semantic theories. The starting point of this chapter is the idea that metaphysical ontologies are inappropriate when it comes to natural language semantics. Moreover, the claim is not only that ontologies motivated by studying the structure of natural language are superior as a basis for building a natural language semantics, but also that the nature of those ontologies is an important intermediary in the more overarching metaphysical discussion (see also Rett, this volume). In other words, given that we use language to reason about the world, the metaphysical questions cannot be sensibly answered without being explicit about the ontological intermediates required by language.
(2) What recent developments in linguistics and philosophy do you think are most exciting in thinking about ontologies for natural language?
There is much recent work in philosophy of language in the area of natural language ontology (Moltmann Reference Moltmann2017; Fine Reference Fine2014; Hinzen Reference Hinzen2016), which is reinvigorating a strand of (heretical) philosophy that questions the hegemony of Quine, Carnap, and Lewis in motivating the building blocks of meaning. On the linguistics side, recent work by e.g. Potts, Henderson, is attempting to extend the dominant frameworks to accommodate phenomena and modalities of expression that are deeply integrated in human linguistic expression, but which are not so easily accommodated within the traditional, classical model (Potts Reference Potts, Barker and Jacobson2007; Henderson Reference Henderson2015; Gehrke & McNally Reference Gehrke and McNally2015).
(3) What do you consider to be the key ingredients in adequately analyzing ontology when it comes to natural language?
A detailed compositional treatment of natural language semantics distinguished by an emphasis on the lexical primes and category types that we can show have psychological reality as primes of memory and computation.
(4) What do you consider to be the outstanding questions pertaining to natural language ontology?
As with all empirical questions that require a close study of natural language phenomena, the question is always the extent to which the evidence from a single language generalizes crosslinguistically. In other words, which of the natural classes of syntactic and semantic primes are universal/cognitively unavoidable within natural language systems, and which are more linguistically parochial and contingent.
10.1 Ontology and the Problem with Truth
Consider the following three sentences of English.
(1)
a. John closed the door. b. John is closing the door. c. The door is closed.
These three sentences all use forms of the same ‘verb’ in English, close, and that verb contributes to the semantics of what is being said in an important and systematic way, as indeed all scholars of meaning in language would concede. It is natural to associate the meaning of the verb close with all the situations that are ‘closing’ events. However, there is something interestingly different about the three cases. While in (1a), the verb meaning can be elucidated in terms of the situations in which it could be said to be true, namely actual closing events, the verb meanings in (1b) and (1c) are not so straightforward to describe in this way. This is because the situations in which we would judge (1b) to be true are not literally ‘closing’ events in the sense that if John stops half way then what he has done is not ‘close the door’. Nevertheless, what he is doing must be sufficiently similar to a fulfilled ‘closing’ event for it to justify the use of that particular verb, and the situation so described must somehow perceptually and cognitively evoke actual closing events as their organic continuation. The task then becomes how we describe in a formal and precise way what it means to evoke a ‘closing’ event even if it never comes to fruition.
The problem with (1c) is essentially the converse. The verbal form used here is closed, the past participle, and it is systematically and paradigmatically part of the same verbal item as in (a) and (b). Situations that make (1c) true are those in which the door is in a particular kind of state, specifically, the state it would be in if somebody had performed a ‘closing’ event on it. But it is easy to see that no such ‘closing’. event need actually have transpired for (1c) to be true – the door could have been manufactured closed, for all we know. The task here is the temporal inverse of the task in the progressive case – how do we describe in a formal and precise way what it means to be the potential outcome of a ‘closing’ event without making reference to that closing event itself as a primitive.
The problem for both cases reflects choices that have been made in the ontology of how we talk about semantics in linguistics: we choose the actualized event of closing as the primitive with respect to which the subparts of ‘criterial activity’ and ‘criterial result’ should be defined. The reason that the actualized event of ‘closing’ is chosen as the primitive here is because it has a clear and transparent role of truth-making for natural language expressions using the close lexical verb. However, every language I have ever worked on or seen discussed has, in addition to its finite verbal forms, at least one related verbal form which is used in nonfinite contexts, often several. In many cases, these nonfinite forms are at least as morphologically simple, sometimes more so, than the finite inflected forms of the same verb. As we have seen, these nonfinite forms create complexities for the articulation of truth conditions for even the simplest of sentences containing them. Specifically, these ‘simple’ forms have classically been treated by invoking possible worlds as an addition to the ontology required for describing natural language meanings (Lewis Reference Lewis1973, Reference Lewis1986). This allows the event of ‘closing’ to live not just in the real world, but in all worlds imaginable, counterfactual, hypothetical and future. With this toolbox we can talk directly of events that could have happened even though they actually didn’t, or which might come to pass if things continued normally.
In this chapter, I will argue that choosing the actualized event, the truthmaker, as the primitive of our semantic theory, while successful and elegant for some descriptive purposes, fails to make sense of natural language in the sense that it gives us no purchase on the details and specifics of what has elsewhere been called natural language ontology. This question has been a central concern in the work of e.g. Friederike Moltmann (see Moltmann Reference Moltmann2017, Reference Moltmann2020 for summaries of the area). In this article I will be fully endorsing her position that natural language ontology is an important domain within ‘descriptive metaphysics’ (using the term from Strawson Reference Strawson1959), which is distinct from the kind of foundational metaphysics that the philosophical tradition tends to engage itself in. This project analyses the ontological commitments implicit in natural language(s) itself, one of the most pervasive and important reflections of our human engagement with the world. As Fine (Reference Fine2017) argues, there is even a case to be made that progress in foundational metaphysics relies on a close and nuanced understanding of the descriptive metaphysics involved in natural language ontologies. But even if that were not the case, as a linguist, it seems to me that the project of natural language ontology is crucial if we are to understand the compositional products of meaning and meaning-building in language, and the mechanisms by which it is embedded in our cognition and cognitive processing more generally. The spare and elegant axiomatization of semantic descriptions anchored just in truth and reference to particulars simply does not do justice to content and partial and incremental contents that we see in language. Exploring natural language ontology in its own right, taking the internal evidence as primary is a prerequisite to getting this kind of deeper understanding.
In other words, I will be arguing that the ontological categories that we require for understanding the domain of natural language meaning (“its properties and the relations between them”), cannot and should not be identified with external objects in the world or the nature of being, and certainly cannot be explicated in terms of the latter’s ontology. Anchoring all denotations in the primitives of extensional reality justified by some kind of foundational metaphysics, ends up ‘flattening out’ and making opaque the different ways in which such meanings are built up. The internal pieces of compositional meaning in such a system are reverse engineered from their connections to metaphysically justified primitives, and are not taken as evidence of natural language ontological primitives. Of course I would not deny that we clearly want to understand the use of language to ‘say things about the world’, and that a good theory of utterance meaning must ultimately anchor itself in a truth-making relation to situations in the world (Fine Reference Fine2014). What I do deny is that natural language meanings for the lexical and functional formatives of a particular linguistic code should be stated directly extensionally. While we sometimes use meanings to build truth-evaluable thoughts or refer to things in the world, the meanings of lexical items themselves are the cognitive precursors of those acts (see Pietroski Reference Pietroski2018).
The central empirical, descriptive ground in this chapter will be ‘nonfinite verbal forms’ as illustrated briefly in this introduction. In Section 10.2, I take the example of the English participle in -ing and discuss its ‘meaning’ in terms of the classic imperfective paradox that it gives rise to in the English progressive construction. In Section 10.3, I take on a slightly less well discussed case of the same sort of paradox, this time involving the English participle in ‑en/ed the passive/perfect. In Section 10.4, I provide an overarching diagnosis of the problem and in Section 10.5 I lay out the structure of the solution. As indicated in my introductory paragraph, the solution will require a rather different set of grounding assumptions than those we have become accustomed to over the past 40 years.
10.2 V-ing: The Imperfective Paradox
I start with the English participle in -ing, productively formable from every single verb in the language and ubiquitous in the speech of children and adults alike.
(2)
(a) John is singing. (b) Mary is riding her bike.
It turns out that even these most basic of English sentences involve a troubling paradox that becomes visible particularly when a verb with a natural telos is chosen. To see the problem, let us compare the simple past tense of the verb cross in (3a) with the past progressive using the -ing participle in (3b).
(3)
(a) John crossed the street. (b) John was crossing the street.
Difficulties that arise when one attempts to express the truth conditions of the English progressive in terms of the truth conditions of the verb it is based on. Bennett and Partee (Reference Bennett and Partee1972) noticed that you could not give a semantics of the progressive that said it was ‘true at a proper subinterval of the interval at which the nonprogressivized sentence is true’. This is because if you say that, you commit yourself to John actually getting across the street at some future time.
10.2.1 Possible Worlds Accounts
To get out of this unwanted completion entailment, one needs to say that the complete interval corresponding to ‘John cross the street’ is not necessarily true in the real world, but rather in some possible world, plausible given the starting conditions (i.e. ‘inertial’ in Dowty’s Reference Dowty1979 terms), leading out from the situation that the progressive sentence describes (Dowty Reference Dowty1979). The solution in terms of possible worlds was considered a satisfying one when it was proposed. Possible worlds after all had already been accepted as a necessary addition to the toolbox in any case (Lewis Reference Lewis1973) because it is a fact about natural language that meanings can be ‘displaced’ – we can and do talk about things that are not happening now because they are in the past, things that may or may not happen in future, things that we know did not happen, but would have, if … etc. etc. So possible worlds (or something equivalent) is a necessary addition to the ontology of natural language semantics in this view. Note that it is not strictly necessary to see this as an aspect of foundational metaphysics, although many semanticists including Lewis would be happy to commit themselves to this idea. Even if we do not believe that we are living in the multiverse (with or without a different Spiderman in each one), it does not diminish the fact that we need to make reference to some such thing to make sense of how we as humans speak. The question for natural language ontology is descriptive and empirical: How does natural language achieve the expression of these ideas and with which ingredients? There are ways of capturing the notions of choice and indeterminacy without using possible worlds as a technical device. I will show as we go on that possible worlds are used to plaster over many of the cracks that emerge from pursuing the extensionalist agenda. They allow the theorist to count and quantify and verify as usual, while farming out the mysteries of intensional content to an explication of the ontology of possible worlds (something which is just as hard as the problem it is trying to solve). But in fact, in the case of the imperfective paradox, we are still far from being out of the woods. It was quickly noticed that even if all of Dowty’s inertial worlds could never lead to a successful street crossing (because for example a large truck was rushing towards John without him realizing it), we can still say that ‘John was crossing the street’. So inertial worlds simpliciter are insufficient; further research shows that what we need is a notion of continuation that is relativized to the event and modulated by event internal properties such as intention and abilities of the agent (Landman Reference Landman1992). And we are still not done because analyses of the imperfective paradox continue to this day, with more and more fine-grained scenarios forcing the semanticist to precisify the nature of the modal base, and/or call into question the nature of the quantification involved. It is fair to say that we have had nearly 40 years of possible worlds accounts of the humble progressive. The overwhelming consensus of the field is that some such analysis involving possible worlds is the correct one for this construction, and for imperfectives in other languages.Footnote 1
Let us look for concreteness at the denotation of the progressive as given in Landman (Reference Landman1992). First of all, the progressive operator is defined in terms of the full actualized event, as in (a), and the denotation of the progressivized sentence is given in (b).
(4)
John was crossing the street. (a) 
(b) 
For this to have truth conditions, we need to specify the conditions under which
, VP(x)) would be true given a particular e and a particular denotation for x. This is given by Landman using three crucial notions: the stage-of relation between events, and the idea of a continuation branch which is a property of the possible worlds network and relations therein, and an intuition about what counts as a ‘reasonable’ possible world. The stage-of relation is a special kind of subpart relation in which we would be willing to say that ‘it is the same event in an earlier state of development’. The definition of continuation and continuation branch is shown in (5).
(5)
(a) e is a stage of f if speakers would be willing to say that e is the same event as f in an earlier stage of development. (b) f is a continuation of e iff e is a stage of f. (c) f stops at j in w iff there is not continuation of f beyond j in w. (d) A continuation branch CON(f, w) of f in w is a chain of world, event pairs such that the each event is a ‘continuation’ of the previous one in the chain, and each world is the closest to the previous one in the chain and is ‘reasonable’.
Note that ‘continuation’ in the continuation branch is now stripped of its normal real-world connotations because it does not reflect normal extensional temporal contiguity, but is defined at its heart in terms of the speaker’s judgment of ‘stage-of’, itself anchored in a judgment of ‘sameness’. With the machinery in (5) in hand, we are now finally in a position to state the truth conditions for Prog, given in 6).
![]()
In other words, the in-progress event of ‘crossing the street’ for example would be true now, if we could find some continuation branch for ‘cross the street’ which maintained sameness of event and continued into some future close and reasonable worlds in which at some point ‘cross the street’ does become true.
Landman’s (Reference Landman1992) main insight, and the reason why his analysis improves on the earlier rather similar account of Dowty (Reference Dowty1979) is that he notices and builds in the importance of event internal properties in underwriting people’s actual on the ground judgments of the truth of the progressive in certain scenarios. He builds it in by defining continuation branches in terms of judgments of event ‘sameness’. But it is important to see that it is this judgment of ‘sameness’ that is doing all the work here. On top of the machinery of possible worlds, we still need ‘stage-of’ primitive, a residue that is never given complete objective explicitness. This is not just true of Landman’s account (although see Landman Reference Landman and Rothstein2008 which is an admirable attempt in that direction). In Portner (Reference Portner1998) it is the relativization to event descriptions, in Hallman (Reference Hallman2009b) situational version it is the relation R called ‘the relevant subpart relation’. In other words, if you look at these accounts very closely, the essential question of “What does it mean to be an in-progress version of an event?” remains essentially primitive.
My critique at this point is not based on a request for more formal explicitness in defining what it means to be the ‘stage-of’ an event, or what underlies our judgments of ‘sameness’. In fact, I suspect that deconstructing this further is liable to put us on a path of explication in which we are constantly placing the burden on a more and more abstract notion that defies formal explicitness, in a kind of ‘turtles all the way down’ kind of situation. It is exactly in this kind of situation that we as formal modelers should recognize the role of the primitive, or atom in our natural language ontology. Once we embrace and acknowledge the notion of event ‘sameness’ as a primitive cognitive judgment, the question then becomes whether there is anything else of the possible worlds machinery that we actually need.
The problem here is not that we foolishly tried to express the progressive in terms that build on the nonprogressivized version – that is just a basic compositional desideratum. The problem is that the toolbox forces us to start with an extensional meaning for the nonprogressivized version. But I would argue that the essential meaning of a verb (unprogressivized) cannot be captured via extensions, so this strategy is doomed to failure, and the progressive is intuitively related to the main verb in a way that is not defined in terms of truth, but in terms of the primitive of event ‘sameness’.Footnote 2
10.2.2 The Progressive as a Stative Construction Built from -ing
The progressive is surely built from the simple verb via some systematic meaning operation, but one in which meanings are not extensions.Footnote 3 If we are really to take language seriously in what it is telling us about how these meanings are built up, we need to look at the syntactic and morphological devices that recur in language after language to do precisely this kind of thing. The existence of nonfinite verbal forms of particular types are interesting because they point to a recurrent pattern in natural language where the conceptual content of a particular event type is severed from the information expressed concerning the temporal and locational actualization of those events.
In short, verbal lexical items seem to be encoded as belonging to particular aktionsartal categories which then have an effect on how they can be used linguistically and what kinds of entailments they eventually give rise to (see Vendler Reference Vendler1967 and Dowty Reference Dowty1979 for the basic aktionsartal system found in natural language, including a fundamental difference between dynamic and static eventualities and dynamic eventualities which either do or do not make essential reference to a final telos). In addition, languages often provide systematic morphological means to create aspectual modulations of those lexically specified roots, where the latter specifies the particular conceptual information that characterizes the event (i.e. provides lexical encyclopedic content), while the affixation contributes information about the way in which the unfolding of that event in time is anchored to reference and utterance intervals. The latter information is contingent and related to the particular instantiation of the event being described whereas the former represents essential content that is reusable over instantiations. In many languages, aspectual and temporal modulation of verbal content is achieved via ‘auxiliary constructions’ where nonfinite versions of the verb are combined with helping verbs that carry the required tense/aspect morphology required for finiteness. Thus, auxiliary constructions in the world’s languages show in very tangible form some sort of division of labor in building up the meaning ingredients of a complex construction. If we consider the sentence in (7),
John is crossing the street.
we see that the situation is described by using the -ing form of the verb cross as well as the tensed form of the copula be, which carries no particular encyclopedic information, but does minimally bear tense/anchoring information. The primary question must therefore be: how do we characterize the meaning of the nonfinite form crossing such that when it combines with the tensed be, gives rise to something that has the truth conditions of the progressive? Note here that I am not asking “What are the truth conditions of the progressive?” (that is the question we have been working on for 40 years with turtles all the way down). I am asking rather, how do we understand the meaning contribution of the participle crossing – it must be built from the stem cross in a systematic way, and it must contribute an important ingredient to what the progressive ends up meaning. Moreover, whatever we choose as the meaning for crossing, we must also be aware of the fact that it shows up in attributive, gerundive and nominalizing constructions. We have no linguistic or cognitive evidence that there are multiple participles in -ing in English, as opposed to a single form that can be used in multiple constructions as an input to multiple ultimate truth-conditional outputs.
There are a number of things we know about the relationship between crossing and the verbal root cross that it intuitively depends on. One is that whatever aktionsart properties are associated with the lexical item cross as the descriptor for an event, they become flattened out once crossing has been chosen. Specifically, any ‘progressive’ sentence built from an -ing participle and the tensed form of be is, from an external distribution point of view, stative, regardless of the initial aktionsart of the verb.
The modern treatment that takes this fact most seriously is the one in Hallman (Reference Hallman2009a), although he essentially builds on insights from Mittwoch (Reference Mittwoch1988) and Vlach (Reference Vlach, Tedeschi and Zaenen1981). In the literature, the interest in the imperfective paradox has meant that semanticists have focused on the differences between aktionsart categories in the way they feed the progressive construction (accomplishments and achievements give rise to the imperfective paradox, while activities do not; states do not make well-formed progressives). There has been less emphasis on the homogenous output in the form of the progressive construction. For this reason, I go over the clear linguistic diagnostics in English that systematically separate stative verbs from dynamic verbs. There are many. In each case, where you find a diagnostic that successfully separates verbs like know on the one hand from verbs like run (activity), break (achievement) and build (accomplishment) on the other, the progressive construction patterns with the former class.
Most obviously, as emphasized by Hallman (Reference Hallman2009a), the progressive patterns with statives in being possible in the present tense in English with the same interpretation as the past tense (unlike eventives which shift to a habitual interpretation, or a narrative present). The reason is, as Hallman also argues, statives and the progressives can be true at a ‘point’ in time, while eventives which have duration cannot.
(8)
a. John looked tired when I saw him yesterday and he looks tired now too. b. John was writing a novel when I saw him yesterday, and he is writing a novel now too. c. John ate a mango when I saw him yesterday, and ?? John eats a mango now too.
In case we are accused of taking evidence from the very phenomenon we are attempting to explain, consider the other stativity diagnostics we find in English. In the following presented contrasts, I use activities as my example of the dynamic predicate to keep all of the verbs ‘homogenous’. Homogeneity is not the deciding factor underlying these diagnostics, stativity is.
Interaction with ‘When’-Clauses, and Narrative Progression
When the progressive construction is the main clause modified by a when-adverbial clause, the situation of the when-clause overlaps with the main clause situation (9a). This is like statives (9b) and unlike dynamics (9c), which get a sequential interpretation. The point about the progressive and when-adverbials was originally made by Leech (Reference Leech1971)
(9)
a. When we arrived she was buying up all the restaurants in town. progressive b. When we arrived, she owned all the restaurants in town. state c. When we arrived, she danced for joy. activity
In terms of narrative progression, in (10c) the event in the middle sentence advances the narrative time while in (10b) and (10a) it does not (cf. also Kamp & Reyle Reference Kamp and Reyle1993).
(10)
a. John arrived. He was sweating. Then he left in a hurry. progressive b. John arrived. He looked hot and bothered. Then he left in a hurry. state c. John arrived. He laughed hysterically. Then he left in a hurry. activity
Semantic Selection
Hallman (Reference Hallman2009a) adds further diagnostics to the stativity claim. For example complements of ECM discover and reveal in English must be specifically stative and are bad with events of all kinds, including activities. Once again, the progressive patterns with the statives with respect to this test. The following data is from (Hallman Reference Hallman2009a: 8)
(11)
a. The inspector revealed/discovered Max to be lying. progressive b. The inspector revealed/discovered Max to be a liar. state c. *The inspector revealed/discovered Max to lie. activity
The Universal Reading of the Perfect
Portner (Reference Portner2003) points out that the universal reading of the perfect is triggered in English for states, and is impossible for events of all stripes including activities. In (12), I use the since 5 o’clock phrase to trigger and force the universal reading of the perfect. Only statives (12b) and progressives (12a) are licit.
(12)
a. John has been jogging since 5 o’clock. progressive b. John has known the answer since 5 o’clock. state c. *John has jogged since 5 o’clock. activity
Epistemic Readings under ‘Must’
Finally, Ramchand (Reference Ramchand2014), points out that the modal must in English can only get an epistemic reading with stative prejacents, where it is ambiguous with a deontic interpretation. For dynamic predicates, only the deontic interpretation is available
(13)
a. Mary must be jogging in the park. progressive b. Mary must know the answer. state c. Mary must jog in the park. activity
10.2.3 Taking Stock
We have seen that attempting to be explicit about the meaning of the English participle in -ing throws up some tough desiderata. Let us summarize what has been learned:
(i) The progressivized eventuality in -ing is related in an organic way to its nonprogressivized counterpart, but does not actually entail it (in the actual world) at a future time.
(ii) The perceived relationship between a progressivized event and the event simpliciter gives rise to variable judgments across speakers. In this regard, internal properties of the participants and their intentions, and the nature of the process evidenced seem to be more important than external circumstances.
(iii) The progressive functions like a state in its temporal semantics and external distribution.
As I have argued in the previous section, all of the possible worlds accounts we have seen fall short of complete objective explicitness when it comes to point (i) above. In all cases, the appeal to possible worlds still leaves an unexplained residue completely independent of the possible worlds mechanisms themselves. The unexplained part is the core of the progressive meaning itself – the idea of what it means to be a subpart of a particular event described in a particular way. It is unclear what the rest of the machinery is contributing.
Instead, we need to assume the equivalent of the unexplained part as the basic cognitive primitive. In other words, the ability to identify an event as being of a certain ‘type’, as represented by the reusable lexical item run, build, or whatever, is a sensory/cognitive judgment that forms the basis of our ability to classify the world based on symbolic labels. There is good evidence that the building of the derived stative participle in -ing is in fact cognitively basic. It has been known for a long time that the progressive participle in -ing is one of the very earliest pieces of morphology acquired by English children. It is acquired between the ages of 19–28 months, and appears before both irregular past tense (which in turn appears often before regular past inflection) and the copula (Brown Reference Brown1973; Owens Reference Owens2001). The use of the -ing participle thus appears before any actual tense inflection or modal expression, and is used correctly immediately.Footnote 4 A fully modal and intensionalized analysis of the progressive would require us to believe that English children acquire a modalized meaning accurately before they are two years old, and always do so before they even have the ability to express tense or use modal auxiliaries. The pragmatic complexity of inferences connected to the setting up of modal bases and ordering sources is supposed to be something that children need some social and interactional maturity in order to develop. But standard accounts seem to assume that they can do this even before they pass theory of mind tests.
The meaning of the participle in -ing must indeed be tethered in some way to the finite verbal form, and it must be so in a direct and cognitively obvious way. Let me state it pretheoretically in terms of descriptive content, independent of how we cash out the intuition in a formal semantics framework.
(14)
The Meaning of V-ing V-ing expresses a stative eventuality that gives cognitive/perceptual evidence for the essential descriptive-classificatory content associated with V.
So the idea here is that the nonfinite verbal form in V-ing is related to V by the most basic of identificatory relationships, with all specific dynamic and instantiational information removed. In other words, V-ing describes a state (which can be true at a moment), which nevertheless is apprised by human cognizers as being qualitatively classifiable by the descriptive content of V. For this informal characterization to work, V’s descriptive content in turn must be that part of the meaning that has to do with its ‘essential nature’ (in the sense of Fine Reference Fine2005) and not defined in terms of the set of eventualities it is true of.Footnote 5
10.3 V-en: The Paradox of Target States
We turn now to the other major nonfinite form in English, the ed/en-participle found in the passive (and perfect) constructions. I will abstract away from its use in the dynamic constructions (eventive passive and perfect) and concentrate on the stative use of this participle, since this is where the paradox emerges (but see Ramchand Reference Ramchand2018 for an account of the participle that encompasses all three uses).
Even confining ourselves to the stative use, in the literature a number of different types of stative participle have been claimed to exist. In Embick (Reference Embick2004), these are called ‘resultative’ vs. ‘stative’ participles (although they are both actually stative, as he acknowledges). In Kratzer (Reference Kratzer2000) a ‘resultant state’ vs. a ‘target state’ passive participle are distinguished. In fact, these two authors are not making precisely the same distinction with these labels, so we cannot simply choose our terminology here. In the case of Embick, what is important in distinguishing the two classes is the presence or absence of ‘event implications’: resultative stative participles have event implications, ‘pure stative’ participles do not. In the case of Kratzer, what is important for the label is the relationship of the stative meaning to the meaning of the verb as a whole: ‘target state’ participles denote a state which is already an internal component of the verbal denotation; ‘resultant state’ participles denote a state that holds forever by virtue of the event in question having occurred, as in the distinction originally proposed by Parsons (Reference Parsons1990) for the perfect. Each of these distinctions comes with its own set of diagnostics. In fact, as we will see, it is not the case that target state participles are the non-event-implicating participles and resultant state participles are the event-denoting participles. Rather, target state participles can be both event-implicating and non-event-implicating in the relevant sense, while resultant state participles are only event-implicating.Footnote 6
The point I am interested in here is in what is meant by ‘event implications’. Let us consider first the following diagnostic, as proposed by Embick (Reference Embick2004).
(15)
a. The door was built open. b. *The door was built opened.
According to Embick and much subsequent work, the problem with (15b) is that the state of being ‘opened’ simply cannot be true in the world unless there has been a prior event of ‘opening’, i.e. it is not something that can be one of the door’s properties before anything has happened to it. This is not true of the underived adjective open, which can be true regardless of how that state so described came about. In the case of open, there is an underived adjective which has the non-event-implicating reading, effectively blocking that interpretation for the participial form. But in English, there are many participles in en/ed which do in fact allow this kind of reading. Consider the participle closed.
The door was built closed.
In other words, it is perfectly sensible to describe a door as ‘closed’ even if that particular state was never actually preceded by a ‘closing’ eventuality. What is important to note about this situation is that the nature of the state described by the participle closed is in fact related in an intimate and organic way to the verbal concept close – the state we are describing is one that would typically result from an action of ‘closing’. To give another example, one can describe a drawn object as a ‘flattened cube’ because it has exactly the configuration of lines and angles that you would get if you took a three dimensional cube and flattened it symmetrically. There is no implication that it is a cube that has been flattened, just that it can be accurately described as such. The description is crucially dependent on a knowledge of what it means to ‘flatten’ something, but not on the actual existence of such an event.
The criterion of event implications in the literature that stems from Embick’s original distinction refers to the property of entailing the previous instantiation of an event particular. Thus, target states do not have event implications in Embick’s sense, but they are conceptually related to the corresponding dynamic event and thus give rise to what I will call the ‘resultative paradox’ – they describe the result of an event, which in fact did not have to happen to produce it.
Can Kratzer’s distinction between ‘resultant state participles’ and ‘target state participles’ get us out of the resultative paradox? Unfortunately not. Kratzer distinguishes between ‘resultant state’ participles and ‘target state’ participles, but as Embick (Reference Embick2004) points out, the phrasal target state reading that she analyses and gives a denotation always has event implications of necessity, since it requires existentially binding the Davidsonian event variable corresponding to the verb. It is only a pure adjectival reading that corresponds to the ‘pure state’ reading in Embick’s terms. The denotation for the adjective cool vs. the target state cooled from Kratzer, cited in Embick (Reference Embick2004) is given below in (17).
(17)
a. 
b. 
We can see that in the case of the non-event-implicating reading of closed and flattened given above, this definition has the unwanted eventuality entailments built in to it and simply won’t do the job. However, Kratzer’s denotation for the target states is importantly correct in one respect, namely that the verb that gives rise to the target state includes specific conceptual information corresponding to the result state that the participle eventually denotes.
The resultant state passives in Kratzer (Reference Kratzer2000), on the other hand, are the ones where there is no readily available state in the denotation of the verb’s meaning. Instead, the state that the participle denotes is the state that Parsons (Reference Parsons1990) calls the ‘resultant state’. The definition from Parsons is given in (18).
(18)
Resultant states “For every event e that culminates, there is a corresponding state that holds forever after. This is ‘the state of e’s having culminated,’ which I call the ‘Resultant state of e,’ or ‘e’s Rstate.’ If Mary eats lunch, then there is a state that holds forever after: The state of Mary’s having eaten lunch.”
Kratzer’s diagnostic to distinguish resultant states in this sense from the others (the target states and the pure adjectives) is the incompatibility with the adverb immer noch-‘still’. This is because the definition of resultant state means that the state persists indefinitely for ever after the event is over, and therefore trivially the adverb ‘still’ cannot meaningfully be applied to it.Footnote 7 The resultant state passives given by Kratzer (Reference Kratzer2000) are shown below in (19).
(19)
a. Das Theorem ist (*immer noch) bewiesen. The theorem is (*still) proven. ? ‘The theorem is (*still) proven.’ b. Der Briefkasten ist (*immer noch) geleert. The mail box is (*still) emptied. ? ‘The mailbox is emptied.’ c. Die Gäste sind (*immer noch) begrüsst. The guests are (*still) greeted. ? ‘The guests are greeted.’ d. Die Töpfe sind (*immer noch) abgespült. The pots are (*still) washed up ? ‘The pots are washed up.’
Kratzer (Reference Kratzer2000)’s semantics for the resultant state does not produce a property of events, but rather a property of times directly. However, it is important to note that her semantics for both the target state and the resultant state require actualization and have real event implications, since for her events are instantiated particulars. So with respect to event implications, target state passives and resultant state passives are on a par. The only difference is the way in which that state is constructed.
The Target State passive participle, the one that is compatible with immer noch ‘still’, has a strong constraint imposed on it. For these to be formed, the verb in question must contain a caused result state in its denotation. Kratzer (Reference Kratzer2000) diagnoses this by the fact that a ‘for’-phrase is felicitous as a measure of the duration of that caused state. It is precisely these verbs that form good target state passives with immer noch, Kratzer argues, that consist of an activity portion and a final state. It is this ‘final state’ that ends up being the denotation of the formed up participle.
If we consider the denotation Kratzer assumes for the target state verb aufpumpen ‘pump up’, we see that it contains the representation of a caused final state.
(20)
das Boot aufpump- - ‘pump up the boat’ 
According to Kratzer’s semantics, the output of the stativizer -en/ed is a predicate of states, exactly the one that is inside the verb’s complex event semantics. The external event variable (the ‘process’ variable in my terms), is existentially bound.
(21)


So the participle morphology in Kratzer’s system does not do very much work except to existentially bind the ‘Davidsonian’ event, and also to license the absence of verbal inflection. Since in the vast majority of cases, ‘a pumped up’ state is actually preceded by a ‘pumping up’, Kratzer does not notice or address the resultative paradox. But it is a fact that the target state stative participles, defined in terms of being a subpart of the whole verb’s event decomposition, are precisely the ones where the resultative paradox emerges. Once again, the existence of the meaning in question is conceptually dependent on knowledge of the whole event description corresponding to the verb, but this cannot be cashed out in terms of actual events.
Going back to the case of the ‘closed door’. To describe the truth conditions precisely, we would have to make reference to possible worlds, counterfactual worlds in this case, preceding the current moment, which would contain dynamic subparts of the event of ‘closing’ that culminate in the final state of ‘closedness’ in the actual world. We must find a continuation branch from a previous possible world into the present, even though in fact, that previous possible world is not the real one. This is all that is required to felicitously refer to something as ‘closed’. Most times when we use the stative participle, there is in fact a dynamic act of ‘closing’ that leads to it, but this does not appear to be necessary. Although I do not know if anyone has stated the semantics in exactly these terms before, I would contend that the case of pure stative participles of the non-event-implicating kind, what we are seeing in the resultative paradox is in fact the perfective counterpart of the ‘imperfective paradox’, and would require the equivalent move to the possible worlds toolbox.
10.3.1 Taking Stock
Here once again we have a problem with relating the meaning of the nonfinite verbal form, in this case the participle in -en/ed, to its finite counterpart. We must somehow keep some core contribution to the truth conditions constant, while removing any actual entailments about the way that event will play out or has played out in real time. But how do we even talk about this part of the ‘meaning’ of the verb when meanings are given only in terms of actualized events? In both cases, it seems, we can preserve the extensionalist-style mapping by augmenting the system with nonreal and hypothetical worlds in the model.
But even this move, as with the imperfective paradox in the previous section, doesn’t add any formal explicitness. This is because there is no real way of making sense of the notion of ‘subpart of an event’ without making reference to teleological intentions and mentalist conceptual characterizations related to the very lexical items used as categorization devices.
(22)
Once again, we seem to need V’s descriptive content to be part of its ‘essential nature’ (in the sense of Fine Reference Fine2005) and not defined in terms of the set of eventualities it is true of.
10.4 Diagnosis
The discussion of the meaning of these two core nonfinite verbal forms in English (and I have no reason to believe that trying to understand participles in any other language would be any easier), has led us to seemingly unresolvable paradoxes, which I have argued cannot actually be defused by helping ourselves to a model invoking possible worlds. Moreover, the paradox in question is actually the same one, and has the same ultimate cause – it results from attempting to express essentially internalist compositional ingredients in externalist terms.
Chomsky (Reference Chomsky1995) argues convincingly to my mind that reference and truth are themselves mystical notions, and that the attempt to fill out some materialist agenda closing the gap between mind and body by bypassing the mental in favor of external reality fails to do justice to the reality and structuredness of mental representations.
It is a hopeless task to ‘complete the materialist world picture’ by translating accounts of ‘mental phenomena’ in terms of a ‘description that is either explicitly physical or uses only terms that can apply to what is entirely physical’ or perhaps give ‘assertability conditions’ on ‘externally observable grounds’. (Nagel 1993: 37)
In other words, by starting from the assumption that sentences denote truth or falsity and that meanings of lexical items are functions from entities to truth values, or some recursive type associated with those extensional primitives, one effectively begs the question. In other words, it tells us nothing about what precisely the cognitive ingredients are, and how human minds combine them to create these complex truth-evaluable expressive acts.
I wish to argue that the difficulties described above in characterizing the meanings of perfectly ordinary elements of natural language should be evidence that we need to rethink the very foundation of our formal theories of semantics, and abandon truth as the language in which we explicate meanings of natural language formatives. We do need to preserve an indirect relationship with truth-making, via the explication of the pragmatically enriched sentence or proposition, but we should simply abandon the idea that ‘meanings’ of language ingredients are stated directly in terms of truth conditions. This heretical position has been argued for by others, most particularly Paul Pietroski in a series of articles and books and most recently in Pietroski (Reference Pietroski2018). Problems with the truth-conditional approach to natural language meanings are well known, but often ignored strategically because of the payoffs gained in getting a robust theory of semantics off the ground. It is time to revisit those problems and reassess the foundations.
10.4.1 Problems Too Hard for Truth Theories
There are certain number of classical paradoxes for the current system that are well known. The first is the fact that even with the addition of possible worlds (Lewis Reference Lewis1973, Reference Lewis1986), truth-conditional meanings do not seem to be quite fine-grained enough to do justice to Fregean ‘senses’. The problem arises with logically equivalent statements, which should give the same truth value in any possible world, but which are obviously distinguishable in terms of sense or naive meaning (cf. Pietroski Reference Pietroski2018: chapter 1).
![]()
It is hard to see how an extensional theory of meaning based on truth conditions avoids this, without resorting to a condition on the use of particular lexical ingredients. See Kratzer (this volume) for an extensive discussion of this problem and for a proposal concerning how to capture the specificity of content in speech and attitude reports even for logically equivalent propositions. In a nutshell, while Kratzer makes a case for putting aside mathematical statements, the problem remains even for regular language. For the two kinds of cases that Kratzer considers, it turns out they can only be solved by giving up on compositionality, and making the truth conditions of the larger speech dependent on more than just the intensions of its parts. In the case of Kratzer’s own solution for speech predicates, this involves including the ‘guise’ of the reported speech, specifically, the intensional structure, or the way in which the proposition is put together. As we will see when it comes to my own proposed ‘solution’ in Section 10.5, I take the intuition behind the idea that the way in which the proposition is structured matters to its extreme conclusion and build explicit reference to the symbols deployed right into the heart of the meaning composition. While Kratzer is showing us how to solve the problem of composition under attitude and speech predicates with minimum disruption to the standard toolbox, invoked only when nothing else will work, I will pursue a different strategy and take the hard cases as evidence that foundational aspects of the composition system for natural language need to be rethought.
Problems for truth theories also arise in interaction with the use of the predicate truth itself which leads to the Liar’s Paradox (cf. Pietroski Reference Pietroski2018: chapter 4). These problems are well known and there has always been a tacit agreement among working semanticists at any rate to just ignore them and get on with things.
There is also the problem of semanticity. Consider the truth theorems that Davidson urges on us as things that a suitable theory of meaning should be able to derive.
‘Snow is white’ is True if and only if snow is white.
This is all very well and good, but it is not just important for the theory to derive (24), it also must not generate (25)
‘Snow is white’ is True if and only if snow is white and 2 + 2 = 4
This leads us inexorably to the final problem, and the real deal breaker in my opinion. This is the problem of the meanings of open class lexical items – their flexibility, their polysemanticity, even while exhibiting productive and generative composition with each other. One could argue that building truth theorems of the type in (24) is already hard, and good enough, and that it really is not the immediate job of the formal semanticist to elucidate the meanings of the individual lexical concepts snow and white. However, it simply will not do if our aim is understand how natural language works. The problem is that the meanings of open class lexical items are not monolithic blobs of coherent conceptual stuff encapsulated and insulated away from the combinatorial system, the fact is that both (i) they are conceptually polysemous, and (ii) they undergo productive compositional processes with each other. The latter point shows that understanding their behavior is an important component of understanding the central properties of the human language system and its powers of productive meaning generation. So, we cannot ignore the word book, to take an example, and we cannot ignore the fact that book in (26a) refers to the content, while in (26b) it refers to the physical artifact.
(26)
a. The book is interesting. b. The book is on the table. c. The interesting book is on the table.
No natural language that I know of fails to have this particular ambiguity or codes it by means of explicit morphology. There is a lot of interesting semantic work on this topic, most prominently Pustejovsky’s work on the generative lexicon (Pustejovsky Reference Pustejovsky1995), and Asher’s work on semantic types (Asher Reference Asher2011). This tradition seems to exist in a separate dimension from much linguistic semantic work, although it has been extremely influential in computational applications, partly because the current toolbox puts this kind of work in a separate, independent module which nonlexical semanticists can decide not to worry about.
The by now standard approach from formal compositional semantics is to reverse engineer the meaning required based on the truth conditions for the sentence as a whole. This would, effectively, create two distinct lexical meanings for (a) and (b), with no real engagement with the systematicity of the alternation. In (26c) natural language seems to allow it to refer to both things simultaneously, throwing up insurmountable problems for standard denotations of type <et> (see the computational literature on dot types, Pustejovsky Reference Pustejovsky1995 and Asher Reference Asher2007, for a discussion of the hardness of this problem).
10.4.2 Problems Too Easy for Truth Theories
Pietroski (Reference Pietroski2018) points out that the Frege–Tarski–Church playbook (having been designed ultimately for rather different purposes) offers the possibility of many more types of function than are ever used in the analysis of human meaning combinations, and even those that are used could be rethought in more minimal terms.
In many ways, actual natural language combinatory relationships are a much simpler subset of what is potentially expressible in the lambda calculus. So it seems that it is the job of linguists to actually make decisions concerning how natural language fits into this machinery.
It is precisely the job of linguistic semanticists to classify the primitive cognitively realistic meaning units and the substantive types of meaning composition we actually find. So for every detailed problem of meaning composition natural language throws up for our consideration, and there are many, the solution offered by lambda calculus is simply too easy, and basically trivial to implement.Footnote 8 In the absence of a real theory of meanings and meaning combination, we hide behind the rich descriptive power of the toolbox which for this reason never fails to deliver as a description, but ultimately fails to explain.
10.5 Towards a Natural Language Ontology for Meaning
I contend that we cannot just acknowledge the flaws the current system and use that wisdom to deploy the system carefully and circumspectly. All the puzzles and paradoxes are the same, all the dead ends have the same deep problem at their heart. All the technical solutions contain the same self-defeating trick (structure the possible world space in an increasingly elaborate, context-sensitive way, often depending on the actual words used in building the construction). This shows that we really have not understood the meanings of words in a way that is helping us to solve the compositional issues.
The bottom line here is that we can grant that formal semantics is a computational theory (in the sense of Marr Reference Marr1982) of meaning and still hold it to certain standards of correctness. Even if successful in its own terms, correctness needs to be understood at least partly in terms of a theory’s potential to inform and constrain theories of representation and algorithm. The current theory, even when wielded by extremely intelligent and knowledgeable linguists ends up forcing us into dead ends and paradoxes. All of which could be avoided by dropping the empirically undermotivated assumption that meanings are immediately extensional.
And it’s not so bad actually to have had 50 years of this particular tradition of formal semantic theorizing. There is a lot we know that we did not know before. In concentrating on the externally verifiable truth conditions of sentences, we have a sort of E-semantics (in the sense of Chomsky’s E-Language) descriptively in place, that can form the foundation for an I-semantics of language, if we are willing to take the next step.
In the case of Pietroski, he proposes that we think of Meanings as instructions to fetch a potentially polysemous network of concepts, and further that Meanings are confined to a restricted version of monadic predication (M predicates) and dyadic predication (D Predicates). They combine in ways that are severely restricted compared to the full combinatorical power of the type-based approach using functions. These are defined so as to ensure that the meaning of two joined meanings properly includes those individual pieces.Footnote 9 Pietroski defines two basic semantic combinatoric operations: M Predicates combine with M Predicates by M junction and M predicates combine with D predicates by D junction.
I welcome the Pietroskian critique of truth, which is close to my own position, but I think that the meanings of nonfinite verbal forms which formed the empirical heart of this chapter tell us something important about what needs to be added to natural language ontologies to get a satisfying account of partial contents and compositionality at the sentence internal level. In what is left of this chapter, I will briefly lay out my own formal solution to the paradox of maintaining the identity of descriptive content across formatives in the absence of actuality implications. To anticipate, the surprising new addition to the ontology required by natural language for the expression of complex meanings will be the linguistic symbol itself.
The starting intuition follows an old line of thought most forcefully articulated in Barwise and Perry (Reference Barwise and Perry1983) concerning the reusability of the sign. For human language to get off the ground, and because of the reusability of the signs of natural language, we need to have (i) common possession of symbols that are abstractions over the different actual situations encountered in the learning phase, and (ii) a speaker to deploy those symbols as a means of characterizing new situations in the world as she comes across them. Put in this way, we can see an important logical separation between the symbols, or elements stored in declarative memory with their rich web of associational content, and the external world which the speaker is using those symbols to describe. Because symbols are reusable they cannot be tied to a particular time and place, and they cannot be overly rigidly tied to particular worldly particulars. Polysemy and flexibility must be seen as features of the symbolic inventory not bugs. Thus symbols, because they are stored as abstractions or generalizations over particulars function as types rather than tokens, although crucially they are not ‘kind level’ in the formal semantic sense as being literally built from particulars, as in some attempts to capture this intuition. To be sure, the process of acquisition must involve the experiencing of particulars, but symbols are then stored and used in the system as abstract associational bundles, in much the same way that the cognitive scientists have assumed and maybe even best implemented in terms of a vector semantics (see Rett, this volume for more discussion). How then can we move from this psychologically plausible internalist conception to a formal semantics describing real-world objects and situations, and how can we model it computationally? Intuitively, the bridge is made by understanding that the symbol is to be literally deployed by a speaker at a particular place and time in order to describe a new real-world particular in the human being’s flow of experience. To do this, symbols must literally become objects in the ontology themselves, artifacts of human language.
I thus propose a radically different computational theory of meaning composition: this logical architecture seeks to reconcile internalist conceptions of meaning with the well-established formal foundations of an externally grounded theory of meaning. I argue for a view whereby the symbol is reified in the formal system and then explicitly deployed via an act of communication. The separation of the act of communication from the symbol itself allows for an internalist conception of the symbol qua symbol to be embedded within an extensionalist semantics for the content that is created after deployment in a context.
In Ramchand (Reference Ramchand2018), I argued that typological meaning layering in the verbal domain and was part of an argument against purely extensional meanings of lexical items (see also Pietroski Reference Pietroski2018). Here I take the idea further and propose that lexical items (items stored in declarative memory) are the reusable symbols whose meaning needs to be represented nonextensionally (internalistically). Functional items, on the other hand, contribute information that is directly relevant to reference in the world and the instantiation of situations (see Figure 10.1).

Figure 10.1 Meaning layers of the Symbol-within-Reference model
To bridge the internalist–externalist divide (chasm, one might say), information about the context of utterance (via an event variable corresponding to the very act of communication) must be used to mediate the internal representation and the externalist one and gives rise to a second zone where externalist content is structured and manipulated.
So in brief, the symbol is deployed by a user in a communicative context, giving rise to content that actually says something about the external world. Actual reference only occurs in the second step, crucially mediated by the speaker, the here and the now. This general architecture is represented schematically in Figure 10.2.

Figure 10.2 Information composition in the (Sym)Ref model
This is different from the standard classical semantic formalism where the most primitive denotations are couched in terms of objectual referents and truth, while intensional concepts have to be built by generalizing these denotations over possible worlds. The present model inverts the primitive-derived distinction found there, and structures the system around a primitive abstract and essence-like core with referential objectual and instantiational facts built up on top of that.
Formal semantic theories as we know them sit squarely in the domain of ‘computational theories’ in the sense of Marr (Reference Marr1982) and as such are independent of the nitty gritty of minds and brains that might implement them, and Marr’s work has been widely read in linguistics as a plea for and a legitimization of this level of scientific theorizing. But Marr was interested ultimately in understanding the functioning of our internalized systems in creating behaviors in response to stimuli, and notationally equivalent computational theories could nevertheless be distinguished by how well they help us understand the other, more algorithmic levels. His model included both an internal representation (the primal sketch) fed directly by light stimulus, and a final 3-D representation of our cognitive uptake. For Marr also, there is crucially a role for the subjective perspective in between the two (what Marr called the 2 1/2-D level), which in a sense allows the transition from the primal sketch to the objectual 3-D representation. In my own system, the reification of the utterance event and the parameters of context c that it brings with it performs the same cognitive function as the folding in of the subjective perspective in Marr’s own algorithm. I do not think it is an accident that the two systems mirror each other in this way. Marr’s own proposed computational system, which was designed to interface with a more algorithmic understanding of visual perception in the mind is a blueprint for cognitive neuroscience more generally. The idea that explicit information about the Self, or Origo is necessary to convert the stimulus into the information necessary to build an objectual 3-D representation is echoed in my claim that the reification of the utterance context is necessary to convert abstract symbolic content into representations with truth-conditional import. This model of meaning puts the identity of the anchoring information, consciousness of Self and the cognitive appreciation of the fact of other minds at the heart of the creative construction and comprehension of meaning respectively, and it makes different predictions for the role of context in the time course of complex meaning construction than other models of meaning.
In logical and philosophical terms, the picture above is a more Kaplanian (Kaplan Reference Kaplan, Almog and Wettstein1989) view of context as mediator between character and content, but it is at odds with the more classical formal semantic strategy of integrating context after linguistic form has been processed. This means that no current theoretical semantic proposals incorporate it – it would require a formal rift between symbol and symbol deployer in the build-up of the compositional system. Conversely, cognitive and internalist theories of meaning struggle with dealing with the aboutness of language and with reliable facts about how speakers judge entailment. In addition, cognitive theories do not concern themselves centrally with the structural meaning that arises procedurally from hierarchical syntactic structure. So far neither view has succeeded in building a genuinely algorithmic theory of meaning composition in the brain, which to my mind must deal with both of these kinds of phenomena.
Because it is the symbol that is being combined and manipulated in the earliest zone of compositional combination, we are not yet dealing with actual events in the world, but with complex concept combination. At this level of formal composition, we need to be manipulating elements of a new domain
which we add to the model, following Potts (Reference Potts, Barker and Jacobson2007), which is essentially the domain of well-formed linguistic symbols, possibly nonatomic, of type µ. To do this formally we need to add to the usual model, a domain which is the domain of well-formed linguistic objects. I will assume that these linguistic objects are triples, consisting of a < phonological information, syntactic features/information, semantic representation >, very similar to what is assumed in the psycholinguistic literature as a structured associational bundle. So we could think of the denotation of the English verbal symbol run as an ordered triple as in (27).
![]()
where we represent the content of the third, semantic, position in terms of a (necessarily incomplete) set of properties of events to indicate that it is a shorthand for a certain abstraction of cognitive properties, a meaning contribution which is partial on the one hand when it comes to describing an event in the world since it lacks temporal and locational specifications, but also potentially excessive and irrelevant on the other in that it includes ‘irrelevant’ and associational information as well (including the fact that meetings can ‘run’ as well as things like water and trains). This triple is also the piece of information that is stored in declarative memory, and which indeed may turn out to be most satisfactorily represented through vector semantic notation, as suggested by e.g. Baggio (Reference Baggio2018).
The point is that the internal structure of the elements in the domain
include a semantic contribution, which can in principle be filled in with whatever associational and incomplete meaning contributions turn out to be psycholinguistically and neurologically justified. Crucially, we do not need to state the meaning member of the triple in terms of the kinds of extensional formulas that formal semantics traffics in. To reiterate, the reason we have the freedom to place an internalist conception of meaning in this part of the ‘denotation’ for the symbol run is because ‘meaning’ in the internalist sense is encapsulated away from the formal computation. The encapsulation is effected by reifying the symbol itself as an object of the ontology. The symbol itself is an artifact of human communicative life, and it is an object in the domain of reference as much as a book, or a piece of string.
Importantly, the ‘meanings’ of the symbols of
need to be devoid of temporal or worldly information. They form the hierarchically inner core which is then clothed with the contingent information of time, place, and world, to link descriptions to actual particulars. True, the symbols of a person’s language are acquired through actual experience of the world, both sensory and cognitive, but they are, importantly, abstractions over the particular instantiations exposed to. The result (which may be fine tuned over the life of the speaker) is a partial description of essential properties for the event that, on grounds of reusability, are necessarily devoid of information related to the particularity of the instantiation (time, place and the reference of the participants). Humans possess an inventory of symbols, which are then consciously deployed in multiple situations. It is reusability, this third factor property of the symbolic system itself that leads to temporal and spatial, and referential information being represented morphosyntactically external to the memorized sign (the topic of Ramchand’s Reference Ramchand2018 monograph).
Now, let us imagine that conceptual content is associated to symbols we store in declarative memory, and that it can be combined by simple processes of primitive semantic combination to create complex concepts. We can call this Stage-One-Semantics for convenience. It is going to be an open question exactly what the possibilities and limits of internal concept combination are, but the hypothesis is going to be that it is qualitatively different from the next stages.
The reification of the symbol qua symbol in the computational representation is what allows us to embed the internalist intuition in the clothing of an externalist computational paradigm. Once the symbolic layer (possibly complex) has been constructed, these partial descriptions (properties of symbols) can be converted into formal properties of events that are identical to the kinds of denotations we are used to from formal semantics. To do this, Stage-One-Semantics must be followed by the integration of meaning corresponding to the Origo, the speaker in the here and the now, in order to convert abstract conceptual content into actual claims about the world. This is done by means of an explicit deployment operation. Our starting point is the innovation proposed in Henderson (Reference Henderson2015) who needs it to incorporate ideophonic, or depictive content into the descriptive system. The conversion operator (called QUOTE) by Henderson, combines with a complex symbol, notated u in (28) below, and a demonstration event d, to produce a property of events. The predicate
relates the demonstration event d to the symbol being deployed, u. The DEMO predicate relates d and a general event variable e, saying that d ‘demonstrates’ or has certain structural properties in common with e.
I propose to carry the QUOTE function over to the general case, proposing that the reusable essential symbolic content of any perfectly ordinary sign is the equivalent of Henderson’s ideophone. In other words, a symbol, as reconceived this way, is a hyper-conventionalized ‘ideophone’ used to invoke and describe an event, with a higher proportion of descriptive as opposed to depictive content. In order to make this fully general, though, we need to replace Henderson’s DEMO predicate with what I will call CONVEY, and replace the exotic idea of a demonstration event with the ordinary garden variety utterance act itself. Thus, Ramchandian externalization operator would have the following general form.
(28)
To build the externalized representation, I follow Champollion (Reference Champollion2015) in closing the event variable low, at the point of inclusion of the deployment act of communication, introducing a variable over spatiotemporal properies of eventualities. In other words, these are now properties of fully objectual eventive particulars.
(29)
The Externalized Representation
[Utterance(d) & ‘u is the symbolic content of d’ & convey(d,e) & f(d,e)]Where f is a variable representing spatiotemporal/worldly properties of events which I propose is a complex property relating d and e. Put another way, f is a property of events e, anchored in d.
In other words, this is now a property of a communication event d and event e in the world, where d deploys the (possibly complex) symbol
in order to convey e. Truth-conditional content is created only after deployment in a context by a speaker. This corresponds to the idea of demonstration in the gestural and ideophonic literature. The novelty of this model is that the symbol is always in some sense ‘demonstrated’, whether that symbol is conventional or possesses some iconic properties.
It is only at the point where an externalized denotation has been built up that what linguists term functional items, or grammatical words, can be integrated with the message. Here I am thinking about functional items like the determiner the or the past tense marker -ed in English. These items have the semantic property that they are related to extensional meanings, or in simpler terms they involve actual reference to entities in the world. The very earliest studies of aphasias and language disorders show a robust double dissociation between function words and richly contentful lexical items. Both of these kinds of items have a ‘semantics’, in my view, but of a radically different kind. The meanings of lexical items are associational and abstract, while lexical items clothed in function words have truly referential meanings. Formal semantic theories with their insistence on extensional denotations for everything including lexical items like dog and run are not in a position to make a distinction in these terms. Because it builds in meaning layering explicitly into the formal semantic account of meaning, with an internalist compositional core embedded within an outward-looking externalist representation, I have referred to the framework outlined above as the Symbol-within-Reference model of meaning composition. This is different from the standard classical semantic formalism where the most primitive denotations are couched in terms of objectual referents and truth, while intensional concepts have to be built by generalizing these denotations over possible worlds. The present model inverts the primitive-derived distinction found there, and structures the system around a primitive abstract and essence-like core with referential objectual and instantiational facts built up on top of that.Footnote 10 When it comes to holistic meaning, there is an additional layer of downstream inferencing based on whole sentence meaning, and conversational assumptions, but this is not directly my concern here.
A reviewer rightly asks how a model such as the Symbol-within-Reference model helps us with the problem of explicating the truth conditions of sentences containing the progressive, for example. For example, the present model allows us to state the relationship between ing forms and root forms of the same verbal symbol in terms of a ‘criterial state with the ‘same’ cognitive recognizers’, thus defusing the immediate imperfective paradox. However, we still do need an explanation of why sentences containing accomplishment verbs give rise to a failure of entailment between the progressivized version and the simple past verb, while activity verbs do allow such an entailment. I fully agree that this is part of our responsibility as semanticists, and elsewhere I have discussed how this theory does allow us to have an account of the empirical facts as just stated (Ramchand Reference Ramchand2018). The central idea here is that although the ‘meanings’ of lexical uninflected verbs are not stated in terms of truth conditions directly, propositions that contain them do have truthmakers, and there are principles that systematically relate the use of these lexical items to the truthmakers of the sentences that contain them. For example, the principles of Taylor (Reference Taylor1977) constitute a theory of how stative vs. activity vs. accomplishment lexical verbs produce different kinds of entailments when clothed in temporal information. This theory, plus the claim that the progressive operator creates a derived state, combine to provide an account of why “John is building the house.” does not entail “John built a house” while “John is running.” does seem to entail “John ran.” at least under certain contextual conditions. It is important to underline that the agenda I am pursuing here does not entail an abandonment of the principle of truth-making in explicating what people take to be what propositions ‘mean’. I am proposing that lexical items and certain operations over them are primitive meaning ingredients in a natural language ontology and occur as cognitive precursors to truth-making.
The lesson of nonfinite verbal forms is that language is telling us something quite specific about natural language ontology, namely that the ‘meanings’ of lexical roots must be stated in a way that is neutral with respect to finiteness. Finite inflection is what anchors the use of the verb to a particular event in a particular world at a particular time. Nonfinite verbs very transparently do not have any such direct entailments, and therefore throw up technical complications for semantic theories rooted in extensional denotations. But internal structuring of complex meaning in natural language shows us that we need to have a common meaning ingredient for both finite and nonfinite verbal forms that is simple and primitive. If I am right, that common meaning ingredient must be stated in nonexternalist terms.
10.6 Conclusion
We started on this journey by noticing the repeating paradoxes thrown up by truth theories, exemplified here by the attempt to relate the denotations of nonfinite verbal forms to the finite verbal counterparts. I have insisted that the job of semantics is to understand the ontology of natural language and how it then relates to truth-making. Natural language ontology forces us to acknowledge that meaning ingredients are not directly extensional. This, combined with typological facts about meaning layering in all natural languages, suggests a layered architecture for natural language semantics involving the integration of internalist and externalist-anchored ingredients, mediated by the utterance context.
While there is something important about establishing the connection to truth-making, there is nothing in the logic of how we got here that says we have to state natural language ingredients in terms of functions which take extensional primitives as their inputs and truth values as their final output. The theory we currently have is essentially reverse engineering based on externally verifiable behavior, which denies any role to internal representations. It is akin to constructing a theory of the sound system of human languages by relating objective acoustic primitives of the message to the articulatory primitives that produce that message, without passing through a mode of organization and internal representation in terms of abstract phonemes. In addition, such a theory has no way of modeling the kind of ‘partial content’ that I think actual linguistic forms traffic in. And as Pietroski (Reference Pietroski2018) points out, for natural language, there is no logical necessity for our truth-evaluability to be so direct. Rather, we need to see how sentences constructed systematically by natural language, eventually end up building something that is truth-evaluable, but this does not mean that meanings themselves, the linguistic ingredients of sentence, are extensions.
Finally, this sort of study of nonfinite verbal forms in natural language is important because of how it can inform both linguistics and philosophy. I endorse the position of Fine (Reference Fine2017) and Moltmann (Reference Moltmann2020), who argue that the descriptive work of natural language ontologies is an important stepping stone to both the understanding of the human mind, and for any subsequent theorizing about foundational metaphysics.

