Hostname: page-component-54dcc4c588-hp6zs Total loading time: 0 Render date: 2025-09-21T10:20:59.282Z Has data issue: false hasContentIssue false

REVISIONISM REVISITED

Published online by Cambridge University Press:  03 September 2025

HARTRY FIELD*
Affiliation:
DEPARTMENT OF PHILOSOPHY https://ror.org/0190ak572NEW YORK UNIVERSITY NEW YORK, NY 10012, USA
*
E-mail: hf18@nyu.edu
Rights & Permissions [Opens in a new window]

Abstract

This paper offers a substantial improvement in the revision-theoretic approach to conditionals in theories of transparent truth. The main modifications are (i) a new limit rule; (ii) a modification of the extension to the continuum-valued case; and (iii) the suggestion of a variation on how universal quantification is handled, leading to more satisfactory laws of restricted quantification.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of The Association for Symbolic Logic

1 Background and preview

It seems central to the notion of truth that there is a kind of equivalence between the attribution of truth to a proposition and the proposition itself. Not merely that ‘ $True(\lambda p)$ ’ and ‘p’ are co-assertable, but that one can be substituted for the other (except inside quotation marks, attitude contexts and the like) without affecting assertability. Call this the transparency of truth. But there seem to be propositions that directly or indirectly attribute untruth to themselves.Footnote 1 Given transparency, their truth is equivalent to their untruth, which is inconsistent in classical logic. Given that there are such propositions, we must decide between restricting transparency and restricting classical logic. To make an informed choice, we must explore both classical theories that restrict transparency, and non-classical theories that keep it.Footnote 2 The present paper, like a great deal of recent literature, is part of an exploration of the latter option.

Instead of speaking of propositions, we can speak of sentences as used on a given occasion.Footnote 3 I think this changes little in the theory of truth. (A small advantage is that the slightly contentious assumption of propositions that attribute untruth to themselves is replaced by the harder to deny assumption of sentences that on a given occasion attribute untruth to themselves as used on that occasion. Another advantage is that it enables us to put aside questions of propositional identity, and of how to extend transparency to accommodate it.)Footnote 4 If we focus on sentences rather than propositions, it’s natural to consider toy languages in which there are no ambiguous or indexical sentences, so that we can take sentence-types of the language as bearers of truth. I will follow most of the literature on paradoxes of truth by pretending that sentence-types are the bearers of truth, but what I say could easily be rewritten in terms of propositional truth.Footnote 5

Far and away the most influential paper in the literature on non-classical approaches to truth is [Reference Kripke13] (despite Kripke’s own protests that his paper has nothing to do with non-classical logic). In particular, his fixed point constructions using Strong Kleene semantics have served as the basis for transparent truth in a variety of non-classical logics. (I will assume that the reader knows these constructions. They start from a ground language, without ‘ $True$ ’, in which a theory of syntax (supplying the bearers of truth) can be developed, and classical ground models for this language that are standard for syntax. We extend these models to 3-valued models by adding a transparent truth predicate; different fixed points extend them in different ways.Footnote 6 Sentences true in a ground model M $_{0}$ get value 1 in its 3-valued extensions; sentences false in M $_{0}$ get 0; and value $\frac {1}{2}$ is used for certain sentences involving ‘ $True$ ’, such as Liar sentences that can’t coherently be assigned 0 or 1.) Here I’ll restrict attention to the Strong Kleene quantificational logic K $_{3}$ (extended to include transparent truth by the construction). Call an inference “TK $_{3}$ -valid” if for all fixed point models, if the premises get value 1 in the model, so does the conclusion. (One gets a slightly stronger logic if one restricts to minimal fixed point models.)

The 3-valued Kleene models treat, $\neg $ , $\wedge $ , $\vee $ , $\forall $ and $\exists $ in very natural ways: the value of $\neg A$ is 1 minus the value of A; $|A\wedge B|$ is $min\{|A|,|B|\}$ and $|\forall xA(x)|$ is $min\{|A(o)|: o$ in the domain}; and $\vee $ and $\exists $ are similar but with $max$ instead of $min$ . But there is no connective that serves as a reasonable conditional. (I’m assuming that a reasonable conditional should license the assertion of all instances of $A\rightarrow A$ , $A\wedge B\rightarrow A$ , and the like, but if $\rightarrow $ were defined from $\neg $ and $\vee $ as in classical logic, some instances of these will get the undesignated value $\frac {1}{2}$ .) So it is natural to want to extend Kripke’s theory to a language involving a reasonable conditional $\rightarrow $ in addition to the Kleene connectives and quantifiers plus ‘ $True$ ’. But this requires one or another substantial alteration of Kripke’s methods.Footnote 7

One such extension was proposed in [Reference Brady3]. Brady first noted that from an arbitrary assignment h of values in $\{0,\frac {1}{2},1\}$ to conditional sentences $A\rightarrow B$ in the language with ‘ $True$ ’, we can give what I’ll call a “Kripkean microconstruction” to generate an assignment $v_{h}$ of values to all sentences in that language. (More accurately, h and $v_{h}$ need to assign values to conditional formulas relative to assignments of objects in the domain to their free variables. For brevity’s sake, I’ll be sloppy about this in my formulations; if you like, you can take my ‘sentence’ to mean ‘quasi-sentence’, defined as a pair of a formula and a function assigning objects to its free variables.)Footnote 8 In this construction, conditionals are treated as atomic, with their values given by h, and otherwise the construction is just like Kripke’s minimal fixed point construction. $v_{h}$ will assign the same values to conditionals as h does, and as long as h is transparent in its assignments to conditionals, $v_{h}$ will be transparent in its assignments to all sentences.

So now the question is, how do we choose h? For this, Brady proposed a fixed point “macro-construction”: we start out with an $h_{0}$ that assigns 1 to all conditionals; then at each subsequent $\alpha $ , we let $h_{\alpha }$ assign to $A\rightarrow B$ the value 1 iff $(\forall \beta <\alpha )(v_{h_{\beta }}(A)\leq v_{h_{\beta }}(B))$ , 0 iff $(\exists \beta <\alpha )(v_{h_{\beta }}(A)=1\wedge v_{h_{\beta }}(B)=0)$ , and $\frac {1}{2}$ otherwise.Footnote 9 Because for every conditional, the value assigned to it by $h_{\alpha }$ never increases as $\alpha $ increases, we eventually reach a fixed point $h_{\Psi }$ . The “final value” of a sentence in the Brady construction is its value in the Kripkean fixed point micro-construction over $h_{\Psi }$ .

A Brady conditional has some nice properties, but also a very odd one: conditionals invalidated at early stages (early micro-constructions like that from $h_{0}$ , where the assignment to conditionals is obviously bad) can never recover. For instance, if $\top $ is $\forall x(x=x)$ and $\bot $ is its negation, $h_{0}$ assigns 1 to $\top \rightarrow \bot $ , so $h_{1}$ assigns value 0 to $(\top \rightarrow \bot )\rightarrow \bot $ ; and because at every subsequent stage $h_{\alpha }$ looks back at all earlier stages, every $h_{\alpha }$ assigns 0 to this sentence.

Maybe there are purposes for which we might want a conditional where this isn’t so bad. But one role for conditionals is to restrict universal quantification: to define “All A are B” as “ $\forall x(Ax\rightarrow Bx)$ ”. For this purpose, the Brady conditional seems plainly inadequate: it invalidates the inference from “There are no A” to “All A are B”. More generally, it fails because the $\rightarrow $ fails to reduce to the ordinary $\supset $ in classical contexts, and in classical contexts, restricted universal quantification goes by $\supset $ . It’s the use of conditionals for defining restricted universal quantification that I’ll be considering in this paper. Footnote 10 (If you think that the restricted universal quantifier should be taken as primitive rather than defined in terms of unrestricted quantification and a conditional, you can easily rewrite what follows in terms of extending Kripke by adding a restricted universal quantifier to the language.)

An obvious diagnosis of why the Brady construction has this odd feature is that in defining $h_{\alpha }$ , it looks back at all previous $v_{h_{\beta }}$ ; what if instead we just look at the recent ones? One way of developing this thought is to use a revision construction.Footnote 11 In the simplest version of this, when $\alpha $ is a successor ordinal, we look only at values generated by the previous stage:

$$ \begin{align*} h_{\beta +1}(A\rightarrow B)=1\ \mathrm{iff} \ v_{h_{\beta }}(A)\leq v_{h_{\beta }}(B).\end{align*} $$

There’s more than one possibility for the 0 clause, but I think the best is the revision-theoretic variation of [Reference Brady3] (which models Łukasiewicz 3-valued logic as well as a revision theory can):

$$ \begin{align*}&h_{\beta +1}(A\rightarrow B)=0 \ \mathrm{iff} \ v_{h_{\beta }}(A)=1\wedge v_{h_{\beta }}(B)=0;\\&\mathrm{so}\ \frac {1}{2}\ \mathrm{iff} \ v_{h_{\beta }}(A)-v_{h_{\beta }}(B)=\frac {1}{2}.\end{align*} $$

(In what follows, I’ll use the notation $|A|_{\beta }$ in place of $v_{h_{\beta }}(A)$ .)

But what about 0 and limit ordinals? The starting assignment $h_{0}$ matters only slightly in a 3-valued revision construction: we want it to be transparent, but other than that the values there largely wash out in the end. But the treatment of $h_{\lambda }$ when $\lambda $ is a limit ordinal matters a great deal. And in all my prior writings on this, I made an ill-advised choice.

In my previous writings, I used a rule that like Kripke’s gives a kind of default status to value $\frac {1}{2}$ : in the 3-valued case, it gave value $\frac {1}{2}$ to a conditional at a limit ordinal unless either the limit inferior of prior values was 1 or the limit superior was 0. That is, for limit $\lambda $ ,

$$ \begin{align*} {\mathbf{(SYM_{3})}} h_{\lambda}(A\rightarrow B) = \begin{cases} 1 & \text{iff }(\exists\beta<\alpha)(\forall\gamma\in[\beta,\alpha))(h_{\gamma}(A\rightarrow B)=1)\\ 0 & \text{iff }(\exists\beta<\alpha)(\forall\gamma\in[\beta,\alpha))(h_{\gamma}(A\rightarrow B)=0)\\ \frac{1}{2} & \text{otherwise.} \end{cases} \end{align*} $$

This has some attractive mathematical properties, with which I was much taken: in particular, what I called “the Fundamental Theorem”. But I now think that a different limit rule that does not yield a “Fundamental Theorem” actually leads to a much better theory. The preferred rule (in the 3-valued case, for limit $\lambda $ ) is

$$ \begin{align*} \mathbf{(LimInf_{3})} h_{\lambda}(A\rightarrow B) = \begin{cases} 1 & \text{iff }(\exists\beta<\alpha)(\forall\gamma\in[\beta,\alpha))(h_{\gamma}(A\rightarrow B)=1)\\ 0 & \text{iff }(\forall\beta<\alpha)(\exists\gamma\in[\beta,\alpha))(h_{\gamma}(A\rightarrow B)=0)\\ \frac{1}{2} & \text{otherwise.} \end{cases} \end{align*} $$

In other words: for $r>0$ , $h_{\lambda }(A\rightarrow B)\geq r$ iff $(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ))(h_{\gamma }(A\rightarrow B)\geq r)$ .

The successor and limit cases can be combined: for any $\alpha>0$ and $r>0$ , $h_{\alpha }(A\rightarrow B)\geq r$ iff $(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ))(|A|_{\gamma }-|B|_{\gamma }\leq 1-r)$ . The equivalence to the official definition is trivial in the successor case, and proved by an easy induction in the limit case. (The combined formulation works even for $\alpha =0$ if the starting hypothesis $h_{0}$ assigns 0 to all conditionals. ‘Hypothesis’ here and later will mean ‘assignment of values to conditionals’.)

In §2, I’ll explore the sentential logic that results from adopting (LimInf $_{3}$ ), and say a bit more about the choice of it over (SYM $_{3}$ ). In §3, I’ll discuss the advantages of moving to a continuum-valued context (the advantages are huge, and the discussion of §2 is little affected), and how the revision theory with a generalization of (LimInf $_{3}$ ) should be developed in such a context. (I discussed continuum-valued revision theory with a generalization of (SYM $_{3}$ ) in [Reference Field8], but in addition to the new limit rule, I will make another substantial change.) In §4, I’ll explore the quantifier logic. In §5, I’ll discuss restricted quantification and raise an issue about the interpretation of the English word ‘all’. §6 concludes, with comparisons to earlier work combining conditionals with transparent truth. The Appendices contain some proofs.

2 The sentential logic

A revision construction doesn’t lead to a fixed point. However, it’s easy to see on cardinality grounds that there are recurrent valuations of conditionals, i.e., valuations h that occur arbitrarily late: $\forall \beta (\exists \lambda>\beta )(h_{\gamma }=h)$ . Indeed there comes a point $\alpha _{0}$ (the “critical ordinal”) such that for every $\alpha \geq \alpha _{0}$ , $h_{\alpha }$ is recurrent.Footnote 12 A valuation v of sentences generally is called recurrent if it’s $v_{h}$ for some recurrent h. (Or more generally if one doesn’t restrict to the minimal Kripke fixed point: if it is a member of $V_{h}$ (see n. 9) for some recurrent h.) Which valuations are recurrent obviously depends on the ground model. An inference is taken as valid iff for all ground models, it preserves value 1 in all recurrent valuations based on the ground model.

Putting it another way, let the value space V $_{M}$ for a ground model M be the set of functions that assign a value in $\{0,\frac {1}{2},1\}$ to every recurrent hypothesis based on M. Then a revision construction based on ground model M assigns to every sentence A a value $||A||$ in V $_{M}$ . The sole designated value of V $_{M}$ is the constant function 1, assigning 1 to every recurrent hypothesis based on M. Validity is the preservation of this value 1, on all ground models. Equivalence of two sentences is their having the same function as value, on all ground models; in other words, validity of the biconditional between them.

Another feature of revision constructions (at least, those with fixed rules at limits, like (SYM) or (LimInf)) is the existence of strong reflection ordinals: ordinals $\Delta $ so complicated that for every $\beta <\Delta $ , every recurrent hypothesis occurs in the interval $[\beta ,\Delta )$ . It’s clear that on both (SYM) and (LimInf), the value of a conditional is 1 at a strong reflection ordinal $\Delta $ iff it is 1 for all recurrent $h_{\alpha }$ .Footnote 13 That is, $||A\rightarrow B||$ = 1 iff $|A\rightarrow B|_{\Delta }$ = 1. That’s a nice feature.

With (SYM), the value of a conditional is 0 at $\Delta $ iff it is 0 for all recurrent $h_{\alpha }$ ; so the value of the negation of a conditional is 1 at $\Delta $ iff it is 1 for all recurrent $h_{\alpha }$ . So the nice feature for conditionals extends to negations of conditionals. In fact, it can be shown that it extends to every sentence of the language: for every sentence, not just for conditionals, $||A|| = \boldsymbol{1}$ iff $|A|_{\Delta }=1$ . That’s what in earlier various papers, I called the Fundamental Theorem. But it depends on (SYM): indeed, with (LimInf), the value of a conditional is 0 at $\Delta $ iff it is 0 for some recurrent $h_{\alpha }$ ; so the value of the negation of a conditional is 1 whenever it is 1 for some recurrent $h_{\alpha }$ . So for (LimInf), the Fundamental Theorem fails very badly.

And that does have a philosophical cost. Typical classical theories based on supervaluationist and revision-theoretic semantics have a feature sometimes thought odd: a disjunction can be valid even though both disjuncts lead to inconsistency. Indeed, in such classical theories, the disjunction of a Liar sentence and its negation is an example. In revision theories based on (SYM), the Fundamental Theorem rules out there being any such examples. When we move to (LimInf), on the other hand, there are such examples, though in the theory ultimately to be recommended, it happens only with rather arcane disjunctions, like $R\vee \neg R,$ where R is the Restall sentence discussed in §3.

Despite this apparent disadvantage of LimInf, it leads to a much nicer logic, one that more closely approximates classical logic and hence is easier to use in formal reasoning involving the restricted quantifier conditional. Here’s a simple example, but it’s just the tip of the iceberg. Consider the schema

  • Adjunction: $(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C)$ .

This is not valid on (SYM), but is on (LimInf). (The rule form $(A\rightarrow B)\wedge (A\rightarrow C)\vdash A\rightarrow B\wedge C$ is valid on both.) Indeed on (SYM), even the special case

$$ \begin{align*} (\top \rightarrow B)\wedge (\top \rightarrow \neg B)\rightarrow (\top \rightarrow B\wedge \neg B) \end{align*} $$

fails. For suppose $|B|_{\alpha }$ is 1 for odd $\alpha $ , 0 when $\alpha $ is an even successor, and $\frac {1}{2}$ when $\alpha $ is a limit.Footnote 14 Then (SYM) gives $\top \rightarrow B$ and $\top \rightarrow \neg B$ value $\frac {1}{2}$ at limits, but gives $\top \rightarrow B\wedge \neg B$ 0 at limits of form $\lambda +\omega $ , so the displayed conditional has value $\frac {1}{2}$ at any ordinal of form $\lambda +\omega +1$ and hence at any multiple of $\omega ^{2}$ . (With (LimInf) that counterexample fails since $\top \rightarrow B$ and $\top \rightarrow \neg B$ have value 0 at all limit ordinals.) The soundness of Adjunction for (LimInf) is easily verified: I’ll give a proof in a more general setting in Appendix A.

A difference between (SYM) and (LimInf) that has big ramifications is that with (SYM) we have the biconditional

$$ \begin{align*} (\top \rightarrow \neg B)\leftrightarrow \neg (\top \rightarrow B), \end{align*} $$

whereas with (LimInf) we have only the left-to-right direction. (Though even with (LimInf) we have the rule form of the right to left: $\neg (\top \rightarrow B)\vdash \top \rightarrow \neg B$ .) Equivalently, writing $A\sqcap B$ for $\neg (A\rightarrow \neg B)$ , (LimInf) yields $(\top \rightarrow B)\rightarrow (\top \sqcap B)$ but not the converse (since at limit ordinals, $|\top \rightarrow B|_{\lambda }$ is the $liminf$ of prior values of $B,$ whereas $|\top \sqcap B|_{\lambda }$ is the $limsup$ of those values). (SYM) leads to both directions, which might at first seem like an advantage to (SYM); but in fact having the distinction between $\top \rightarrow B$ and $\top \sqcap B$ is crucial to getting approximations to many important classical laws.

For instance, one such law is

  • Classical Permutation: $[A\rightarrow (B\rightarrow C)]\rightarrow [B\rightarrow (A\rightarrow C)]$ .

There’s no way one can get this on any revision construction that doesn’t lead to a fixed point, basically due to the fact that A differs in its depth of conditional embedding on the two sides of the main conditional, and so does B. The obvious idea of how to fix this is to “depth-level it”, replacing the occurrence of A on the left by $\top \rightarrow A$ or $\top \sqcap A$ , and analogously for the occurrence of B on the right. It turns out that on (LimInf), you need $\sqcap $ on the left and $\rightarrow $ on the right:

  • Depth-Leveled Permutation: $[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)]$ .

That this is sound with (LimInf) will be shown in Appendix A. With (SYM) on the other hand, there is no distinction between $\top \sqcap B$ and $\top \rightarrow B$ , and as a result, no depth-leveled version of Permutation is available. (For a counterexample in the 3-valued construction, let A be the $Q_{2}$ of n. 14, C be $\neg A$ , and B be $\top $ .)

I won’t say a lot more about (SYM). With (LimInf), the following derivation system for the sentential logic with ‘ $True$ ’ is sound in the given semantics (and also in the continuum-based semantics to be given later).

  • S1 (Identity). $A\rightarrow A.$

  • S2 ( $\wedge $ -Elim). $A\wedge B\rightarrow A$ , and $A\wedge B\rightarrow B.$

  • S3 (Distribution). $[A\wedge (B\vee C)]\rightarrow [(A\wedge B)\vee (A\wedge C)].$

  • S4 (Double Negation). $\neg \neg A\rightarrow A.$

  • S5 (Contraposition). $(A\rightarrow \neg B)\rightarrow (B\rightarrow \neg A).$

  • S6 (Adjunction). $(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C).$

  • S7a (Depth-Leveled Suffixing). $[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)].$

  • S7b (Depth-Leveled Prefixing). $[\top \rightarrow (A\rightarrow B)]\rightarrow [(C\rightarrow A)\rightarrow (C\rightarrow B)].$

  • S7c. $[\top \rightarrow (A\sqcap B)]\rightarrow [(B\rightarrow C)\rightarrow (A\sqcap C)].$

  • S8 (Depth-Leveled Weakening). $[\top \rightarrow (\neg A\vee B)]\rightarrow [A\rightarrow B].$

  • S9 (Depth-Leveled Permutation). $[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)].$

  • S10. $(\top \rightarrow \neg A)\rightarrow \neg (\top \rightarrow A).$

  • S11 (Depth-Leveled special Łukasiewicz law). $[(A\rightarrow B)\rightarrow (\top \rightarrow B)]\rightarrow [\top \rightarrow ((\top \sqcap A)\vee (\top \sqcap B))].$

  • T-Schema. $True(\langle A\rangle )\leftrightarrow A$ [When we add quantifiers we’ll want to strengthen this and add composition laws.]

  • R1 (Modus Ponens). $A,A\rightarrow B\vdash B.$

  • R2 ( $\wedge $ -Introd). $A,B\vdash A\wedge B.$

  • R3a (Positive Weakening Rule). $B\vdash \top \rightarrow B.$

  • R3b (Negative Weakening Rule). $\neg (\top \rightarrow B)\vdash \neg B.$

  • R4a (No decreases). $(\top \rightarrow B)\rightarrow B\vdash B\rightarrow (\top \rightarrow B).$

  • R4b (No increases). $B\rightarrow (\top \sqcap B)\vdash (\top \sqcap B)\rightarrow B.$

Mostly these require little comment beyond what I’ve given already. R4a and R4b reflect the fact that if a sentence doesn’t reach a fixed point in a revision construction, it must increase at some final ordinals (ordinals greater than the critical ordinal) and decrease at others: if, e.g., it never decreased from $\alpha $ on, but did increase, then the valuation at $\alpha $ couldn’t be recurrent. It easily follows from R4a and b (with S10 and the Prefixing and Suffixing rules below) that their reverse directions are valid and that $(\top \rightarrow B)\leftrightarrow B\vdash (\top \sqcap B)\leftrightarrow B$ and conversely. S11 is depth-leveled versions of the L to R of the Łukasiewicz (and classical) equivalence $[(A\rightarrow B)\rightarrow B]\leftrightarrow [A\vee B]$ . (The system proves depth-level versions of the R to L as well: $[\top \rightarrow (\top \rightarrow A)]\rightarrow [(A\rightarrow B)\rightarrow (\top \rightarrow B)]$ and $[\top \rightarrow (\top \rightarrow B)]\rightarrow [(A\rightarrow B)\rightarrow (\top \rightarrow B)]$ .Footnote 15

What makes the logic more complicated than the Łukasiewicz is the need to depth-level, due to the lack of a general equivalence between $\top \rightarrow B$ and B. (If you were to add that general equivalence as an additional axiom schema, the Łukasiewicz continuum-valued laws as given in [Reference Priest14, pp. 228–229], would be immediate consequences: all those laws are obtained from those here by deleting some occurrences of ‘ $\top \rightarrow $ ’ and ‘ $\top \sqcap $ ’.) Any revision construction (not leading to a fixed point) will have failures of this equivalence, e.g., when A is the Restall sentence discussed below.

What we might hope, though, is that for “most” sentences, even “most” paradoxical ones, the “regularity assumption” $(\top \rightarrow B)\leftrightarrow B$ is valid. (As noted, that’s equivalent to the validity of $(\top \sqcap B)\leftrightarrow B$ , so that could equally serve as a regularity assumption; so could either direction of either of these biconditionals.)Footnote 16 With a revision theory based on 3-valued logic, a great many paradoxical sentence are irregular. But if we move to an analog that replaces $\{0,\frac {1}{2},1\}$ with the unit interval $[0,1]$ , regularity can be assumed for all but the very arcane sentences (ones that employ quantification in a very specific way). In particular, in the continuum-valued semantics, regularity and hence Łukasiewicz continuum-valued laws can be assumed for “quantifier-independent” sentences, in a sense I’ll define in the next section. And since the axioms of [Reference Priest14] are complete for Łukasiewicz semantics (for inferences involving only finitely many premises), this means that if we supplement the system above with a schema stating the regularity of quantifier-independent sentences, the system so supplemented is complete for (finite premise) inferences involving only such sentences.Footnote 17

For now, I just note that the unsupplemented system is powerful enough to derive lots of desirable consequences. Here are a few obvious ones:

  • Suffixing Rule: $A\rightarrow B\vdash (B\rightarrow C)\rightarrow (A\rightarrow C)$ .

Proof. $A\rightarrow B\vdash \top \rightarrow (A\rightarrow B)$ by the Positive Weakening rule, so with Depth-Leveled Suffixing and Modus Ponens, we get the result.

Similarly for the analogous Prefixing Rule; and of course from either one, we get the

  • Transitivity Rule: $A\rightarrow B,B\rightarrow C\vdash A\rightarrow C$ .

Using Contraposition, we get negative versions of Suffixing and Prefixing, e.g., $A\rightarrow B\vdash \neg (A\rightarrow C)\rightarrow \neg (B\rightarrow C)$ . Using this, we can get

  • Intersubstitution Law: For any $n\geq 0$ : $(\top \rightarrow ^{n}(A\rightarrow B))\rightarrow [X(A)\rightarrow X(B)],$ where A occurs only positively and of depth n in $X({\ldots})$ ; and similarly with the consequent replaced by $X(B)\rightarrow X(A)$ , if A occurs only negatively and of depth n. ( $\top \rightarrow ^{n}(A\rightarrow B)$ means the result of prefixing $A\rightarrow B$ with ‘ $\top \rightarrow $ n times. This can be explicitly defined for variable n, by recursion.) If the antecedent is strengthened to $(\top \rightarrow ^{n}(A\leftrightarrow B))$ we only need to assume that substitutions are of depth n.

  • Various laws relating $\wedge $ and $\rightarrow $ , such as (i) $[(A\rightarrow B)\wedge (C\rightarrow D)]\rightarrow [A\wedge C\rightarrow B\wedge D]$ and (ii) $(A\rightarrow B)\rightarrow [(A\wedge C)\rightarrow (B\wedge C)]$ .

These would not be valid under (SYM); their proofs rest on Adjunction.

Proof of (i):

Using $A\wedge C\rightarrow A$ and $A\wedge C\rightarrow C$ , the antecedent is easily shown to strongly entail $(A\wedge C\rightarrow B)\wedge (A\wedge C\rightarrow D)$ ; we get $A\wedge C\rightarrow B\wedge D$ by Adjunction.

Proof of (ii):

Take D to be C in (i).

  • Reverse Adjunction: $(A\rightarrow C)\wedge (B\rightarrow C)\rightarrow (A\vee B\rightarrow C)$ .

(This follows from Adjunction using contraposition.) Given this, obvious disjunctive analogs of the laws relating $\wedge $ and $\rightarrow $ are derivable in an analogous fashion.

Note that depth-level prefixing with $C=\top $ is

$$ \begin{align*} \vdash [\top \rightarrow (A\rightarrow B)]\rightarrow [(\top \rightarrow A)\rightarrow (\top \rightarrow B)], \end{align*} $$

which makes the prefix $\top \rightarrow $ “necessity-like”, with a necessity that in this system doesn’t obey the T axiom corresponding to a reflexive accessibility operation. Of course this “necessity” has nothing to do with necessity as normally understood, and for sentences that can be assumed regular, it is totally vacuous.

If we define $ $A$ as $(\top \rightarrow A)\wedge A$ , $ does of course obey the T axiom, and it too is necessity-like under (LimInf):

  • $-Introduction: , and

  • K-law for $: .

($-Introduction is immediate from R3a and R2, and the K-law for $ follows from the K-law for “ $\top \rightarrow $ ”, together with law (i) relating $\wedge $ and $\rightarrow $ . As the dependence on the law relating $\wedge $ and $\rightarrow $ suggests, it isn’t sound under (SYM).) is clearly derivable. Of course when A can be assumed regular, the “necessity” operators $\top \rightarrow $ and $ are vacuous.

  • Strengthened Positive Weakening Rule: $\neg A\vee B\vdash A\rightarrow B.$

Proof. Use R3a and either Depth-Leveled Suffixing or Depth-Leveled Weakening.

  • Strengthened Negative Weakening Rule: $\neg (A\rightarrow B)\vdash A\wedge \neg B$ .

Proof. By Depth-Leveled Weakening, Contraposition, and transitivity rule, $\vdash \neg (A\rightarrow B)\rightarrow \neg [\top \rightarrow (\neg A\vee B)]$ ; so $\neg (A\rightarrow B)\vdash \neg [\top \rightarrow (\neg A\vee B)]$ and so by R3b, $\neg (A\rightarrow B)\vdash \neg (\neg A\vee B)$ , which yields the result.

The Strengthened Positive Weakening rule also yields explosion (it entails $\neg A\vdash A\rightarrow B$ , which yields $A,\neg A\vdash B$ by modus ponens).

The absence of a full weakening axiom $(\neg A\vee B)\rightarrow (A\rightarrow B)$ might seem as if it would make trouble for restricted quantification. However, when we add quantifiers we’ll be able to define an operator , where behaves something like an infinite conjunction of A, $\top \rightarrow A$ , $\top \rightarrow (\top \rightarrow A)$ , and so on. (See third paragraph of §4.) And we’ll be able to derive the law

  • (#) .

This is the secret behind the treatment of restricted quantification that I’ll suggest.

I’ve noted that on sufficient regularity assumptions, all the laws of Łukasiewicz [0,1]-valued sentential logic hold. This enables us to immediately see that we must be able to prove such things as

$$ \begin{align*} &Reg(A),Reg(B)\vdash (A\rightarrow B)\vee (B\rightarrow A), \end{align*} $$

with $Reg(A)$ defined as $(\top \rightarrow A)\leftrightarrow A$ . (One could verify this directly by taking a standard Łukasiewicz derivation of itFootnote 18 and using the regularity assumptions as needed.)

Of course at the moment I’m working in a 3-valued revisionist framework, where relatively few such regularity assumptions are valid; so it may be unclear why I’ve appealed to the Łukasiewicz [0,1]-valued logic. The answer is that nothing in the derivation system exploits the limitation to three values, as we’ll soon see;Footnote 19 and there are strong reasons to avoid such a limitation.

3 Values in [0,1]

In the 3-valued revision theory, there are a great many rather simple sentences B that don’t get constant values: equivalently, neither $(\top \rightarrow B)\rightarrow B$ nor $B\rightarrow (\top \rightarrow B)$ get value 1. This leads to complexities in reasoning with them. One example is the sentence $Q_{2}$ that says of itself that if it’s true it isn’t true. On the LimInf rule in 3-valued revision theory, it eventually goes in $\omega $ -cycles of form $ <0,1,0,1,0,1,\ldots >$ . But it’s natural to try to avoid the cycling. We could do this by going to a 7-valued revision theory with values $\frac {k}{6}$ where $0\leq k\leq 6$ : then we could (with a suitable starting hypothesis) give $Q_{2}$ the value $\frac {2}{3}$ at each stage.Footnote 20 But there are other simple sentences that would still cycle, e.g., a sentence that asserts that if it’s true then so is the Liar sentence: to avoid cycling, we’d need a value space with the value $\frac {3}{4}$ .

An obvious idea for avoiding cycling to the extent possible is to do a revision theory over the unit interval [0,1]. This requires that we generalize the Kripkean micro-constructions, to allow them to take arbitrary values in [0,1], but doing so is easy. If we use minimal Kripke fixed points then at the initial Kripke stage we still assign each $True(t)$ value $\frac {1}{2}$ (relative to an assignment to its free variables) when (relative to that assignment) t denotes (the Gödel number of) a sentence A; at a successor stage $\sigma +1,$ we assign $True(t)$ the value that A gets (via the obvious generalization of the Kleene rules to [0,1]) at stage $\sigma $ ; and at a limit ordinal $\lambda ,$ we assign to $True(t)$ the limit of the values at prior stages. The limit exists, because it’s easily shown that any change of values in the construction is always away from $\frac {1}{2}$ (in the same direction).Footnote 21

We adapt this to the revision theory by allowing the assignments h to conditionals to take on arbitrary values in [0,1]; then the Kripke construction over h yields a [0,1]-valued assignment $v_{h}$ to all sentences in the language.

We now need to generalize the revision-theoretic macro-construction. Let’s defer the issue of the starting hypothesis for the moment. The obvious generalization of the successor stages is:

  • (Suc) ,

where is 1 if $|A|_{\beta }\leq |B|_{\beta }$ , $1-(|A|_{\beta }-|B|_{\beta })$ otherwise. (This is the obvious adaptation of Łukasiewicz continuum-valued semantics to revision theory.) And the obvious generalization of the LimInf rule for limit stages is that $h_{\lambda }(A\rightarrow B)$ is the limit inferior of the $h_{\gamma }(A\rightarrow B)$ as $\gamma $ approaches $\lambda $ . In other words,

  • (LimInf $[0,1]$ ) For limit $\lambda $ , $h_{\lambda }(A\rightarrow B)\geq r$ iff $(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(h_{\gamma }(A\rightarrow B)\geq r-\epsilon )$ .Footnote 22

I tentatively propose that we use (Suc) and (LimInf $_{[0,1]}$ ), which taken together say in effect that for any $\alpha>0$ , $h_{\alpha }(A\rightarrow B)$ is (where that means the liminf of the sequence of values prior to $\alpha $ ). On this proposal, the derivational system in §2 is still sound, as shown in Appendix A.

But for this proposal to serve its desired purpose of eliminating unnecessary cycling, we need care in selecting the starting valuation $h_{0}$ . For instance, if $h_{0}$ assigns every conditional a value in $\{0,\frac {1}{2},1\}$ , then no sentence will ever get a value outside of $\{0,\frac {1}{2},1\}$ . If it assigns every conditional the value $\frac {p}{q}$ for positive integers p and q with $p<q$ , then no sentence will ever get a value that isn’t a multiple of $\frac {1}{q}$ if q is even, or $\frac {1}{2q}$ if q is odd. This makes it unevident what to choose as the starting valuation, especially given that we need the starting valuation to be transparent.

For this reason, [Reference Field8] proposed a modification of (Suc), that eliminated cycling for a huge variety of sentences even with simple $h_{0}$ like those that assign 0 (or $\frac {1}{2}$ ) to all conditionals. (That paper used a generalization (SYM $_{[0,1]}$ ) of (SYM $_{3}$ ), rather than (LimInf $_{[0,1]}$ ), but the modification would work as well with (LimInf $_{[0,1]}$ ).) The modification was called “slow corrections”. Instead of taking $h_{\beta +1}(A\rightarrow B)$ to be , as (Suc) does, the “slow correction” revision rule takes it to be the average of this and its previous value:

  • (SC) .

Under (SC) (and with either (SYM) or (LimInf)), a great many sentences like $Q_{2}$ reach fixed point values other than 0, $\frac {1}{2}$ and 1 (and do so even by stage $\omega $ ).

As we’ll see, there are some rather arcane sentences that can’t reach fixed points in any revision construction. (A typical example is the Restall sentence R, constructed by diagonalization to be equivalent to the claim $\exists n(True(\langle R\rightarrow ^{n}\bot \rangle )$ , where $R\rightarrow ^{n}\bot $ is the result of prefixing $\bot $ with ‘ $R\rightarrow $ n times. I think all other examples will involve a similar quantification over depth of embedding in the scope of $\rightarrow $ ’s, combined with truth, though I’m not clear just how to make this precise.) In dealing with such arcane sentences, (SC) is much less convenient to work with than (Suc), and leads to somewhat messier laws. So it would be better if we could stick to (Suc) by using a more complicated $h_{0}$ .

One way to do that would be to use a “pre-construction” based on (SC), for the sole purpose of arriving at a starting valuation for the “real construction” based on (Suc). The starting valuation for the preconstruction could be a simple one, e.g., assigning 0 to all conditionals; we could use the assignment of values to conditionals at a reflection ordinal of the pre-construction as the starting value of the real construction with the simpler rule for successors.

It would be nice if this procedure led, at the very least, to all “quantifier-independent” sentences having constant value. (The importance of doing so is discussed at the end of this Section.) More precisely, call a sentence quantifier-dependent if it is in the smallest set X that contains all sentences with quantifiers and also contains all sentences with $True(t)$ where t denotes a member of X. Call it quantifier-independent otherwise.Footnote 23 If all quantifier-independent sentences have constant value in the slow-correction pre-construction, then of course they will retain that value in a real construction that uses that pre-construction for its starting valuation; but I don’t know whether the antecedent is true.

In case it isn’t true, there’s an alternative way to specify a starting valuation for the revision construction based on (Suc), that does lead to constant values for all quantifier-independent sentences (provided that we make a slight alteration in the Kripkean micro-constructions). It uses the Brouwer fixed point theorem. That theorem has previously been used to showFootnote 24 that in a language without quantifiers (but which may have a way to construct self-referential sentences without them),Footnote 25 any ground model (that’s standard for syntax) can be extended to a model for the language of ‘ $True$ ’ satisfying Łukasiewicz sentential logic.Footnote 26 We can adapt that result here, by temporarily treating as atomic all formulas whose main connective is a quantifier, and giving them uninteresting values—say assigning them all value 0 (relative to any assignment of objects to their free variables), or assigning all universals 0 and all existentials 1. Let g be any function assigning values in $[0,1]$ to every sentence. Let $M_{g}$ be the Łukasiewicz sentential model in which (i) values of atomic sentences with predicates other than ‘ $True$ ’ are given in the ground model; (ii) sentences whose main connectives are quantifiers are given their chosen uninteresting values; (iii) sentences of form ‘ $True(t)$ ’ where t doesn’t denote the Gödel number of a sentence are assigned value 0; and (iv) for any sentence x, the value $g(x)$ is assigned to any sentence of form ‘ $True(t)$ ’ where t denotes the Gödel number of x. Let F be the function that, applied to any g, yields the Łukasiewicz valuation function on $M_{g}$ : that is, yields the function that assigns to each sentence its value in $M_{g}$ under those sentential rules (hence assigns the uninteresting values to quantified sentences). It’s evident from those sentential rules that this function is continuous on the product space $[0,1]^{SENT}$ . By the Brouwer theorem, F has fixed points: there are assignments g such that $F(g)=g$ . Pick any such g (there will be many),Footnote 27 and let $h_{0}$ assign to each conditional what this g does.

We’d like that all subsequent $h_{\alpha }$ will agree with $h_{0}$ on all quantifier-independent conditionals (or equivalently, that the valuations they generate will agree with that generated by $h_{0}$ on all quantifier-independent sentences). In that case, all such sentences clearly get constant value in the macro-construction. But to get such agreement, we need to move beyond minimal Kripke fixed points.Footnote 28 For instance, if the Brouwer fixed point function g assigned a value other than $\frac {1}{2}$ to a Truth-Teller sentence U (taken as quantifier-independent: see n. 25), we’d need a non-minimal fixed point that gives it the same value: otherwise $h_{0}$ and $h_{1}$ would disagree on $\top \rightarrow U$ , and there’s no reason to think that the macro-construction would eventually settle down on all quantifier-independent sentences. But the fix is obvious: in the initial stage of the micro-constructions, we still assign $True(t)$ value $\frac {1}{2}$ if t denotes a quantifier-dependent sentence; but if t denotes a quantifier-independent sentence x, $True(t)$ is assigned $g(x)$ for the chosen Brouwerian function g. If we do this, all stages of the macro-construction will agree with g on quantifier-independent sentences, and so they will get constant values.

Invoking the Brouwer theorem in this way gives little information as to what values a fixed point might assign. To the extent that the pre-construction approach works, it can be used to actually determine fixed point values for sentences; so the question of how widely it determines constant values is worth investigating.

But one way or another, we have a revision theory based on (Suc) and (LimInf) that leads to a wide range of constant values. So the [0,1]-based revision theory validates the soundness of not only the derivation system of §2, but also

  • Regularity Schema $(\top \rightarrow A)\leftrightarrow A$ , for quantifier-independent A.

(This could doubtless be extended beyond the quantifier-independent, though I don’t know how far beyond.) And to repeat, adding this to the derivation system is enough to derive Łukasiewicz continuum-valued logic as restricted to such sentences.

4 Quantifiers

I will now expand the derivation system (that of §2 plus the Regularity Schema) by adding quantifier laws. I’ll take $\exists $ as primitive. As we’ll see, there’s an issue about how best to represent the English ‘all’, so I’ll define more than one candidate. I’ll use $\forall ^{0}$ for the familiar $\neg \exists \neg $ , and define $\forall ^{1}xA$ as . (Recall that $ $A$ is $A\wedge (\top \rightarrow A)$ .) So $|\forall ^{1}xA|_{\alpha }\geq r$ iff $(\forall \epsilon>0)(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ])(\forall c\in |M|)(|A(c)|_{\gamma }\geq r-\epsilon )$ . (Note the use of a fully closed interval $[\beta ,\alpha ]$ .) Obviously this valuation rule is very much dependent on using the (LimInf) rule: under (SYM), $ is a very unnatural operator, and so is $\forall ^{1}$ . But in the derivation system to be developed for (LimInf), $\forall ^{1}$ will actually play a more salient role than $\forall ^{0}$ .

Even without the quantifier rules, we can see that $\vdash \forall ^{1}xA\rightarrow \forall ^{0}xAx$ (by definition and $\wedge $ -elimination) and $\forall ^{0}xAx\vdash \forall ^{1}xAx$ (by $-Introduction, based on R3a).

A third candidate for ‘all’ is a quantifier I’ll call $\forall ^{\omega }$ . To define it, I first use $\forall ^{0}$ plus the truth predicate to define as something like an infinite conjunction of A, $\top \rightarrow A$ , $\top \rightarrow (\top \rightarrow A)$ , etc. More exactly, ; equivalently, , where is just A. (As remarked earlier, $\rightarrow ^{n}$ can be explicitly defined by recursion as a function of variable n, so this definition of makes sense.) Semantically, we have

(Note again the use of a fully closed interval $[\beta ,\alpha ]$ , due to the inclusion of n=0 in the definition.)

It’s obvious from this semantics that the inference from A to is valid, as is the sentence for any specific n; though we’ll need quantifier rules before we can discuss their derivability in a finitary system. Similarly for ; its validity reflects the fact that $1+\omega =\omega $ . The fact that $\omega +1\neq \omega $ suggests, correctly, that is a genuine strengthening of ; it might be called .Footnote 29

I’ll take $\forall ^{\omega }xA$ to be . Like $\forall ^{1}$ (and unlike $\forall ^{0})$ , it will play a salient role in the quantifier rules. In the semantics,

$|\forall ^{\omega }xA|_{\alpha }\geq r$ iff $(\forall \epsilon>0)(\exists \beta [\beta +\omega \leq \alpha \wedge (\forall \gamma \in [\beta ,\alpha ])(\forall c\in |M|)(|A(c)|_{\gamma }\geq r-\epsilon )]$ .

Of course $\forall ^{\omega }$ , $\forall ^{1}$ and $\forall ^{0}$ are all equivalent when applied to “ordinary” formulas A, those that can be assumed to be regular.

If you think it odd to have unrestricted quantifiers $\forall $ inequivalent to $\neg \exists \neg $ , it may help to observe that we already have such an inequivalence for restricted quantification: the restricted universal is defined via $\rightarrow $ and the restricted existential via $\wedge $ . (Also, many other non-classical logics, like intuitionist, have $\forall $ stronger than $\neg \exists \neg $ , as is the case with $\forall ^{1}$ and $\forall ^{\omega }$ .)

We use three axioms and a meta-rule, and strengthen the truth rule:

  • Q1 ( $\exists $ -Intro) $B(y)\rightarrow \exists xB(x)$ when no free occurrence of x in $B(x)$ is in the scope of an $\exists y.$

  • Q2 (Central Quantifier Rule) $\forall ^{1}x(Bx\rightarrow C)\rightarrow (\exists xBx\rightarrow C)$ when x not free in $C.$

  • Q3 (InfDist) $C\wedge \exists xBx\rightarrow \exists x(C\wedge Bx).$

  • Metarule (UnivGen): If $A_{1},{\ldots},A_{n}\vdash Bx$ and x isn’t free in any of $A_{1},{\ldots},A_{n}$ then $A_{1},{\ldots},A_{n}\vdash \forall ^{\omega }xBx.$

We generalize the truth schema, and add composition principles. In these cases, it won’t matter which version of the universal quantifier to use (as we’ll see), so I write them without a superscript.

  • Generalized Truth Schema $(\forall t_{1},{\ldots}t_{n})[TERM(t_{1})\wedge {\ldots}\wedge TERM(t_{n})\supset $ $ [True(\langle A(t_{1},{\ldots},t_{n})\rangle )\leftrightarrow A(den(t_{1}),{\ldots},den(t_{n}))]]$

  • Generalized Composition Schema for $\rightarrow $ : $(\forall x)(\forall y)[QSENT(x)\wedge QSENT(y)\supset [True(x\dot {\rightarrow }y)\leftrightarrow (True(x)\rightarrow True(y))]$ ],

with analogous composition rules for other connectives. (‘ $QSENT$ ’ means ‘quasi-sentence’: see the text to which n. 8 is attached.) Given the compositional rules, it’s obviously enough to postulate the generalized truth schema for atomic A.

Here and in §2, I’ve written most of the axioms and rules schematically, but now that we have quantifiers and transparent truth, we should really use generalized versions. For instance, we should replace S1 with the generalization $\forall x[QSENT(x)\supset True(x\dot {\rightarrow }x)]$ . (Again, it makes little difference which version of the quantifier we use.) From this, we can easily derive specific instances $A\rightarrow A$ using Q1 and the truth rules (and a bit more if we use $\forall ^{1}$ or $\forall ^{\omega }$ ). We can do this for each of the axiom schemas, except that we need a schema in the vicinity of Q1 as well as the generalizations so as to be able to derive the other schemas from their corresponding generalizations. We also should generalize the rules, e.g., replacing Modus Ponens by $True(x),True(x\dot {\rightarrow }y)\vdash True(y)$ , and supplementing UnivGen as written by “If $True(y_{1}),{\ldots},True(y_{n})\vdash True(z)$ then for every variable x not free in any member of the $y_{i}$ , ”. If this is done, the proofs of the schematic theorems and schematic derived rules are easily transformed to proofs of the corresponding generalizations. For ease of reading, I will continue to write in schematic terms, except in one place in Appendix B, where the generalized form is needed.

Q2 could be replaced by the following pair:

  • Q2a. $\forall ^{1}x(Bx\rightarrow Cx)\rightarrow (\exists xBx\rightarrow \exists xCx).$

  • Q2b. $\exists xB\rightarrow B$ when x isn’t free in $B.$

(Q1 entails the converse of Q2b.)

In Łukasiewicz logic, Q3 is derivable from other quantifier laws, but that seems unlikely here.Footnote 30

Obviously , by definition of plus the truth rules. More generally,

  • -Elim For any n,

(For a special case of the previous is , and it’s not hard to prove the equivalence of to ; see Appendix B.)

And using UnivGen, we get a rule form of the converse:

  • -Introduction .

Proof. $A\vdash A$ , so by UnivGen $A\vdash \forall ^{\omega }xA$ for any x not free in A, i.e., ; and using Q2b (plus contraposition, etc.) we get .

In Q2, on the other hand, the use of $\forall ^{1}$ is essential: the analog with $\forall ^{\omega }$ is weaker than we want, and the analog with $\forall ^{0}$ isn’t sound. For suppose $B(n)$ is $R\rightarrow ^{n}\bot $ , where R is the Restall sentence. (More exactly: $B(n)$ is “either n isn’t a natural number, or n is a natural number and the result of prefixing $\bot $ with ‘ $R\rightarrow $ n times is true”.) Then for each n, and each limit ordinal $\lambda $ and natural numbers k, $|B(n)|_{\lambda +k}$ is 1 iff $n\geq k\geq 1$ , 0 otherwise. Then $|\exists nB(n)|_{\lambda +k}$ is 1 for all $k\geq 1$ (since $|B(k)|_{\lambda +k}$ is 1 when $k\neq 0$ ), and so $|\exists nB(n)\rightarrow \bot |_{\lambda +k}$ is 0 for all $k\geq 1$ , and so $|\exists nB(n)\rightarrow \bot |_{\lambda +\omega }$ is 0. But for each n, $|B(n)\rightarrow \bot |_{\lambda +k+1}$ is 1 if $|B(n)\rightarrow \bot |_{\lambda +k}=0$ , that is, if $k>n$ ; so for each n, $|B(n)\rightarrow \bot |_{\lambda +\omega }$ is 1, so $|\forall ^{0}n(B(n)\rightarrow \bot )|_{\lambda +\omega }$ is 1.

That Q2 as stated is sound in the semantics is proved in Appendix A. (It wouldn’t be sound in a semantics based on (SYM), since it implies the K-law for $.) The soundness of the others is obvious.

Of course we don’t need $\forall ^{1}$ for the rule forms of Q2 and Q2a: $\forall ^{0}x(Bx\rightarrow C)\vdash \exists xBx\rightarrow C$ follows from Q2, because of $-introduction, and similarly for Q2a.

A crucial feature of $\forall ^{\omega }$ (not shared by $\forall ^{1}$ or $\forall ^{0}$ ) is

  • (RQ $_{\rightarrow }$ ) $\forall ^{\omega }x(\neg Ax\vee Bx)\rightarrow \forall ^{\omega }x(Ax\rightarrow Bx)$ .

That this is not only valid, but provable in our system, is shown in Appendix B. I’ll discuss its ramifications in §5.

It’s also useful to introduce n-ary quantifiers, that are candidates for understanding “for all $x_{1}$ ,…, $x_{n}$ , $A(x_{1},{\ldots},x_{n})$ ”. I take $(\forall ^{0}x_{1},{\ldots},x_{n})$ to be defined as a string of n unary $\forall ^{0}$ s: $(\forall ^{0}x_{1},{\ldots},x_{n})A(x_{1},{\ldots},x_{n})$ is just $(\forall ^{0} x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$ . (Similarly for $(\exists x_{1},{\ldots},x_{n})$ .) But I take $(\forall ^{1}x_{1},{\ldots},x_{n})$ to be defined as : that is, $(\forall ^{1}x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$ , which for $n>1$ is weaker than a string of $\forall ^{1}$ s. Strings of more than one $\forall ^{1}$ s will turn out to be rather less useful. Similarly, $(\forall ^{\omega }x_{1},{\ldots},x_{n})$ is to be defined as : that is, $(\forall ^{\omega }x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$ .

Here are some salient facts about what’s derivable in the system, even without using the Regularity Schema. (More detail, with proofs, in Appendix B.) And about laws you might expect but that aren’t sound.

  • Vacuous quantification: If x isn’t free in C, then $\exists xC$ and $\forall ^{0}xC$ are each equivalent to C. But $\forall ^{1}xC$ is equivalent to $ $C$ , and $\forall ^{\omega }xC$ is equivalent to . When $n>1$ and $x_{1}$ isn’t free in A, $(\forall ^{1}x_{1},{\ldots},x_{n})A$ is equivalent to $(\forall ^{1}x_{2},{\ldots},x_{n})A$ (and analogously for $x_{i}$ when $i>1$ ), since the one occurrence of $ remains when the $x_{1}$ is deleted; analogously for $\forall ^{\omega }$ .

  • Change of bound variables: Works just as expected.

  • vs. : We get the analog of the converse Barcan formula for $\forall ^{0}$ , that is, ; but only the rule form of the Barcan formula.Footnote 31

  • vs. , and vs. : Analogously, the converse Barcan holds with $\forall ^{1}$ and $\forall ^{\omega }$ , but the Barcan only in rule form.

  • Commutation of quantifiers: We have $\exists x\exists yAxy\rightarrow \exists y\exists xAxy$ , and similarly when all the $\exists $ ’s are replaced by $\forall ^{0}$ . With $\forall ^{1}$ and $\forall ^{\omega }$ we have only the rule forms, which is part of the reason why the n-ary quantifiers $(\forall ^{1}x_{1},{\ldots},x_{n})$ and $(\forall ^{\omega }x_{1},{\ldots},x_{n})$ are more useful than strings of the corresponding unary quantifiers. With the n-ary quantifiers, permuting the variables in the quantifier is unproblematic.

  • Universal analogs of Q2 (and Q2a): We have $\forall ^{1}x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{0}xBx)$ (and $\forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{0}xAx\rightarrow \forall ^{0}xBx)$ ). (We can’t replace $\forall ^{0}$ in the consequent with $\forall ^{1}$ : see Appendix B.) Similarly $\forall ^{\omega }x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{n}xBx)$ , for any n (where $\forall ^{n}$ is ); and

  • $\exists $ -Importation: If C doesn’t contain x free, $\vdash \exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ .

  • Failure of $\exists $ -Exportation Rule: The rule $C\rightarrow \exists xBx\vdash \exists x(C\rightarrow Bx)$ (for C not containing free x) is unsound in the semantics.

That $\exists $ -Exportation fails is a good thing! For (as noted in [Reference Bacon1]) its validity in Łukasiewicz logic is the source of the inconsistency mentioned in n. 26. (Taking C to be the Restall sentence R and $B(n)$ to be $R\rightarrow ^{n}\bot $ , $\exists -$ Exportation (together with transparency) immediately leads to the proof of R; and with that and the compositional rule for $\rightarrow $ , $\neg R$ is easily proved as well.)Footnote 32

These give enough of the flavor of the system to enable us to move to restricted quantification.

5 Restricted quantification

I mentioned early on that my concern in this paper was with a conditional $\rightarrow $ adequate to defining universal restricted quantification: “All A are B” was to be understood as ‘ $\forall x(Ax\rightarrow Bx)$ ’. But we’ve seen at least three candidates for ‘ $\forall $ ’, so three candidates for “All A are B”: we could take it as $\forall ^{0}x(Ax\rightarrow Bx)$ , as $\forall ^{1}x(Ax\rightarrow Bx)$ , or as $\forall ^{\omega }x(Ax\rightarrow Bx)$ . Similarly for multi-place restricted: we could take “All $x_{1}$ …, $x_{n}$ that stand in relation A stand in relation B” as $(\forall ^{0}x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$ , or as $(\forall ^{1}x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$ , or as $(\forall ^{\omega }x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$ . I’ll use the notations $\forall ^{0}x_{Ax}Bx$ , $\forall ^{1}x_{Ax}Bx$ and $\forall ^{\omega }x_{Ax}Bx$ for the unary restricted quantifiers (and analogously for the n-ary).

I think there’s a big advantage to the $\forall ^{\omega }$ readings for both the restricted and unrestricted quantifiers: together these validate the obviously desirable law “If everything is either not-A or B then all A are B”:

  • (RQ) $\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}Bx$ .

(Using an ordinal bigger than $\omega $ for both unrestricted and restricted (see n. 29) would have the same advantage, but there seems little point to that.)

I don’t of course mean this as a claim about ordinary English: ordinary speakers don’t have coherent ways of dealing with truth-theoretic paradoxes, and certainly not of the arcane paradoxes involving Restall-like sentences; and only for such sentences is there any real distinction between $\forall ^{0}$ , $\forall ^{1}$ and $\forall ^{\omega }$ . What I mean, rather, is that by and large, ordinary reasoning is best preserved even for arcanely paradoxical sentences if we understand both the unrestricted and the restricted ‘all’ as $\forall ^{\omega }$ .

If we were to understand $\forall $ in (RQ) as either $\forall ^{0}$ or $\forall ^{1}$ in both occurrences, (RQ) wouldn’t be valid. We could get it by an ad hoc adjustment to the definition of restricted quantification, e.g., taking $\forall x_{Ax}Bx$ not as $\forall x(Ax\rightarrow Bx)$ but as $\forall x[(Ax\rightarrow Bx)\vee (\neg Ax\vee Bx)]$ ; but while that would deliver (RQ) it would fail to deliver related laws, such as

  • (RQ*) $\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}(Ax\wedge Bx)$ .

(For instance, suppose that for any object c, $|Ac|_{\alpha }$ is $\frac {1}{2}$ for all $\alpha $ and $|Bc|_{\alpha }$ is 1 for odd $\alpha $ and 0 for even $\alpha $ . On the proposed ad hoc reading, (RQ*) then amounts to $(\neg Ac\vee Bc)\rightarrow [(Ac\rightarrow Ac\wedge Bc)\vee \neg Ac\vee (Ac\wedge Bc)]$ . But if $\alpha +1$ is odd, then $|\neg Ac\vee Bc|_{\alpha +1}$ is 1 while $|Ac\rightarrow Ac\wedge Bc|_{\alpha +1}$ and $|\neg Ac|_{\alpha +1}$ are $\frac {1}{2}$ and $|Ac\wedge Bc|_{\alpha +1}$ is 0; so this instance of (RQ*) will get value $\frac {1}{2}$ at $\alpha +2$ .) To get laws (RQ) and (RQ*) without going to infinite iterations of $, we need one more $ on the left than on the right: e.g., $\forall ^{1}$ on the left and $\forall ^{0}$ on the right. If we want a single unrestricted ‘ $\forall $ ’, and define restricted quantification in terms of it, we need to use $\forall ^{\omega }$ (or $\forall ^{\alpha }$ for some bigger $\alpha $ ).

Might we do better by taking the restricted quantifier as primitive, and defining the unrestricted in terms of it ( $\forall xBx$ as $\forall x_{x=x}Bx$ )? We’d need valuation rules for the restricted quantifier, but the obvious idea would mimic the revision construction here, evaluating $\forall x_{Ax}Bx$ at an ordinal directly in terms of the values of $Ac$ and $Bc$ at prior ordinals. But in that case, explaining $\forall xBx$ as $\forall x_{x=x}Bx$ would be unacceptable: if the $Bc$ ’s all have high values at ordinals prior to $\alpha $ , $\forall xBx$ would automatically get a high value at $\alpha $ , even if some or all of its instances $Bc$ have low values at $\alpha $ . This can be so even when $\alpha $ is past the critical ordinal; so universal instantiation would fail.

Of course we could get the equivalence of “everything is B” to “every self-identical thing is B” by reading the former as $\forall ^{1}xBx$ and the latter as $\forall ^{0}x_{Ax}Bx$ (or more generally, the former as $\forall ^{n+1}xBx$ and the latter as $\forall ^{n}x_{Ax}Bx$ , for finite n). But that seems rather unnatural, which is why I prefer going to $\forall ^{\omega }$ .

6 Final remarks

I expect that the derivation system presented in §2 and §4 could usefully be expanded (even beyond the added regularity principle suggested at the end of §3; and I expect that that principle too could usefully be extended to a wider class of sentences). A system that is fully complete on this semantics is not in the cards.Footnote 33 But it would certainly be desirable to strengthen the one here; this paper is only a first step.

But I think it is an important step in a number of ways. Probably most centrally, it shows that we can have a theory of transparent truth that has a conditional obeying natural laws and where in particular we can have the law

  • (RQ) $\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}Bx$ ,

reading $\forall x_{Ax}Bx$ as “All A are B”, and taking it to abbreviate $\forall x(Ax\rightarrow Bx)$ . Admittedly, we get (RQ) only by treating the universal quantifier in an unusual way, not as $\neg \exists \neg $ ; but non-classical theories already pretty much required us to treat the universal restricted quantifier as different from $\neg \exists x_{Ax}\neg Bx$ (where $\exists x_{Ax}Cx$ is the restricted existential quantifier $\exists x(Ax\wedge Cx)$ ), so treating the universal unrestricted quantifier as different from $\neg \exists \neg $ doesn’t seem totally outlandish.

Previously, it was reasonable to worry that a theory of transparent truth with (RQ) might be unobtainable (except in trivial ways, like adding ’ as a disjunct, that don’t deliver related laws like $\forall x(\neg Ax\vee Bx)\rightarrow \forall x(Ax\rightarrow Ax\wedge Bx)$ .) Brady’s pioneering papers on conditionals in transparent truth theories, and his book [Reference Brady4], don’t validate even the Weakening Rule

  • (WR) $\forall xBx\vdash \forall x(Ax\rightarrow Bx)$

(and though he didn’t consider varying the semantics of the quantifier, it is pretty clear that there is no remotely acceptable treatment of quantifiers that would give this rule given his treatment of conditionals). In [Reference Bacon2] Bacon modifies Brady’s construction, and says that with his modification, we get not only that rule but a conditional form of it:

  • (WC) $\vdash \forall xBx\rightarrow \forall x(Ax\rightarrow Bx)$ .

But in the main body of his paper (all but the last section), Bacon gets (WC) only in a very limited language, with no proper negation operator—so for instance, it has no means to express ordinary Liar sentences.Footnote 34 The final section of his paper does contain interesting though quite undeveloped suggestions about how to extend his account to a language with a real negation, with a semantics that uses what is essentially the “Routley star” from relevance logic. But though this delivers (WC), it does not deliver (RQ) (which doesn’t follow from (WC) in its logic since its conditional is non-contraposable). It doesn’t deliver even the rule

  • (WCR) $\forall x\neg Ax\vdash \forall x(Ax\rightarrow Bx)$ ;

it takes it as consistent that there are no A’s and yet not all A’s are B’s (e.g., in his preferred version, when $Bx$ is $x\neq x$ , and $Ax$ is $x=x\wedge Q_{2}$ , with $Q_{2}$ as in n. 14).

In earlier work, e.g., [Reference Field7], I made a different sort of attempt to get a law “If everything is either not A or B then all A are B”: I read it not as RQ but rather as

  • (RQ*) $\forall x(\neg Ax\vee Bx)\triangleright \forall x(Ax\rightarrow Bx)$ ,

where $\triangleright $ was a different sort of conditional from $\rightarrow $ : $\rightarrow $ is a restricted quantifier conditional, $\triangleright $ is an ordinary English conditional (perhaps something roughly like the epistemic conditionals in the tradition of Stalnaker 1968). But, while I don’t reject the idea that some pre-theoretic laws about conditionals depend on reading some ‘if…then’s as ordinary English conditionals, the logic that I managed to get by these means was rather weak.

In this paper, I think I’ve achieved much more satisfactory results, and done so without need of a separate conditional $\triangleright $ . Part of this is due to my suggestion of a different reading for ‘all’. But that wouldn’t have led to useful results in a theory based on the revision-theoretic semantics in my earlier papers, because those papers made a poor choice for how the revision procedure behaved at limit ordinals. The earlier part of the present paper shows, I think, that even independent of quantification, the new limit rule (LimInf) is far preferable to the old one; and the treatment of quantifiers in later sections of this paper wouldn’t have made any sense on the old limit rule.

I remarked at the beginning of the paper that we have a choice in how to deal with the paradoxes: we can keep the logic classical and complicate the truth theory, or keep the truth theory simple and complicate the logic. We can’t know which is the better package of logic plus truth theory until we have developed both packages as best we can. The present paper is part of an ongoing effort to do this, focusing especially on the non-classical side. While just a first step, I think that by getting a closer approximation to classical laws, especially laws of restricted quantification, it significantly improves the competitive advantages of the combination of transparent truth and non-classical logic over classical logic with restrictions on transparency. A further step in improving the competitive advantage would involve the addition of predicates “well-behaved” (guaranteeing classical behavior) and “minimally well-behaved” (guaranteeing regularity), as in [Reference Field9].Footnote 35 But the improvement achieved by adding such predicates is somewhat orthogonal to the issues discussed in this paper.

Appendix A: Soundness

The soundness of many of the axioms and rules, with respect to the [0,1]-based revision theory with LimInf, is obvious. I’ll give proofs for the less obvious ones: S6, S7a (S7b is analogous), S7c, S8 (though it too is pretty obvious), S9, S11, R4a (R4b is analogous) and Q2. The system is also sound with respect to the revision theory based on three values (and on 2n+1 values more generally), and the proofs are simpler there. Of course the assumption of regularity for quantifier-independent sentences requires the continuum-based version.

In many of the proofs below, I’ll tacitly rely on a fact noted in the main text, that the value of a conditional $A\rightarrow B$ at a limit ordinal can be characterized either as the $\text {liminf}$ of the values of that conditional at prior ordinals, or as the $\text {liminf}$ of the values at prior ordinals (where these are the values obtained from $|A|_{\gamma }$ and $|B|_{\gamma }$ by the Łukasiewicz evaluation rule). The former “continuity formulation” is what was built into the LimInf rule as I originally stated it; the latter “general rule” has the advantage of holding for successor ordinals as well as for limits. An easy transfinite induction shows their equivalence for limit ordinals. The general rule can be written more explicitly as

  • (GR) For any $\alpha>0$ and $r<1$ , $h_{\alpha }(A\rightarrow B)\geq 1-r$ iff $(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(|A|_{\gamma }-|B|_{\gamma }\leq r+\epsilon )$ .

The equivalence of the two formulations leads immediately to the following lemma.

Lemma A1. $\top \rightarrow (A\rightarrow B)$ has the same value as $A\rightarrow B$ at limit ordinals. (So $|\top \rightarrow (A\rightarrow B)|_{\lambda +1}=|\top \rightarrow (A\rightarrow B)|_{\lambda }$ .)

Proof. By the “general rule”, $|\top \rightarrow (A\rightarrow B)|_{\lambda }$ is , which is $liminf_{\lambda }\{|A\rightarrow B|_{\gamma }\}$ , which by “continuity formulation” is $|A\rightarrow B|_{\lambda }$ .

Lemma A2. For any r: if for some $\beta $ , $|A\rightarrow B|_{\alpha }\geq r$ for all successor ordinals greater than $\beta $ , then for all final ordinals (successor or limit), $|A\rightarrow B|_{\alpha }\geq r$ (where a final ordinal is an ordinal greater than or equal to the critical ordinal).

Proof. The premise guarantees that for all final ordinals $\beta $ , $|A|_{\beta }\leq |B|_{\beta }$ and hence ; so by the “general rule” for evaluating conditionals, the conclusion follows.

  • Soundness of S6: $(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C).$

Proof. By Lemma A2, we need only establish this for successor ordinals $\geq \alpha _{0}+1$ (where $\alpha _{0}$ is the critical ordinal), and that amounts to showing that for all final $\alpha $ , and all reals $r<1 $ , if $|A\rightarrow B|_{\alpha }\geq 1-r$ and $|A\rightarrow C|_{\alpha }\geq 1-r$ then $|A\rightarrow B\wedge C|_{\alpha }\geq 1-r$ . By (GR), the premises tell us that for any $\epsilon>0$ , there are $\beta _{1}$ and $\beta _{2}$ such that $(\forall \gamma \in [\beta _{1},\alpha ))(|A|_{\gamma }-|B|_{\gamma }\leq r+\epsilon )$ and $(\forall \gamma \in [\beta _{2},\alpha )) (|A|_{\gamma }-|C|_{\gamma }\leq r+\epsilon )$ . Letting $\beta $ be their maximum, we get that $(\forall \gamma \in [\beta _{,}\alpha ))(|A|_{\gamma }-min\{|B|_{\gamma },|C|_{\gamma }\}\leq r+\epsilon )$ (since $min\{|B|_{\gamma },|C|_{\gamma }\}$ is either $|B|_{\gamma }$ or $|C|_{\gamma }$ ). So $(\forall \gamma \in [\beta _{,}\alpha ))(|A|_{\gamma }-|B\wedge C|_{\gamma })\leq r+\epsilon $ , and so by (GR) again, $|A\rightarrow B\wedge C|_{\lambda }\geq 1-r$ .

Lemma A3. For any $\alpha>0$ , $|A\rightarrow B|_{\alpha }+|B\rightarrow C|_{\alpha }\leq 1+|A\rightarrow C|_{\alpha }$ . (So if $|B\rightarrow C|_{\alpha }=1$ , $|A\rightarrow B|_{\alpha }\leq |A\rightarrow C|_{\alpha }$ .)

[We obviously need the ‘ $1+$ ’: e.g., if A implies B and B implies C, both terms on left are 1.]

Proof. Using the well-known inequality $liminf\{s_{n}+t_{n}\}\geq liminf\{s_{n}\}+ liminf\{t_{n}\}$ , we get that the left-hand side is . And is easily seen to be (they’re equal when neither summand is 1). So the LHS is , and that’s the RHS.

Lemma A4. For any $\alpha>0$ , $|A\sqcap C|_{\alpha }+|A\rightarrow B|_{\alpha }\leq 1+|B\sqcap C|_{\alpha }$ .

Proof. Writing D for $\neg C$ , Lemma A3 gives $|A\rightarrow B|_{\alpha }+|B\rightarrow D|_{\alpha }\leq 1+|A\rightarrow D|_{\alpha }$ . But $|A\sqcap C|_{\alpha }$ is $1-|A\rightarrow D|_{\alpha }$ , and similarly for $|B\sqcap C|_{\alpha }$ , so $|A\rightarrow B|_{\alpha }-|B\sqcap C|_{\alpha }\leq 1-|A\sqcap C|_{\alpha }$ , which is equivalent to the corollary.

  • Soundness of S7a: $[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)].$

Proof. Obviously $|A\rightarrow B|_{\alpha }\leq 1$ , and so by Lemma A3, when $\alpha>0$ , $|A\rightarrow B|_{\alpha }\leq min\{1,1-|B\rightarrow C|_{\alpha }+|A\rightarrow C|_{\alpha }\}$ . So (for any $\alpha>0$ ) $|\top \rightarrow (A\rightarrow B)|_{\alpha +1}\leq |(B\rightarrow C)\rightarrow (A\rightarrow C)|_{\alpha +1}$ . So by Lemma A2, we get that for any final $\alpha $ , $|\top \rightarrow (A\rightarrow B)|_{\alpha }\leq |(B\rightarrow C)\rightarrow (A\rightarrow C)|_{\alpha }$ . So for any final successor ordinal, $|[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)]|_{\alpha }=1$ . So by Lemma A2 again, $|[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)]|_{\alpha }=1$ for any final ordinal.

  • Soundness of S7c: Analogous to proof for S7a, but using Lemma A4 instead of A3.

  • Soundness of S8: $[\top \rightarrow (\neg A\vee B)]\rightarrow [A\rightarrow B].$

Proof. For any $\alpha $ , $|\top \rightarrow (\neg A\vee B)|_{\alpha +1}$ is $max\{1-|A|_{\alpha },|B|_{\alpha }\}$ , which is $\leq min\{1,1-|A|_{\alpha }+|B|_{\alpha }\}$ , which is $|A\rightarrow B|_{\alpha +1}$ . Using Lemma A2, we get the result.

  • Soundness of S9: $[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)].$

Proof. We need that for all final $\alpha $ , $|(\top \sqcap A)\rightarrow (B\rightarrow C)|_{\alpha }\leq |(\top \rightarrow B)\rightarrow (A\rightarrow C)|_{\alpha }$ . By Lemma A2, it suffices to show this when $\alpha $ is a successor $\beta +1$ . In that case, what we need is that

  • (*) $|\top \rightarrow B|_{\beta }-|A\rightarrow C|_{\beta }\leq max\{0,|\top \sqcap A|_{\beta }-|B\rightarrow C|_{\beta }\}$ .

We prove (*) by induction on $\beta>0.$

When $\beta $ is itself a successor $\delta +1$ , the LHS of (*) is $B_{\delta }-min\{1,1-A_{\delta }+C_{\delta }\}$ , i.e., $max\{B_{\delta }-1,B_{\delta }-1+A_{\delta }-C_{\delta }\}$ (which may be negative).

And the RHS is $max\{0,A_{\delta }-min\{1,1-B_{\delta }+C_{\delta }\}\}$ , i.e., $max\{0,A_{\delta }-1,A_{\delta }-1+B_{\delta }-C_{\delta }\}$ .

If the LHS is $B_{\delta }-1$ , then it’s $\leq 0$ , so $\leq $ RHS. And if it’s $\leq B_{\delta }-1+A_{\delta }-C_{\delta }$ then its also $\leq $ RHS, since that’s equivalent to the last term on right-hand side by rearrangement.

When $\beta $ is a limit $\lambda $ , we invoke induction hypothesis: for all $\gamma <\lambda $ , (*) holds, or rewriting it, $|A\rightarrow C|_{\gamma }\geq |\top \rightarrow B|_{\gamma }-max\{0,|\top \sqcap A|_{\gamma }-|B\rightarrow C|_{\gamma }\}$ , i.e., $|A\rightarrow C|_{\gamma }\geq min\{|\top \rightarrow B|_{\gamma },|\top \rightarrow B|_{\gamma }-|\top \sqcap A|_{\gamma }+|B\rightarrow C|_{\gamma }\}$ .

Taking the $liminf_{\lambda }$ of both sides, we get

$$ \begin{align*}|A\rightarrow C|_{\lambda }\geq liminf_{\lambda }\{min\{|\top \rightarrow B|_{\gamma },|\top \rightarrow B|_{\gamma }-|\top \sqcap A|_{\gamma }+|B\rightarrow C|_{\gamma }\}\}; \end{align*} $$

and since the $liminf$ of the minimum of two sequences is the minimum of their $liminf$ s,

$$ \begin{align*}|A\rightarrow C|_{\lambda }\geq min\{|\top \rightarrow B|_{\lambda },liminf_{\lambda }\{|\top \rightarrow B|_{\gamma }-|\top \sqcap A|_{\gamma }+|B\rightarrow C|_{\gamma }\}. \end{align*} $$

But since $liminf(-y_{\gamma })=-limsup(y_{\gamma })$ , we have

$$ \begin{align*}liminf(x_{\gamma }-y_{\gamma }+z_{\gamma })\geq liminf(x_{\gamma })-limsup(y_{\gamma })+liminf(z_{\gamma }); \end{align*} $$

so the above yields

$$ \begin{align*}|A\rightarrow C|_{\lambda }\geq min\{|\top \rightarrow B|_{\lambda },|\top \rightarrow B|_{\lambda }-|\top \sqcap A|_{\lambda }+|B\rightarrow C|_{\lambda }\}, \end{align*} $$

which means

$$ \begin{align*}|A\rightarrow C|_{\lambda }-|\top \rightarrow B|_{\lambda }\geq min\{0,|B\rightarrow C|_{\lambda }-|\top \sqcap A|_{\lambda }\}. \end{align*} $$

Taking the negatives of both sides, we get (*) for $\lambda $ .

  • Soundness of S11: $[(A\rightarrow B)\rightarrow (\top \rightarrow B)]\rightarrow [\top \rightarrow ((\top \sqcap A)\vee (\top \sqcap B))].$

Proof. By Lemma A2, it suffices to show that at successor ordinals, the LHS has value $\leq $ the RHS; that is, that for all $\beta $ ,

$$ \begin{align*}1-|A\rightarrow B|_{\beta }+|\top \rightarrow B|_{\beta }\leq |\top \sqcap (A\vee B)|_{\beta }. \end{align*} $$

(The fact that $|A\rightarrow B|_{\beta }\geq |\top \rightarrow B|_{\beta }$ avoids any need to minimize the LHS with 1. On the RHS, I’ve used $\top \sqcap (A\vee B)$ instead of the equivalent $(\top \sqcap A)\vee (\top \sqcap B)$ , to simplify the proof of the limit stage of the induction below.)

If $\beta $ is itself a successor $\delta +1$ , the LHS is $1-min\{1,1-|A|_{\delta }+|B|_{\delta }\}+|B|_{\delta }$ ; that is, $max\{B_{\delta },A_{\delta }$ }. That’s the value of the RHS.

When $\beta $ is a limit $\lambda $ we use induction. We can rewrite the inequality as $1-|\top \sqcap (A\vee B)|_{\beta }+|\top \rightarrow B|_{\beta }\leq |A\rightarrow B|_{\beta }$ . Suppose this holds prior to $\lambda $ . Taking $liminf$ s (at $\lambda $ ), we have $liminf\{1-|\top \sqcap (A\vee B)|_{\beta }+|\top \rightarrow B|_{\beta }\}\leq |A\rightarrow B|_{\lambda }$ . But by the same liminf law used in proving S9, this yields $1-limsup\{|\top \sqcap (A\vee B)|_{\beta }\}+liminf\{|\top \rightarrow B|_{\beta }\}\leq |A\rightarrow B|_{\lambda }$ ; that is, $1-|\top \sqcap (A\vee B)|_{\lambda }+|\top \rightarrow B|_{\lambda }\leq |A\rightarrow B|_{\lambda }$ , as desired.

  • Soundness of R4a: $(\top \rightarrow B)\rightarrow B\vdash B\rightarrow (\top \rightarrow B).$

Proof. If $|(\top \rightarrow B)\rightarrow B|_{\alpha }$ is 1 for all final $\alpha $ then for all final $\beta $ , $|B|_{\beta }\geq |\top \rightarrow B|_{\beta }$ . From this, it’s easily seen by induction that the sequence $|B|_{\beta }$ is non-decreasing once the critical ordinal is reached. But then it must be constant: if it strictly increased then the valuations at earlier stages wouldn’t be recurrent. So $|B\rightarrow (\top \rightarrow B)|_{\alpha }$ is 1 for all final $\alpha $ .

  • Soundness of Q2: $\forall ^{1}x(Bx\rightarrow C)\rightarrow (\exists xBx\rightarrow C)$ when x not free in $C.$

Proof. It suffices to prove

(1): For successor $\alpha +1$ , $|\forall ^{0}x(Bx\rightarrow C)|_{\alpha +1}\leq |\exists xBx\rightarrow C|_{\alpha +1}$ , and

(2): For limit $\lambda $ , $|\top \rightarrow \forall ^{0}x(Bx\rightarrow C)|_{\lambda }\leq |\exists xBx\rightarrow C|_{\lambda }$ .

(1) is obvious: If $|\exists xBx\rightarrow C|_{\alpha +1}<r$ then $|\exists xBx|_{\alpha }-|C|_{\alpha }>1-r$ , so there’s a c such that $|Bc|_{\alpha }-|C|_{\alpha }>1-r$ , and so for which $|Bc\rightarrow C|_{\alpha +1}<r$ . So $|\forall ^{0}x(Bx\rightarrow C)|_{\alpha +1}<r$ .

(2): If $|\top \rightarrow \forall ^{0}x(Bx\rightarrow C)|_{\lambda }\geq r$ then $(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(\forall c\in |M|)(|Bx\rightarrow C|_{\gamma }\geq r-\epsilon /2)$ ; this holds in particular at successor $\gamma $ , so

$$ \begin{align*}(\forall \epsilon>0)(\forall \delta \in [\beta _{\epsilon},\lambda ))(\forall c\in |M|)(1-|Bc|_{\delta }+|C|_{\delta }\geq r-\epsilon /2). \end{align*} $$

For any $\delta $ , there’s a c such that

$$ \begin{align*} |Bc|_{\delta }\geq |\exists xBx|_{\delta }-\epsilon /2; \end{align*} $$

so $(\forall \epsilon>0)(\forall \delta \in [\beta _{\epsilon},\lambda ))(1-|\exists xBx|_{\delta }+|C|_{\delta }\geq r-\epsilon )$ . So $|\exists xBx\rightarrow C|_{\lambda }\geq r$ .

Appendix B: Quantifier derivations and invalidities

Proofs of most of the sentential laws mentioned in §2 were sketched there, so I’ll just sketch derivations of some of the quantifier laws mentioned but not proved in §4, and a few others. I’ll also elaborate on a few of the invalidities mentioned there.

  • K-law for : .

Proof. This is the one place where we need to use the generalized form of the laws, mentioned early in §4, rather than the schemas. That enables us to turn the schematic proof of the K-law for $, given in §2, into a proof of

(A) .

The next step is to generalize this to

(B) .

Since we have induction in the theory (see n. 6), and the $n=0$ case reduces to the composition law for $\rightarrow $ , it suffices for (B) that we prove

(C) $[B(n)\rightarrow (C(n)\rightarrow D(n))]\rightarrow [B(n+1)\rightarrow (C(n+1)\rightarrow D(n+1))]$ ,

where $B(n)$ is , $C(n)$ is , and $D(n)$ is . Using contraposition, this amounts to . But using $-introduction followed by three applications of (A), we get (C) and hence (B).

Using UnivGen on (B) (we only need the weaker form with $\forall ^{1}$ ) and Q2a, we get

(D) .

By $-Introduction and (A), this gives

(E) .

And then an application of Q2 to the consequent (together with the Suffixing Rule) yields

(F) .

In other words, .

The K-laws obviously entail the following rules (given -Introduction and -Introduction, both proved in the text).

  • K-rules: and .

N.B.: We can’t strengthen the K-law for to . For take A to be $\top $ and B to be the Restall sentence R. Since $|R|_{\alpha }$ is 0 whenever $\alpha $ is a limit ordinal, is 0 for all $\alpha $ ; so is as well. But since $|R|_{\beta +n}$ is 1 for any $\beta $ and positive integer n, $|\top \rightarrow R|_{\beta +n}$ is 1 for any $\beta $ and $n\geq 2$ , so has value 1 at any $\beta +\omega $ (that is, at any limit ordinal that isn’t a multiple of $\omega ^{2}$ ).

  • Crucial feature of and $\forall ^{\omega }$ : (i) and (ii) .

Proof. By definition, the consequent of (i) is $(\forall n)True(\top \rightarrow ^{n}(\top \rightarrow B))$ , which is obviously equivalent to $(\forall n>0)True(\top \rightarrow ^{n}B)$ ; which follows from $(\forall n)True(\top \rightarrow ^{n}B)$ , which is the antecedent. (ii) easily follows from (i).

  • Vacuous quantification: If x isn’t free in C, $\vdash \exists x\neg C\leftrightarrow \neg C$ , by Q2b and Q1, from which we derive $\vdash \forall ^{0}xC\leftrightarrow C$ using contraposition. The K-rules for $ and then yield and .

  • Change of variables: $\vdash \exists xAx\rightarrow \exists yAy$ (when no y in $Ay$ is in the scope of an $\exists x$ ), and analogously for $\forall ^{0}$ , and $\forall ^{1}$ and $\forall ^{\omega }$ .

Proof. $\vdash Ax\rightarrow \exists yAy$ , so by UnivGen and Q2, $\vdash \exists xAx\rightarrow \exists yAy$ . (Recall that UnivGen has been stated in terms of $\forall ^{1}$ .) The result for $\forall ^{0}$ follows from this using contraposition, and the results for $\forall ^{1}$ and $\forall ^{\omega }$ then follow by the K-rules for $ and .

  • Commutation of unary $\exists $ and $\forall ^{0}$ quantifiers: $\vdash \exists x\exists yAxy\rightarrow \exists y\exists xAxy$ , and similarly when all the $\exists $ ’s are replaced by $\forall ^{0}$ .

Proof. $\vdash Axy\rightarrow \exists xAxy$ by Q1. So $\vdash \exists yAxy\rightarrow \exists y\exists xAxy$ by UnivGen on y plus Q2a. So $\vdash \exists x\exists yAxy\rightarrow \exists y\exists xAxy$ by UnivGen on x plus Q2. The result for $\forall ^{0}$ follows from this using contraposition.

  • Commutation within an n-ary $\forall ^{1}$ quantifier or $\forall ^{\omega }$ quantifier: $\vdash (\forall ^{1}x_{1},x_{2},{\ldots}x_{n})A\rightarrow (\forall ^{1}x_{\pi (1)},x_{\pi (2)},{\ldots}x_{\pi (n)})A,$ where $\pi $ is any permutation of $\{1,{\ldots},n\}$ . [Equivalently, $\forall ^{1}x_{1}\forall ^{0}x_{2}{\ldots}\forall ^{0}x_{n}A\rightarrow \forall ^{1}x_{\pi (1)}\forall ^{0}x_{\pi (2)},{\ldots},\forall ^{0}x_{\pi (n)})A.$ ] Analogously for $\forall ^{\omega }$ .

Proof. An n-ary $\forall ^{0}$ quantifier is obviously equivalent to a string of n unary $\forall ^{0}$ quantifiers, so commutation in the string yields commutation within the n-ary $\forall ^{0}$ . The result for the n-ary $\forall ^{1}$ quantifier is then immediate by $-introduction and the K-law for $.

  • Commutation rule for unary $\forall ^{1}$ quantifiers or $\forall ^{\omega }$ quantifiers: $\forall ^{1}x\forall ^{1}yAxy\vdash \forall ^{1}y\forall ^{1}xAxy$ , and analogously for $\forall ^{\omega }$ .

Proof. $\forall ^{1}x\forall ^{1}yAxy\vdash Axy$ . Apply UnivGen first to x, then to y.

N.B.: We can’t strengthen these to conditional form, e.g., to $\forall ^{1}x\forall ^{1}yA(x,y)\rightarrow \forall ^{1}y\forall ^{1}xA(x,y)$ . For let $Axy$ be of form $Bx\wedge y=y$ . Clearly $\forall ^{0}yAxy$ is then equivalent to $\forall ^{0}yBx$ , which is equivalent to $Bx$ , so $\forall ^{1}yAxy$ is equivalent to , so $\forall ^{1}x\forall ^{1}yA(x,y)$ is equivalent to . And clearly $\forall ^{1}xAxy$ is equivalent to $\forall {}^{1}xBx$ , so $\forall ^{1}y\forall ^{1}xA(x,y)$ is equivalent to . So this instance of commutation for $\forall ^{1}$ is equivalent to the Barcan conditional for $\forall ^{1}$ , ruled out in n. 31.

The one-way interchange laws $\forall ^{1}\forall ^{0}\rightarrow \forall ^{0}\forall ^{1}$ and $\forall ^{\omega }\forall ^{0}\rightarrow \forall ^{0}\forall ^{\omega }$ will be derived below.

  • Universal analogs of Q2a:

(1) $\vdash \forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{0}xAx\rightarrow \forall ^{0}xBx)$

(1*) $\forall ^{1}x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{0}xBx)$

(2)

(2*)

(3)

(3*) .

Proofs. (1) is immediate from Q2a, by contraposition of its consequent and the prefixing rule. From (1), the K-rule for $ yields ; the conditional form of K-law for $ yields , and by Transitivity Rule, we get (2). For (3), we use the K-rule for $ $^{\omega }$ followed by that for $ to get ; the “crucial feature” allows replacing the antecedent by $\forall ^{\omega +1}x(Ax\rightarrow Bx)$ , and the conditional form of K-law for then allows replacing the consequent by $\forall ^{\omega }xAx\rightarrow \forall ^{\omega }xBx$ . (1*)–(3*) follow from (1) to (3) by vacuous quantifier rules.

N.B.: We can’t strengthen (2) to $\forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{1}xAx\rightarrow \forall ^{1}xBx)$ . This is evident, since when $Ax$ is $\top $ this is equivalent to $\forall ^{1}x(\top \rightarrow Bx)\rightarrow (\top \rightarrow \forall ^{1}xBx)$ , and that implies the Barcan formula for $\forall ^{1}$ which we’ve seen invalid (n. 31). [The antecedent of that Barcan formula, , is $\forall ^{1}x[Bx\wedge (\top \rightarrow Bx)]$ , which implies $\forall ^{1}xBx\wedge \forall ^{1}x(\top \rightarrow Bx)$ , which by the bad strengthening of (2) would imply $\forall ^{1}xBx\wedge (\top \rightarrow \forall ^{1}xBx)$ , which is the consequent of that Barcan formula.] Similarly for (3).

  • Partial converse of Q2: $\vdash (\exists xBx\rightarrow C)\rightarrow \forall ^{0}x(Bx\rightarrow C)$ , when x not free in $C.$

Proof. $\vdash Bx\rightarrow \exists xBx$ , so by Suffixing Rule, $\vdash (\exists xBx\rightarrow C)\rightarrow (Bx\rightarrow C)$ , so by UnivGen and (1*), $\vdash (\exists xBx\rightarrow C)\rightarrow \forall ^{0}x(Bx\rightarrow C)$ .

Corollary. $\exists xBx\rightarrow C\vdash \forall ^{1}x(C\rightarrow Bx)$ .

  • Quantifiers with disjunction and conjunction:

(4) $\vdash \exists x(Ax\vee Bx)\leftrightarrow \exists xAx\vee \exists xBx$

(5) $\vdash (\forall ^{0}xAx\wedge \forall ^{0}xBx)\rightarrow \forall ^{0}x(Ax\wedge Bx)$

(6) $\vdash (\forall ^{1}xAx\wedge \forall ^{1}xBx)\rightarrow \forall ^{1}x(Ax\wedge Bx)$

(7) $\vdash \forall ^{0}x(C\vee Bx)\rightarrow (C\vee \forall ^{0}xBx)$ , when x isn’t free in $C.$

Proofs. (4) The R to L is easy, using Reverse Adjunction. L to R: $\vdash Ax\rightarrow (\exists xAx\vee \exists xBx)$ , using Q1, $\vee $ -Introduction and transitivity; similarly $\vdash Bx\rightarrow (\exists xAx\vee \exists xBx)$ . So by Reverse Adjunction, $\vdash (Ax\vee Bx)\rightarrow (\exists xAx\vee \exists xBx)$ . So by UnivGen and Q2, $\vdash \exists x(Ax\vee Bx)\rightarrow (\exists xAx\vee \exists xBx)$ .

(5): By (4) plus contraposition.

(6): From (5), use $-Introduction plus K-law for $ to get . Using , conclusion follows.

(7) By Q3, $\vdash \neg C\wedge \exists x\neg Bx\rightarrow \exists x(\neg C\wedge \neg Bx)$ . From this, we easily get $\neg (C\vee \neg \exists x\neg Bx)\rightarrow \exists x\neg (C\vee Bx)$ , i.e., $\neg (C\vee \forall ^{0}xBx)\rightarrow \exists x\neg (C\vee Bx)$ . The result follows by contraposition (plus definition of $\forall ^{0}$ and double negation laws).

  • Three analogs of converse Barcan formula:

(8a) $\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow \forall ^{0}x(\top \rightarrow Bx)$

(8b) (that is, )

(8c) (that is, ).

Proof of (8a): $\vdash \forall ^{0}xBx\rightarrow Bx$ , so by prefixing rule, $\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow (\top \rightarrow Bx)$ . By UnivGen $\vdash \forall ^{1}x[(\top \rightarrow \forall ^{0}xBx)\rightarrow (\top \rightarrow Bx)]$ , and by (1 $\star$ ), $\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow \forall ^{0}x(\top \rightarrow Bx)$ .

The proofs of (8b) and (8c) are analogous, using the K-rules for $ and instead of the prefixing rule.

Rule versions of converses of these (that is, rule-analogs of the Barcan formula) are trivial.

  • One-way interchange laws: $\vdash \forall ^{1}x\forall ^{0}yBxy\rightarrow \forall ^{0}x\forall ^{1}yBxy$ , and analogously with $\forall ^{\omega }$ for $\forall ^{1}$ in both occurrences.

Proof. Immediate from (8b) and (8c), taking $Bx$ to be $\forall ^{0}yBxy$ .

  • Further analogs of converse Barcan:

(9)

(10)

Proof of (9): use $-Introduction and K-law for $ on (8b).

Proof of (10): is , which is equivalent to , i.e., . But by (8b) that entails , which is .

  • Importation: $\vdash \exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ when x isn’t free in C.

Proof. $\vdash Bx\rightarrow \exists xBx$ by Q1, so $(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ by Prefixing rule; then by UnivGen and Q2, $\exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ .

  • (#): .

Proof. S8 gives $\vdash (\top \rightarrow \neg A\vee B)\rightarrow (A\rightarrow B)$ . By K-rule for , , so by the “crucial feature”, we get the result.

  • (RQ): $\forall ^{\omega }x(\neg Ax\vee Bx)\rightarrow \forall ^{\omega }x(Ax\rightarrow Bx)$ .

Proof. S8 gives $\vdash (\top \rightarrow \neg Ax\vee Bx)\rightarrow (Ax\rightarrow Bx)$ . By UnivGen, $\vdash \forall ^{1}x[(\top \rightarrow \neg Ax\vee Bx)\rightarrow (Ax\rightarrow Bx)]$ , and so by (1), $\vdash \forall ^{0}x(\top \rightarrow \neg Ax\vee Bx)\rightarrow \forall ^{0}x(Ax\rightarrow Bx)$ . By the K-rule for and the “crucial feature”, we get the result.

Acknowledgements

I’m grateful to two referees for suggesting useful clarifications.

Footnotes

1 This is sometimes challenged, but for a rebuttal to the challenges see [Reference Field10]. That paper also elaborates some of the points in notes 3 and 5 below.

2 Poincare once argued that Euclidean geometry is so much simpler than alternative geometries that it would never make sense to take seriously a physical theory based on an alternative geometry. There are many today who have the attitude toward classical logic that Poincare had toward Euclidean geometry: keep it no matter what the cost elsewhere. Maybe there are also some who are “Poincarean” in the opposite direction: keep the naive laws of truth and/or property-instantiation come what may. But I think the proper attitude, here as with geometry, is to choose the simplest overall theory (logic and truth rules together). To do that, we must know what the simplest overall theory is, which requires that we develop each one as best we can.

3 A sentence as used on a given occasion can be indeterminate, but instead of regarding that as involving its expressing multiple propositions, we can regard it as expressing an indeterminate proposition, and taking the attribution of truth to an indeterminate proposition p to be indeterminate in just the way that p is.

4 But in the sentential case, the notion of equivalence involved in the definition of transparency requires a bit more explanation: it is something like equivalence modulo the existence of the utterance. (In the propositional case, it’s also modulo their existence, but that might be thought unimportant since propositions “exist necessarily”.)

5 Talk of propositions is sometimes regarded as merely a convenient paraphrase for quantification into sentential position. The paradoxes of truth arise equally in this framework. (If quantification both in and out of intentional contexts is allowed, one can use sentences like $\exists p$ (The person in Room 202 is now saying that p, but not p), said by the person on Room 202. Even if that sort of sentential quantification is disallowed, one can mimic Gödel–Tarski diagonalization; at least, one can if quantification into predicate position as well as sentential position is allowed, and the latter is needed anyway to deal with related issues about properties.) I think that what follows could also be rewritten more laboriously in terms of sentential quantification.

6 Since the extended models merely add a new predicate, they too are standard, which means that a rule form of induction is guaranteed to hold even for predicates that contain ‘ $True$ ’.

7 [Reference Rossi16] improves on Kripke’s construction using basically Kripkean methods (a multi-stage monotonic construction), but its conditional is not “reasonable” in the sense of this paragraph: some instances of $A\rightarrow A$ come out undesignated. Maybe for some purposes that doesn’t make it unreasonable, but as I’ll soon make explicit, my interest in this paper is with conditionals that can be used to restrict universal quantification, and I don’t think it acceptable not to be able to assert in full generality such laws as “All A are A”, “Everything that’s A and B is A”, and so forth.

8 In fact, I’ll take ‘ $True$ ’ to apply to quasi-sentences as well as ordinary sentences, so that the theory to be discussed really treats of satisfaction rather than just truth as normally understood (though truth as normally understood is a special case).

9 If I (following Brady) hadn’t restricted to minimal fixed points, then each h wouldn’t generate a single $v_{h}$ but a large set $V_{h}$ of them; then presumably $h_{\alpha }(A\rightarrow B)$ would be 1 iff for all $\beta <\alpha $ and all v in $V_{h_{\beta }}$ , $v(A)\leq v(B)$ .

10 Restricted existential quantification is of course $\exists x(Ax\wedge Bx)$ , so the two restricted quantifiers aren’t interdefinable except in classical settings.

11 [Reference Gupta and Belnap11] is a well-known discussion of revision theories. But their focus is on revision theories for truth directly, as an alternative to Kripke (and also, revision theories for circular definitions). The application of revision theory to conditionals, using a Kripkean background, has quite a different flavor.

12 These facts, and the existence of reflection ordinals as stated in the paragraph after next, are standard results in revision theory (see, for instance, [Reference Gupta and Belnap11]).

13 A weak reflection ordinal is any ordinal that satisfies this biconditional. The easiest way to prove the existence of weak reflection ordinals is to prove the existence of strong ones.

14 A sentence B which has this feature on (SYM $_{3}$ ) is the sentence $Q_{2}$ that says $True(\langle Q_{2}\rangle )\rightarrow \neg True(\langle Q_{2}\rangle )$ . (A continuum-valued theory would give $Q_{2}$ a constant value, viz., 2/3, but there are other sentences that would have this valuation pattern there.)

15 The second is immediate by the Suffixing Rule below; the first is a special case of Depth-Leveled Suffixing.

The Łukasiewicz (and classical) equivalence in the text entails $[(A\rightarrow B)\rightarrow B]\rightarrow [(B\rightarrow A)\rightarrow A]$ . Here we get an approximation: $[(A\rightarrow B)\rightarrow (\top \rightarrow B)]\rightarrow [(B\rightarrow A)\rightarrow (\top \sqcap A)]$ . This is almost immediate from S11 and S7c.

16 Note that the formulas themselves are not equivalent; only their validity is.

17 The complexity of the revision-theoretic model theory precludes any sound axiomatic system satisfying a more general form of completeness for finite-premise sentential inferences: see [Reference Welch20] (for an earlier version of the revision-theoretic semantics, but the point still applies).

18 For example, [Reference Schechter17, pp. 419–421].

19 Łukasiewicz 3-valued logic is Łukasiewicz continuum-valued plus one extra (and rather unnatural) axiom that limits the values to three; and I haven’t included any analog of that axiom in my derivational system.

20 The 4-valued logic with values $\{0,\frac {1}{3},\frac {2}{3},1\}$ wouldn’t work, because we need the value $\frac {1}{2}$ to accommodate Liar sentences.

21 For a more careful formulation, see [Reference Field8].

22 The simpler

(?) $h_{\lambda }(A\rightarrow B)\geq r$ iff $(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(h_{\gamma }(A\rightarrow B)\geq r)$

is incoherent in the standard real numbers: for instance, if $h_{\gamma }(A\rightarrow B)$ are forever increasing prior to $\lambda $ , then if r is their limit, (?) would say that the value at $\lambda $ is $\geq $ each real less than r, but not $\geq r$ .

23 Defining “Z-independent” for other notions Z is usually more complicated because of the need to exclude occurrences of ‘ $True(x)$ ’, where the range of ‘x’ includes sentences that are Z-dependent. But a quantifier-free sentence can’t contain ‘ $True(x)$ ’ for any variable x.

24 This is spelled out in [Reference Field6, pp. 97–99], but I’m sure the point was well known long before. [Reference Skolem18] used the theorem in a similar way in connection with the axiom of comprehension in set theory, and [Reference Chang5] extends his results, in ways that could doubtless be used to extend the result as stated in the text to partial results in the language with quantifiers, and thus generalize the claim about quantifier-independent sentences made later in this paragraph.

25 For instance using direct naming, or the diagonalization operator of [Reference Smorynski19].

26 There can be no such result for Łukasiewicz quantifier logic: [Reference Restall15] pointed out (using the aforementioned Restall sentence) that adding the truth schema to Łukasiewicz quantifier logic gives an $\omega $ -inconsistent theory; and [Reference Hajek, Paris and Sheperdson12] showed that this becomes a flat out inconsistency if we beef up the truth schema to include the compositional principle for conditionals, that is, $\forall x\forall y[True(cond(x,y))\leftrightarrow (True(x)\rightarrow True(y))]$ . In a revision construction with (Suc) and (LimInf) then (whatever the starting valuation) the Restall sentence eventually gets value 0 at all limit ordinals and 1 at all successor ordinals. (This pattern is impossible for conditionals, but not for existential quantifications of conditionals.)

27 Most obviously, a fixed point g can assign any value to a Truth-teller, and any value $\geq \frac {1}{2}$ to a sentence that says of itself that if it’s true then so is the Truth-teller. For an example with a very different flavor, suppose A is equivalent to $B\rightarrow \neg A$ , B to $A\rightarrow B$ . Some fixed points have $|A|=1$ , $|B|=0$ ; others have $|A|=\frac {1}{2}$ , $|B|=1$ ; these are the only possibilities. (The slow correction pre-constructions that starts from giving all conditionals the same value lead to the latter.)

28 I’m grateful to Quinton Wood for bringing this issue to my attention.

29 Indeed, we could extend this to define a for every ordinal notation in the language. They’d be strictly increasing in strength; no limit of them is definable in the language.

30 In Łukasiewicz, the contraposed form $\forall ^{0}x(C\vee Bx)\rightarrow C\vee \forall ^{0}xBx$ is derived by deriving $\forall ^{0}x[(C\rightarrow Bx)\rightarrow Bx]\rightarrow [(C\rightarrow \forall ^{0}xBx)\rightarrow \forall ^{0}xBx]$ , and using equivalence of each side to disjunction. But the equivalence of each side to disjunction fails here, in the absence of regularity assumptions.

31 You can get counterexamples to the conditional form of the Barcan for $\forall ^{0}$ (that is, of ) even with just values 1 and 0. E.g., with domain the natural numbers, suppose that $B(n)$ is $\neg (R\rightarrow ^{n}\bot )$ , with R the Restall sentence, so that for each n, $|B(n)|_{\lambda +k}$ is 1 iff either $k=0$ or $k>n$ , and 0 otherwise. Then $|B(n)|_{\lambda +\omega }$ is 1 also, so each $B(n)$ goes in cycles of length $\omega $ . $|\forall ^{0}nB(n)|_{\lambda +k}$ is 0 except when k=0, so is 0 for all k. But $|\top \rightarrow B(n)|_{\lambda }$ (= $|\top \rightarrow B(n)|_{\lambda +\omega }$ ) is 1 since that’s the limit of the $|\top \rightarrow B(n)|_{\lambda +k}$ , and $|\top \rightarrow B(n)|_{\lambda +1}$ is 1 since $|B(n)|_{\lambda }$ is 1; so is 1 for all n, so . So at $\lambda +1$ , has higher value than .Variants of this also give counterexamples to Barcan for $\forall ^{1}$ and for $\forall ^{\omega }$ . For $\forall ^{1}$ , let $B(n)$ be $\top \rightarrow \neg (R\rightarrow ^{n}\bot )$ ; for $\forall {}^{\omega }$ , the counterexamples are a bit more complicated, involving sentences that have cycles of length $\omega ^{2}$ .

32 But given regularity assumptions for C and $Bx$ , we get even the conditional form $(C\rightarrow \exists xBx)\rightarrow \exists x(C\rightarrow Bx)$ .

33 (i) It’s well known that the set of sentences valid in Łukasiewicz continuum-valued quantifier semantics is not recursively enumerable, and this system reduces to that on sufficient regularity assumptions. (ii) Even in the sentential case, the complexity is too great, as observed in n. 17.

34 Bacon notes that it does have a hierarchy of pseudo-negation operators $\neg _{n}A=_{df}A\rightarrow ^{n}\bot $ for $n\geq 1$ . But they turn out to be very unlike real negation: e.g., for any A, $\neg _{n}\neg _{n}A$ leads to contradiction, and for any A and $B,$ we can prove $(\neg _{n}\neg _{n}A)\leftrightarrow (\neg _{n}\neg _{n}B)$ .

35 A somewhat analogous “well-behaved” predicate improves the classical theory as well, but the two predicates together mark a bigger improvement in the non-classical case.

References

Bacon, A. (2013a). Curry’s paradox and $\omega$ -inconsistency. Studia Logica, 101, 19.Google Scholar
Bacon, A. (2013b). A new conditional for naive truth theory. Notre Dame Journal of Formal Logic, 54, 87104.Google Scholar
Brady, R. (1983). The simple consistency of a set theory based on the logic CSQ. Notre Dame Journal of Formal Logic, 24, 431439.Google Scholar
Brady, R. (2006). Universal Logic. Stanford: CSLI Publications.Google Scholar
Chang, C. C. (1963). The axiom of comprehension in infinite-valued logic. Mathematica Scandinavica, 13, 930.Google Scholar
Field, H. (2008). Saving Truth from Paradox. Oxford: Oxford University Press.Google Scholar
Field, H. (2016). Indicative conditionals, restricted quantifiers, and naive truth. Review of Symbolic Logic, 9, 181208.Google Scholar
Field, H. (2020). Properties, propositions and conditionals. Australasian Philosophical Review, 4, 112146.Google Scholar
Field, H. (2026). Well-behaved truth . To appear in Notre Dame Journal of Formal Logic.Google Scholar
Field, H. (forthcoming b). “Propositions, properties and paradox.” Manuscript.Google Scholar
Gupta, A., & Belnap, N. (1993). The Revision Theory of Truth. Cambridge: MIT Press.Google Scholar
Hajek, P., Paris, J., & Sheperdson, J. (2000). The liar paradox and fuzzy logic. Journal of Symbolic Logic, 65, 339346.Google Scholar
Kripke, S. (1975). Outline of a theory of truth. Journal of Philosophy, 72, 690716.Google Scholar
Priest, G. (2008). An Introduction to Non-classical Logic. Cambridge: Cambridge University Press.Google Scholar
Restall, G. (1993). Arithmetic and truth in Łukasiewicz’s infinitely valued logic. Logique et Analyse 36, 2538.Google Scholar
Rossi, L. (2016). Adding a conditional to Kripke’s theory of truth. Journal of Philosophical Logic, 45, 485529.Google Scholar
Schechter, E. (2005). Classical and Nonclassical Logics. Princeton: Princeton University Press.Google Scholar
Skolem, T. (1957). Bemerkungen zum Komprehensionsaxiom. Zeitschrift fur Mathematische Logik, 3, 117 Google Scholar
Smorynski, C. (1985). Self-Reference and Modal Logic. New York: Springer-Verlag.Google Scholar
Welch, P. (2010). Ultimate truth vs. stable truth. Review of Symbolic Logic, 1, 126142.Google Scholar