1 Background and preview
It seems central to the notion of truth that there is a kind of equivalence between the attribution of truth to a proposition and the proposition itself. Not merely that ‘
$True(\lambda p)$
’ and ‘p’ are co-assertable, but that one can be substituted for the other (except inside quotation marks, attitude contexts and the like) without affecting assertability. Call this the transparency of truth. But there seem to be propositions that directly or indirectly attribute untruth to themselves.Footnote
1
Given transparency, their truth is equivalent to their untruth, which is inconsistent in classical logic. Given that there are such propositions, we must decide between restricting transparency and restricting classical logic. To make an informed choice, we must explore both classical theories that restrict transparency, and non-classical theories that keep it.Footnote
2
The present paper, like a great deal of recent literature, is part of an exploration of the latter option.
Instead of speaking of propositions, we can speak of sentences as used on a given occasion.Footnote 3 I think this changes little in the theory of truth. (A small advantage is that the slightly contentious assumption of propositions that attribute untruth to themselves is replaced by the harder to deny assumption of sentences that on a given occasion attribute untruth to themselves as used on that occasion. Another advantage is that it enables us to put aside questions of propositional identity, and of how to extend transparency to accommodate it.)Footnote 4 If we focus on sentences rather than propositions, it’s natural to consider toy languages in which there are no ambiguous or indexical sentences, so that we can take sentence-types of the language as bearers of truth. I will follow most of the literature on paradoxes of truth by pretending that sentence-types are the bearers of truth, but what I say could easily be rewritten in terms of propositional truth.Footnote 5
Far and away the most influential paper in the literature on non-classical approaches to truth is [Reference Kripke13] (despite Kripke’s own protests that his paper has nothing to do with non-classical logic). In particular, his fixed point constructions using Strong Kleene semantics have served as the basis for transparent truth in a variety of non-classical logics. (I will assume that the reader knows these constructions. They start from a ground language, without ‘
$True$
’, in which a theory of syntax (supplying the bearers of truth) can be developed, and classical ground models for this language that are standard for syntax. We extend these models to 3-valued models by adding a transparent truth predicate; different fixed points extend them in different ways.Footnote
6
Sentences true in a ground model M
$_{0}$
get value 1 in its 3-valued extensions; sentences false in M
$_{0}$
get 0; and value
$\frac {1}{2}$
is used for certain sentences involving ‘
$True$
’, such as Liar sentences that can’t coherently be assigned 0 or 1.) Here I’ll restrict attention to the Strong Kleene quantificational logic K
$_{3}$
(extended to include transparent truth by the construction). Call an inference “TK
$_{3}$
-valid” if for all fixed point models, if the premises get value 1 in the model, so does the conclusion. (One gets a slightly stronger logic if one restricts to minimal fixed point models.)
The 3-valued Kleene models treat,
$\neg $
,
$\wedge $
,
$\vee $
,
$\forall $
and
$\exists $
in very natural ways: the value of
$\neg A$
is 1 minus the value of A;
$|A\wedge B|$
is
$min\{|A|,|B|\}$
and
$|\forall xA(x)|$
is
$min\{|A(o)|: o$
in the domain}; and
$\vee $
and
$\exists $
are similar but with
$max$
instead of
$min$
. But there is no connective that serves as a reasonable conditional. (I’m assuming that a reasonable conditional should license the assertion of all instances of
$A\rightarrow A$
,
$A\wedge B\rightarrow A$
, and the like, but if
$\rightarrow $
were defined from
$\neg $
and
$\vee $
as in classical logic, some instances of these will get the undesignated value
$\frac {1}{2}$
.) So it is natural to want to extend Kripke’s theory to a language involving a reasonable conditional
$\rightarrow $
in addition to the Kleene connectives and quantifiers plus ‘
$True$
’. But this requires one or another substantial alteration of Kripke’s methods.Footnote
7
One such extension was proposed in [Reference Brady3]. Brady first noted that from an arbitrary assignment h of values in
$\{0,\frac {1}{2},1\}$
to conditional sentences
$A\rightarrow B$
in the language with ‘
$True$
’, we can give what I’ll call a “Kripkean microconstruction” to generate an assignment
$v_{h}$
of values to all sentences in that language. (More accurately, h and
$v_{h}$
need to assign values to conditional formulas relative to assignments of objects in the domain to their free variables. For brevity’s sake, I’ll be sloppy about this in my formulations; if you like, you can take my ‘sentence’ to mean ‘quasi-sentence’, defined as a pair of a formula and a function assigning objects to its free variables.)Footnote
8
In this construction, conditionals are treated as atomic, with their values given by h, and otherwise the construction is just like Kripke’s minimal fixed point construction.
$v_{h}$
will assign the same values to conditionals as h does, and as long as h is transparent in its assignments to conditionals,
$v_{h}$
will be transparent in its assignments to all sentences.
So now the question is, how do we choose h? For this, Brady proposed a fixed point “macro-construction”: we start out with an
$h_{0}$
that assigns 1 to all conditionals; then at each subsequent
$\alpha $
, we let
$h_{\alpha }$
assign to
$A\rightarrow B$
the value 1 iff
$(\forall \beta <\alpha )(v_{h_{\beta }}(A)\leq v_{h_{\beta }}(B))$
, 0 iff
$(\exists \beta <\alpha )(v_{h_{\beta }}(A)=1\wedge v_{h_{\beta }}(B)=0)$
, and
$\frac {1}{2}$
otherwise.Footnote
9
Because for every conditional, the value assigned to it by
$h_{\alpha }$
never increases as
$\alpha $
increases, we eventually reach a fixed point
$h_{\Psi }$
. The “final value” of a sentence in the Brady construction is its value in the Kripkean fixed point micro-construction over
$h_{\Psi }$
.
A Brady conditional has some nice properties, but also a very odd one: conditionals invalidated at early stages (early micro-constructions like that from
$h_{0}$
, where the assignment to conditionals is obviously bad) can never recover. For instance, if
$\top $
is
$\forall x(x=x)$
and
$\bot $
is its negation,
$h_{0}$
assigns 1 to
$\top \rightarrow \bot $
, so
$h_{1}$
assigns value 0 to
$(\top \rightarrow \bot )\rightarrow \bot $
; and because at every subsequent stage
$h_{\alpha }$
looks back at all earlier stages, every
$h_{\alpha }$
assigns 0 to this sentence.
Maybe there are purposes for which we might want a conditional where this isn’t so bad. But one role for conditionals is to restrict universal quantification: to define “All A are B” as “
$\forall x(Ax\rightarrow Bx)$
”. For this purpose, the Brady conditional seems plainly inadequate: it invalidates the inference from “There are no A” to “All A are B”. More generally, it fails because the
$\rightarrow $
fails to reduce to the ordinary
$\supset $
in classical contexts, and in classical contexts, restricted universal quantification goes by
$\supset $
. It’s the use of conditionals for defining restricted universal quantification that I’ll be considering in this paper.
Footnote
10
(If you think that the restricted universal quantifier should be taken as primitive rather than defined in terms of unrestricted quantification and a conditional, you can easily rewrite what follows in terms of extending Kripke by adding a restricted universal quantifier to the language.)
An obvious diagnosis of why the Brady construction has this odd feature is that in defining
$h_{\alpha }$
, it looks back at all previous
$v_{h_{\beta }}$
; what if instead we just look at the recent ones? One way of developing this thought is to use a revision construction.Footnote
11
In the simplest version of this, when
$\alpha $
is a successor ordinal, we look only at values generated by the previous stage:

There’s more than one possibility for the 0 clause, but I think the best is the revision-theoretic variation of [Reference Brady3] (which models Łukasiewicz 3-valued logic as well as a revision theory can):

(In what follows, I’ll use the notation
$|A|_{\beta }$
in place of
$v_{h_{\beta }}(A)$
.)
But what about 0 and limit ordinals? The starting assignment
$h_{0}$
matters only slightly in a 3-valued revision construction: we want it to be transparent, but other than that the values there largely wash out in the end. But the treatment of
$h_{\lambda }$
when
$\lambda $
is a limit ordinal matters a great deal. And in all my prior writings on this, I made an ill-advised choice.
In my previous writings, I used a rule that like Kripke’s gives a kind of default status to value
$\frac {1}{2}$
: in the 3-valued case, it gave value
$\frac {1}{2}$
to a conditional at a limit ordinal unless either the limit inferior of prior values was 1 or the limit superior was 0. That is, for limit
$\lambda $
,

This has some attractive mathematical properties, with which I was much taken: in particular, what I called “the Fundamental Theorem”. But I now think that a different limit rule that does not yield a “Fundamental Theorem” actually leads to a much better theory. The preferred rule (in the 3-valued case, for limit
$\lambda $
) is

In other words: for
$r>0$
,
$h_{\lambda }(A\rightarrow B)\geq r$
iff
$(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ))(h_{\gamma }(A\rightarrow B)\geq r)$
.
The successor and limit cases can be combined: for any
$\alpha>0$
and
$r>0$
,
$h_{\alpha }(A\rightarrow B)\geq r$
iff
$(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ))(|A|_{\gamma }-|B|_{\gamma }\leq 1-r)$
. The equivalence to the official definition is trivial in the successor case, and proved by an easy induction in the limit case. (The combined formulation works even for
$\alpha =0$
if the starting hypothesis
$h_{0}$
assigns 0 to all conditionals. ‘Hypothesis’ here and later will mean ‘assignment of values to conditionals’.)
In §2, I’ll explore the sentential logic that results from adopting (LimInf
$_{3}$
), and say a bit more about the choice of it over (SYM
$_{3}$
). In §3, I’ll discuss the advantages of moving to a continuum-valued context (the advantages are huge, and the discussion of §2 is little affected), and how the revision theory with a generalization of (LimInf
$_{3}$
) should be developed in such a context. (I discussed continuum-valued revision theory with a generalization of (SYM
$_{3}$
) in [Reference Field8], but in addition to the new limit rule, I will make another substantial change.) In §4, I’ll explore the quantifier logic. In §5, I’ll discuss restricted quantification and raise an issue about the interpretation of the English word ‘all’. §6 concludes, with comparisons to earlier work combining conditionals with transparent truth. The Appendices contain some proofs.
2 The sentential logic
A revision construction doesn’t lead to a fixed point. However, it’s easy to see on cardinality grounds that there are recurrent valuations of conditionals, i.e., valuations h that occur arbitrarily late:
$\forall \beta (\exists \lambda>\beta )(h_{\gamma }=h)$
. Indeed there comes a point
$\alpha _{0}$
(the “critical ordinal”) such that for every
$\alpha \geq \alpha _{0}$
,
$h_{\alpha }$
is recurrent.Footnote
12
A valuation v of sentences generally is called recurrent if it’s
$v_{h}$
for some recurrent h. (Or more generally if one doesn’t restrict to the minimal Kripke fixed point: if it is a member of
$V_{h}$
(see n. 9) for some recurrent h.) Which valuations are recurrent obviously depends on the ground model. An inference is taken as valid iff for all ground models, it preserves value 1 in all recurrent valuations based on the ground model.
Putting it another way, let the value space V
$_{M}$
for a ground model M be the set of functions that assign a value in
$\{0,\frac {1}{2},1\}$
to every recurrent hypothesis based on M. Then a revision construction based on ground model M assigns to every sentence A a value
$||A||$
in V
$_{M}$
. The sole designated value of V
$_{M}$
is the constant function 1, assigning 1 to every recurrent hypothesis based on M. Validity is the preservation of this value 1, on all ground models. Equivalence of two sentences is their having the same function as value, on all ground models; in other words, validity of the biconditional between them.
Another feature of revision constructions (at least, those with fixed rules at limits, like (SYM) or (LimInf)) is the existence of strong reflection ordinals: ordinals
$\Delta $
so complicated that for every
$\beta <\Delta $
, every recurrent hypothesis occurs in the interval
$[\beta ,\Delta )$
. It’s clear that on both (SYM) and (LimInf), the value of a conditional is 1 at a strong reflection ordinal
$\Delta $
iff it is 1 for all recurrent
$h_{\alpha }$
.Footnote
13
That is,
$||A\rightarrow B||$
= 1 iff
$|A\rightarrow B|_{\Delta }$
= 1. That’s a nice feature.
With (SYM), the value of a conditional is 0 at
$\Delta $
iff it is 0 for all recurrent
$h_{\alpha }$
; so the value of the negation of a conditional is 1 at
$\Delta $
iff it is 1 for all recurrent
$h_{\alpha }$
. So the nice feature for conditionals extends to negations of conditionals. In fact, it can be shown that it extends to every sentence of the language: for every sentence, not just for conditionals,
$||A|| = \boldsymbol{1}$
iff
$|A|_{\Delta }=1$
. That’s what in earlier various papers, I called the Fundamental Theorem. But it depends on (SYM): indeed, with (LimInf), the value of a conditional is 0 at
$\Delta $
iff it is 0 for some recurrent
$h_{\alpha }$
; so the value of the negation of a conditional is 1 whenever it is 1 for some recurrent
$h_{\alpha }$
. So for (LimInf), the Fundamental Theorem fails very badly.
And that does have a philosophical cost. Typical classical theories based on supervaluationist and revision-theoretic semantics have a feature sometimes thought odd: a disjunction can be valid even though both disjuncts lead to inconsistency. Indeed, in such classical theories, the disjunction of a Liar sentence and its negation is an example. In revision theories based on (SYM), the Fundamental Theorem rules out there being any such examples. When we move to (LimInf), on the other hand, there are such examples, though in the theory ultimately to be recommended, it happens only with rather arcane disjunctions, like
$R\vee \neg R,$
where R is the Restall sentence discussed in §3.
Despite this apparent disadvantage of LimInf, it leads to a much nicer logic, one that more closely approximates classical logic and hence is easier to use in formal reasoning involving the restricted quantifier conditional. Here’s a simple example, but it’s just the tip of the iceberg. Consider the schema
-
Adjunction:
$(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C)$ .
This is not valid on (SYM), but is on (LimInf). (The rule form
$(A\rightarrow B)\wedge (A\rightarrow C)\vdash A\rightarrow B\wedge C$
is valid on both.) Indeed on (SYM), even the special case

fails. For suppose
$|B|_{\alpha }$
is 1 for odd
$\alpha $
, 0 when
$\alpha $
is an even successor, and
$\frac {1}{2}$
when
$\alpha $
is a limit.Footnote
14
Then (SYM) gives
$\top \rightarrow B$
and
$\top \rightarrow \neg B$
value
$\frac {1}{2}$
at limits, but gives
$\top \rightarrow B\wedge \neg B$
0 at limits of form
$\lambda +\omega $
, so the displayed conditional has value
$\frac {1}{2}$
at any ordinal of form
$\lambda +\omega +1$
and hence at any multiple of
$\omega ^{2}$
. (With (LimInf) that counterexample fails since
$\top \rightarrow B$
and
$\top \rightarrow \neg B$
have value 0 at all limit ordinals.) The soundness of Adjunction for (LimInf) is easily verified: I’ll give a proof in a more general setting in Appendix A.
A difference between (SYM) and (LimInf) that has big ramifications is that with (SYM) we have the biconditional

whereas with (LimInf) we have only the left-to-right direction. (Though even with (LimInf) we have the rule form of the right to left:
$\neg (\top \rightarrow B)\vdash \top \rightarrow \neg B$
.) Equivalently, writing
$A\sqcap B$
for
$\neg (A\rightarrow \neg B)$
, (LimInf) yields
$(\top \rightarrow B)\rightarrow (\top \sqcap B)$
but not the converse (since at limit ordinals,
$|\top \rightarrow B|_{\lambda }$
is the
$liminf$
of prior values of
$B,$
whereas
$|\top \sqcap B|_{\lambda }$
is the
$limsup$
of those values). (SYM) leads to both directions, which might at first seem like an advantage to (SYM); but in fact having the distinction between
$\top \rightarrow B$
and
$\top \sqcap B$
is crucial to getting approximations to many important classical laws.
For instance, one such law is
-
Classical Permutation:
$[A\rightarrow (B\rightarrow C)]\rightarrow [B\rightarrow (A\rightarrow C)]$ .
There’s no way one can get this on any revision construction that doesn’t lead to a fixed point, basically due to the fact that A differs in its depth of conditional embedding on the two sides of the main conditional, and so does B. The obvious idea of how to fix this is to “depth-level it”, replacing the occurrence of A on the left by
$\top \rightarrow A$
or
$\top \sqcap A$
, and analogously for the occurrence of B on the right. It turns out that on (LimInf), you need
$\sqcap $
on the left and
$\rightarrow $
on the right:
-
Depth-Leveled Permutation:
$[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)]$ .
That this is sound with (LimInf) will be shown in Appendix A. With (SYM) on the other hand, there is no distinction between
$\top \sqcap B$
and
$\top \rightarrow B$
, and as a result, no depth-leveled version of Permutation is available. (For a counterexample in the 3-valued construction, let A be the
$Q_{2}$
of n. 14, C be
$\neg A$
, and B be
$\top $
.)
I won’t say a lot more about (SYM). With (LimInf), the following derivation system for the sentential logic with ‘
$True$
’ is sound in the given semantics (and also in the continuum-based semantics to be given later).
-
S1 (Identity).
$A\rightarrow A.$
-
S2 (
$\wedge $ -Elim).
$A\wedge B\rightarrow A$ , and
$A\wedge B\rightarrow B.$
-
S3 (Distribution).
$[A\wedge (B\vee C)]\rightarrow [(A\wedge B)\vee (A\wedge C)].$
-
S4 (Double Negation).
$\neg \neg A\rightarrow A.$
-
S5 (Contraposition).
$(A\rightarrow \neg B)\rightarrow (B\rightarrow \neg A).$
-
S6 (Adjunction).
$(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C).$
-
S7a (Depth-Leveled Suffixing).
$[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)].$
-
S7b (Depth-Leveled Prefixing).
$[\top \rightarrow (A\rightarrow B)]\rightarrow [(C\rightarrow A)\rightarrow (C\rightarrow B)].$
-
S7c.
$[\top \rightarrow (A\sqcap B)]\rightarrow [(B\rightarrow C)\rightarrow (A\sqcap C)].$
-
S8 (Depth-Leveled Weakening).
$[\top \rightarrow (\neg A\vee B)]\rightarrow [A\rightarrow B].$
-
S9 (Depth-Leveled Permutation).
$[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)].$
-
S10.
$(\top \rightarrow \neg A)\rightarrow \neg (\top \rightarrow A).$
-
S11 (Depth-Leveled special Łukasiewicz law).
$[(A\rightarrow B)\rightarrow (\top \rightarrow B)]\rightarrow [\top \rightarrow ((\top \sqcap A)\vee (\top \sqcap B))].$
-
T-Schema.
$True(\langle A\rangle )\leftrightarrow A$ [When we add quantifiers we’ll want to strengthen this and add composition laws.]
-
R1 (Modus Ponens).
$A,A\rightarrow B\vdash B.$
-
R2 (
$\wedge $ -Introd).
$A,B\vdash A\wedge B.$
-
R3a (Positive Weakening Rule).
$B\vdash \top \rightarrow B.$
-
R3b (Negative Weakening Rule).
$\neg (\top \rightarrow B)\vdash \neg B.$
-
R4a (No decreases).
$(\top \rightarrow B)\rightarrow B\vdash B\rightarrow (\top \rightarrow B).$
-
R4b (No increases).
$B\rightarrow (\top \sqcap B)\vdash (\top \sqcap B)\rightarrow B.$
Mostly these require little comment beyond what I’ve given already. R4a and R4b reflect the fact that if a sentence doesn’t reach a fixed point in a revision construction, it must increase at some final ordinals (ordinals greater than the critical ordinal) and decrease at others: if, e.g., it never decreased from
$\alpha $
on, but did increase, then the valuation at
$\alpha $
couldn’t be recurrent. It easily follows from R4a and b (with S10 and the Prefixing and Suffixing rules below) that their reverse directions are valid and that
$(\top \rightarrow B)\leftrightarrow B\vdash (\top \sqcap B)\leftrightarrow B$
and conversely. S11 is depth-leveled versions of the L to R of the Łukasiewicz (and classical) equivalence
$[(A\rightarrow B)\rightarrow B]\leftrightarrow [A\vee B]$
. (The system proves depth-level versions of the R to L as well:
$[\top \rightarrow (\top \rightarrow A)]\rightarrow [(A\rightarrow B)\rightarrow (\top \rightarrow B)]$
and
$[\top \rightarrow (\top \rightarrow B)]\rightarrow [(A\rightarrow B)\rightarrow (\top \rightarrow B)]$
.Footnote
15
What makes the logic more complicated than the Łukasiewicz is the need to depth-level, due to the lack of a general equivalence between
$\top \rightarrow B$
and B. (If you were to add that general equivalence as an additional axiom schema, the Łukasiewicz continuum-valued laws as given in [Reference Priest14, pp. 228–229], would be immediate consequences: all those laws are obtained from those here by deleting some occurrences of ‘
$\top \rightarrow $
’ and ‘
$\top \sqcap $
’.) Any revision construction (not leading to a fixed point) will have failures of this equivalence, e.g., when A is the Restall sentence discussed below.
What we might hope, though, is that for “most” sentences, even “most” paradoxical ones, the “regularity assumption”
$(\top \rightarrow B)\leftrightarrow B$
is valid. (As noted, that’s equivalent to the validity of
$(\top \sqcap B)\leftrightarrow B$
, so that could equally serve as a regularity assumption; so could either direction of either of these biconditionals.)Footnote
16
With a revision theory based on 3-valued logic, a great many paradoxical sentence are irregular. But if we move to an analog that replaces
$\{0,\frac {1}{2},1\}$
with the unit interval
$[0,1]$
, regularity can be assumed for all but the very arcane sentences (ones that employ quantification in a very specific way). In particular, in the continuum-valued semantics, regularity and hence Łukasiewicz continuum-valued laws can be assumed for “quantifier-independent” sentences, in a sense I’ll define in the next section. And since the axioms of [Reference Priest14] are complete for Łukasiewicz semantics (for inferences involving only finitely many premises), this means that if we supplement the system above with a schema stating the regularity of quantifier-independent sentences, the system so supplemented is complete for (finite premise) inferences involving only such sentences.Footnote
17
For now, I just note that the unsupplemented system is powerful enough to derive lots of desirable consequences. Here are a few obvious ones:
-
Suffixing Rule:
$A\rightarrow B\vdash (B\rightarrow C)\rightarrow (A\rightarrow C)$ .
Proof.
$A\rightarrow B\vdash \top \rightarrow (A\rightarrow B)$
by the Positive Weakening rule, so with Depth-Leveled Suffixing and Modus Ponens, we get the result.
Similarly for the analogous Prefixing Rule; and of course from either one, we get the
-
Transitivity Rule:
$A\rightarrow B,B\rightarrow C\vdash A\rightarrow C$ .
Using Contraposition, we get negative versions of Suffixing and Prefixing, e.g.,
$A\rightarrow B\vdash \neg (A\rightarrow C)\rightarrow \neg (B\rightarrow C)$
. Using this, we can get
-
Intersubstitution Law: For any
$n\geq 0$ :
$(\top \rightarrow ^{n}(A\rightarrow B))\rightarrow [X(A)\rightarrow X(B)],$ where A occurs only positively and of depth n in
$X({\ldots})$ ; and similarly with the consequent replaced by
$X(B)\rightarrow X(A)$ , if A occurs only negatively and of depth n. (
$\top \rightarrow ^{n}(A\rightarrow B)$ means the result of prefixing
$A\rightarrow B$ with ‘
$\top \rightarrow $ ’ n times. This can be explicitly defined for variable n, by recursion.) If the antecedent is strengthened to
$(\top \rightarrow ^{n}(A\leftrightarrow B))$ we only need to assume that substitutions are of depth n.
-
Various laws relating
$\wedge $ and
$\rightarrow $ , such as (i)
$[(A\rightarrow B)\wedge (C\rightarrow D)]\rightarrow [A\wedge C\rightarrow B\wedge D]$ and (ii)
$(A\rightarrow B)\rightarrow [(A\wedge C)\rightarrow (B\wedge C)]$ .
These would not be valid under (SYM); their proofs rest on Adjunction.
Proof of (i):
Using
$A\wedge C\rightarrow A$
and
$A\wedge C\rightarrow C$
, the antecedent is easily shown to strongly entail
$(A\wedge C\rightarrow B)\wedge (A\wedge C\rightarrow D)$
; we get
$A\wedge C\rightarrow B\wedge D$
by Adjunction.
Proof of (ii):
Take D to be C in (i).
-
Reverse Adjunction:
$(A\rightarrow C)\wedge (B\rightarrow C)\rightarrow (A\vee B\rightarrow C)$ .
(This follows from Adjunction using contraposition.) Given this, obvious disjunctive analogs of the laws relating
$\wedge $
and
$\rightarrow $
are derivable in an analogous fashion.
Note that depth-level prefixing with
$C=\top $
is

which makes the prefix
$\top \rightarrow $
“necessity-like”, with a necessity that in this system doesn’t obey the T axiom corresponding to a reflexive accessibility operation. Of course this “necessity” has nothing to do with necessity as normally understood, and for sentences that can be assumed regular, it is totally vacuous.
If we define $
$A$
as
$(\top \rightarrow A)\wedge A$
, $ does of course obey the T axiom, and it too is necessity-like under (LimInf):
-
$-Introduction:
, and
-
K-law for $:
.
($-Introduction is immediate from R3a and R2, and the K-law for $ follows from the K-law for “
$\top \rightarrow $
”, together with law (i) relating
$\wedge $
and
$\rightarrow $
. As the dependence on the law relating
$\wedge $
and
$\rightarrow $
suggests, it isn’t sound under (SYM).)
is clearly derivable. Of course when A can be assumed regular, the “necessity” operators
$\top \rightarrow $
and $ are vacuous.
-
Strengthened Positive Weakening Rule:
$\neg A\vee B\vdash A\rightarrow B.$
Proof. Use R3a and either Depth-Leveled Suffixing or Depth-Leveled Weakening.
-
Strengthened Negative Weakening Rule:
$\neg (A\rightarrow B)\vdash A\wedge \neg B$ .
Proof. By Depth-Leveled Weakening, Contraposition, and transitivity rule,
$\vdash \neg (A\rightarrow B)\rightarrow \neg [\top \rightarrow (\neg A\vee B)]$
; so
$\neg (A\rightarrow B)\vdash \neg [\top \rightarrow (\neg A\vee B)]$
and so by R3b,
$\neg (A\rightarrow B)\vdash \neg (\neg A\vee B)$
, which yields the result.
The Strengthened Positive Weakening rule also yields explosion (it entails
$\neg A\vdash A\rightarrow B$
, which yields
$A,\neg A\vdash B$
by modus ponens).
The absence of a full weakening axiom
$(\neg A\vee B)\rightarrow (A\rightarrow B)$
might seem as if it would make trouble for restricted quantification. However, when we add quantifiers we’ll be able to define an operator
, where
behaves something like an infinite conjunction of A,
$\top \rightarrow A$
,
$\top \rightarrow (\top \rightarrow A)$
, and so on. (See third paragraph of §4.) And we’ll be able to derive the law
-
(#)
.
This is the secret behind the treatment of restricted quantification that I’ll suggest.
I’ve noted that on sufficient regularity assumptions, all the laws of Łukasiewicz [0,1]-valued sentential logic hold. This enables us to immediately see that we must be able to prove such things as

with
$Reg(A)$
defined as
$(\top \rightarrow A)\leftrightarrow A$
. (One could verify this directly by taking a standard Łukasiewicz derivation of itFootnote
18
and using the regularity assumptions as needed.)
Of course at the moment I’m working in a 3-valued revisionist framework, where relatively few such regularity assumptions are valid; so it may be unclear why I’ve appealed to the Łukasiewicz [0,1]-valued logic. The answer is that nothing in the derivation system exploits the limitation to three values, as we’ll soon see;Footnote 19 and there are strong reasons to avoid such a limitation.
3 Values in [0,1]
In the 3-valued revision theory, there are a great many rather simple sentences B that don’t get constant values: equivalently, neither
$(\top \rightarrow B)\rightarrow B$
nor
$B\rightarrow (\top \rightarrow B)$
get value 1. This leads to complexities in reasoning with them. One example is the sentence
$Q_{2}$
that says of itself that if it’s true it isn’t true. On the LimInf rule in 3-valued revision theory, it eventually goes in
$\omega $
-cycles of form
$ <0,1,0,1,0,1,\ldots >$
. But it’s natural to try to avoid the cycling. We could do this by going to a 7-valued revision theory with values
$\frac {k}{6}$
where
$0\leq k\leq 6$
: then we could (with a suitable starting hypothesis) give
$Q_{2}$
the value
$\frac {2}{3}$
at each stage.Footnote
20
But there are other simple sentences that would still cycle, e.g., a sentence that asserts that if it’s true then so is the Liar sentence: to avoid cycling, we’d need a value space with the value
$\frac {3}{4}$
.
An obvious idea for avoiding cycling to the extent possible is to do a revision theory over the unit interval [0,1]. This requires that we generalize the Kripkean micro-constructions, to allow them to take arbitrary values in [0,1], but doing so is easy. If we use minimal Kripke fixed points then at the initial Kripke stage we still assign each
$True(t)$
value
$\frac {1}{2}$
(relative to an assignment to its free variables) when (relative to that assignment) t denotes (the Gödel number of) a sentence A; at a successor stage
$\sigma +1,$
we assign
$True(t)$
the value that A gets (via the obvious generalization of the Kleene rules to [0,1]) at stage
$\sigma $
; and at a limit ordinal
$\lambda ,$
we assign to
$True(t)$
the limit of the values at prior stages. The limit exists, because it’s easily shown that any change of values in the construction is always away from
$\frac {1}{2}$
(in the same direction).Footnote
21
We adapt this to the revision theory by allowing the assignments h to conditionals to take on arbitrary values in [0,1]; then the Kripke construction over h yields a [0,1]-valued assignment
$v_{h}$
to all sentences in the language.
We now need to generalize the revision-theoretic macro-construction. Let’s defer the issue of the starting hypothesis for the moment. The obvious generalization of the successor stages is:
-
(Suc)
,
where is 1 if
$|A|_{\beta }\leq |B|_{\beta }$
,
$1-(|A|_{\beta }-|B|_{\beta })$
otherwise. (This is the obvious adaptation of Łukasiewicz continuum-valued semantics to revision theory.) And the obvious generalization of the LimInf rule for limit stages is that
$h_{\lambda }(A\rightarrow B)$
is the limit inferior of the
$h_{\gamma }(A\rightarrow B)$
as
$\gamma $
approaches
$\lambda $
. In other words,
-
(LimInf
$[0,1]$ ) For limit
$\lambda $ ,
$h_{\lambda }(A\rightarrow B)\geq r$ iff
$(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(h_{\gamma }(A\rightarrow B)\geq r-\epsilon )$ .Footnote 22
I tentatively propose that we use (Suc) and (LimInf
$_{[0,1]}$
), which taken together say in effect that for any
$\alpha>0$
,
$h_{\alpha }(A\rightarrow B)$
is
(where that means the liminf of the sequence of
values prior to
$\alpha $
). On this proposal, the derivational system in §2 is still sound, as shown in Appendix A.
But for this proposal to serve its desired purpose of eliminating unnecessary cycling, we need care in selecting the starting valuation
$h_{0}$
. For instance, if
$h_{0}$
assigns every conditional a value in
$\{0,\frac {1}{2},1\}$
, then no sentence will ever get a value outside of
$\{0,\frac {1}{2},1\}$
. If it assigns every conditional the value
$\frac {p}{q}$
for positive integers p and q with
$p<q$
, then no sentence will ever get a value that isn’t a multiple of
$\frac {1}{q}$
if q is even, or
$\frac {1}{2q}$
if q is odd. This makes it unevident what to choose as the starting valuation, especially given that we need the starting valuation to be transparent.
For this reason, [Reference Field8] proposed a modification of (Suc), that eliminated cycling for a huge variety of sentences even with simple
$h_{0}$
like those that assign 0 (or
$\frac {1}{2}$
) to all conditionals. (That paper used a generalization (SYM
$_{[0,1]}$
) of (SYM
$_{3}$
), rather than (LimInf
$_{[0,1]}$
), but the modification would work as well with (LimInf
$_{[0,1]}$
).) The modification was called “slow corrections”. Instead of taking
$h_{\beta +1}(A\rightarrow B)$
to be
, as (Suc) does, the “slow correction” revision rule takes it to be the average of this and its previous value:
-
(SC)
.
Under (SC) (and with either (SYM) or (LimInf)), a great many sentences like
$Q_{2}$
reach fixed point values other than 0,
$\frac {1}{2}$
and 1 (and do so even by stage
$\omega $
).
As we’ll see, there are some rather arcane sentences that can’t reach fixed points in any revision construction. (A typical example is the Restall sentence R, constructed by diagonalization to be equivalent to the claim
$\exists n(True(\langle R\rightarrow ^{n}\bot \rangle )$
, where
$R\rightarrow ^{n}\bot $
is the result of prefixing
$\bot $
with ‘
$R\rightarrow $
’ n times. I think all other examples will involve a similar quantification over depth of embedding in the scope of
$\rightarrow $
’s, combined with truth, though I’m not clear just how to make this precise.) In dealing with such arcane sentences, (SC) is much less convenient to work with than (Suc), and leads to somewhat messier laws. So it would be better if we could stick to (Suc) by using a more complicated
$h_{0}$
.
One way to do that would be to use a “pre-construction” based on (SC), for the sole purpose of arriving at a starting valuation for the “real construction” based on (Suc). The starting valuation for the preconstruction could be a simple one, e.g., assigning 0 to all conditionals; we could use the assignment of values to conditionals at a reflection ordinal of the pre-construction as the starting value of the real construction with the simpler rule for successors.
It would be nice if this procedure led, at the very least, to all “quantifier-independent” sentences having constant value. (The importance of doing so is discussed at the end of this Section.) More precisely, call a sentence quantifier-dependent if it is in the smallest set X that contains all sentences with quantifiers and also contains all sentences with
$True(t)$
where t denotes a member of X. Call it quantifier-independent otherwise.Footnote
23
If all quantifier-independent sentences have constant value in the slow-correction pre-construction, then of course they will retain that value in a real construction that uses that pre-construction for its starting valuation; but I don’t know whether the antecedent is true.
In case it isn’t true, there’s an alternative way to specify a starting valuation for the revision construction based on (Suc), that does lead to constant values for all quantifier-independent sentences (provided that we make a slight alteration in the Kripkean micro-constructions). It uses the Brouwer fixed point theorem. That theorem has previously been used to showFootnote
24
that in a language without quantifiers (but which may have a way to construct self-referential sentences without them),Footnote
25
any ground model (that’s standard for syntax) can be extended to a model for the language of ‘
$True$
’ satisfying Łukasiewicz sentential logic.Footnote
26
We can adapt that result here, by temporarily treating as atomic all formulas whose main connective is a quantifier, and giving them uninteresting values—say assigning them all value 0 (relative to any assignment of objects to their free variables), or assigning all universals 0 and all existentials 1. Let g be any function assigning values in
$[0,1]$
to every sentence. Let
$M_{g}$
be the Łukasiewicz sentential model in which (i) values of atomic sentences with predicates other than ‘
$True$
’ are given in the ground model; (ii) sentences whose main connectives are quantifiers are given their chosen uninteresting values; (iii) sentences of form ‘
$True(t)$
’ where t doesn’t denote the Gödel number of a sentence are assigned value 0; and (iv) for any sentence x, the value
$g(x)$
is assigned to any sentence of form ‘
$True(t)$
’ where t denotes the Gödel number of x. Let F be the function that, applied to any g, yields the Łukasiewicz valuation function on
$M_{g}$
: that is, yields the function that assigns to each sentence its value in
$M_{g}$
under those sentential rules (hence assigns the uninteresting values to quantified sentences). It’s evident from those sentential rules that this function is continuous on the product space
$[0,1]^{SENT}$
. By the Brouwer theorem, F has fixed points: there are assignments g such that
$F(g)=g$
. Pick any such g (there will be many),Footnote
27
and let
$h_{0}$
assign to each conditional what this g does.
We’d like that all subsequent
$h_{\alpha }$
will agree with
$h_{0}$
on all quantifier-independent conditionals (or equivalently, that the valuations they generate will agree with that generated by
$h_{0}$
on all quantifier-independent sentences). In that case, all such sentences clearly get constant value in the macro-construction. But to get such agreement, we need to move beyond minimal Kripke fixed points.Footnote
28
For instance, if the Brouwer fixed point function g assigned a value other than
$\frac {1}{2}$
to a Truth-Teller sentence U (taken as quantifier-independent: see n. 25), we’d need a non-minimal fixed point that gives it the same value: otherwise
$h_{0}$
and
$h_{1}$
would disagree on
$\top \rightarrow U$
, and there’s no reason to think that the macro-construction would eventually settle down on all quantifier-independent sentences. But the fix is obvious: in the initial stage of the micro-constructions, we still assign
$True(t)$
value
$\frac {1}{2}$
if t denotes a quantifier-dependent sentence; but if t denotes a quantifier-independent sentence x,
$True(t)$
is assigned
$g(x)$
for the chosen Brouwerian function g. If we do this, all stages of the macro-construction will agree with g on quantifier-independent sentences, and so they will get constant values.
Invoking the Brouwer theorem in this way gives little information as to what values a fixed point might assign. To the extent that the pre-construction approach works, it can be used to actually determine fixed point values for sentences; so the question of how widely it determines constant values is worth investigating.
But one way or another, we have a revision theory based on (Suc) and (LimInf) that leads to a wide range of constant values. So the [0,1]-based revision theory validates the soundness of not only the derivation system of §2, but also
-
Regularity Schema
$(\top \rightarrow A)\leftrightarrow A$ , for quantifier-independent A.
(This could doubtless be extended beyond the quantifier-independent, though I don’t know how far beyond.) And to repeat, adding this to the derivation system is enough to derive Łukasiewicz continuum-valued logic as restricted to such sentences.
4 Quantifiers
I will now expand the derivation system (that of §2 plus the Regularity Schema) by adding quantifier laws. I’ll take
$\exists $
as primitive. As we’ll see, there’s an issue about how best to represent the English ‘all’, so I’ll define more than one candidate. I’ll use
$\forall ^{0}$
for the familiar
$\neg \exists \neg $
, and define
$\forall ^{1}xA$
as
. (Recall that $
$A$
is
$A\wedge (\top \rightarrow A)$
.) So
$|\forall ^{1}xA|_{\alpha }\geq r$
iff
$(\forall \epsilon>0)(\exists \beta <\alpha )(\forall \gamma \in [\beta ,\alpha ])(\forall c\in |M|)(|A(c)|_{\gamma }\geq r-\epsilon )$
. (Note the use of a fully closed interval
$[\beta ,\alpha ]$
.) Obviously this valuation rule is very much dependent on using the (LimInf) rule: under (SYM), $ is a very unnatural operator, and so is
$\forall ^{1}$
. But in the derivation system to be developed for (LimInf),
$\forall ^{1}$
will actually play a more salient role than
$\forall ^{0}$
.
Even without the quantifier rules, we can see that
$\vdash \forall ^{1}xA\rightarrow \forall ^{0}xAx$
(by definition and
$\wedge $
-elimination) and
$\forall ^{0}xAx\vdash \forall ^{1}xAx$
(by $-Introduction, based on R3a).
A third candidate for ‘all’ is a quantifier I’ll call
$\forall ^{\omega }$
. To define it, I first use
$\forall ^{0}$
plus the truth predicate to define
as something like an infinite conjunction of A,
$\top \rightarrow A$
,
$\top \rightarrow (\top \rightarrow A)$
, etc. More exactly,
; equivalently,
, where
is just A. (As remarked earlier,
$\rightarrow ^{n}$
can be explicitly defined by recursion as a function of variable n, so this definition of
makes sense.) Semantically, we have

(Note again the use of a fully closed interval
$[\beta ,\alpha ]$
, due to the inclusion of n=0 in the definition.)
It’s obvious from this semantics that the inference from A to is valid, as is the sentence
for any specific n; though we’ll need quantifier rules before we can discuss their derivability in a finitary system. Similarly for
; its validity reflects the fact that
$1+\omega =\omega $
. The fact that
$\omega +1\neq \omega $
suggests, correctly, that
is a genuine strengthening of
; it might be called
.Footnote
29
I’ll take
$\forall ^{\omega }xA$
to be
. Like
$\forall ^{1}$
(and unlike
$\forall ^{0})$
, it will play a salient role in the quantifier rules. In the semantics,
$|\forall ^{\omega }xA|_{\alpha }\geq r$
iff
$(\forall \epsilon>0)(\exists \beta [\beta +\omega \leq \alpha \wedge (\forall \gamma \in [\beta ,\alpha ])(\forall c\in |M|)(|A(c)|_{\gamma }\geq r-\epsilon )]$
.
Of course
$\forall ^{\omega }$
,
$\forall ^{1}$
and
$\forall ^{0}$
are all equivalent when applied to “ordinary” formulas A, those that can be assumed to be regular.
If you think it odd to have unrestricted quantifiers
$\forall $
inequivalent to
$\neg \exists \neg $
, it may help to observe that we already have such an inequivalence for restricted quantification: the restricted universal is defined via
$\rightarrow $
and the restricted existential via
$\wedge $
. (Also, many other non-classical logics, like intuitionist, have
$\forall $
stronger than
$\neg \exists \neg $
, as is the case with
$\forall ^{1}$
and
$\forall ^{\omega }$
.)
We use three axioms and a meta-rule, and strengthen the truth rule:
-
Q1 (
$\exists $ -Intro)
$B(y)\rightarrow \exists xB(x)$ when no free occurrence of x in
$B(x)$ is in the scope of an
$\exists y.$
-
Q2 (Central Quantifier Rule)
$\forall ^{1}x(Bx\rightarrow C)\rightarrow (\exists xBx\rightarrow C)$ when x not free in
$C.$
-
Q3 (InfDist)
$C\wedge \exists xBx\rightarrow \exists x(C\wedge Bx).$
-
Metarule (UnivGen): If
$A_{1},{\ldots},A_{n}\vdash Bx$ and x isn’t free in any of
$A_{1},{\ldots},A_{n}$ then
$A_{1},{\ldots},A_{n}\vdash \forall ^{\omega }xBx.$
We generalize the truth schema, and add composition principles. In these cases, it won’t matter which version of the universal quantifier to use (as we’ll see), so I write them without a superscript.
-
Generalized Truth Schema
$(\forall t_{1},{\ldots}t_{n})[TERM(t_{1})\wedge {\ldots}\wedge TERM(t_{n})\supset $
$ [True(\langle A(t_{1},{\ldots},t_{n})\rangle )\leftrightarrow A(den(t_{1}),{\ldots},den(t_{n}))]]$
-
Generalized Composition Schema for
$\rightarrow $ :
$(\forall x)(\forall y)[QSENT(x)\wedge QSENT(y)\supset [True(x\dot {\rightarrow }y)\leftrightarrow (True(x)\rightarrow True(y))]$ ],
with analogous composition rules for other connectives. (‘
$QSENT$
’ means ‘quasi-sentence’: see the text to which n. 8 is attached.) Given the compositional rules, it’s obviously enough to postulate the generalized truth schema for atomic A.
Here and in §2, I’ve written most of the axioms and rules schematically, but now that we have quantifiers and transparent truth, we should really use generalized versions. For instance, we should replace S1 with the generalization
$\forall x[QSENT(x)\supset True(x\dot {\rightarrow }x)]$
. (Again, it makes little difference which version of the quantifier we use.) From this, we can easily derive specific instances
$A\rightarrow A$
using Q1 and the truth rules (and a bit more if we use
$\forall ^{1}$
or
$\forall ^{\omega }$
). We can do this for each of the axiom schemas, except that we need a schema in the vicinity of Q1 as well as the generalizations so as to be able to derive the other schemas from their corresponding generalizations. We also should generalize the rules, e.g., replacing Modus Ponens by
$True(x),True(x\dot {\rightarrow }y)\vdash True(y)$
, and supplementing UnivGen as written by “If
$True(y_{1}),{\ldots},True(y_{n})\vdash True(z)$
then for every variable x not free in any member of the
$y_{i}$
,
”. If this is done, the proofs of the schematic theorems and schematic derived rules are easily transformed to proofs of the corresponding generalizations. For ease of reading, I will continue to write in schematic terms, except in one place in Appendix B, where the generalized form is needed.
Q2 could be replaced by the following pair:
-
Q2a.
$\forall ^{1}x(Bx\rightarrow Cx)\rightarrow (\exists xBx\rightarrow \exists xCx).$
-
Q2b.
$\exists xB\rightarrow B$ when x isn’t free in
$B.$
(Q1 entails the converse of Q2b.)
In Łukasiewicz logic, Q3 is derivable from other quantifier laws, but that seems unlikely here.Footnote 30
Obviously , by definition of
plus the truth rules. More generally,
-
-Elim For any n,
(For a special case of the previous is , and it’s not hard to prove the equivalence of
to
; see Appendix B.)
And using UnivGen, we get a rule form of the converse:
-
-Introduction
.
Proof.
$A\vdash A$
, so by UnivGen
$A\vdash \forall ^{\omega }xA$
for any x not free in A, i.e.,
; and using Q2b (plus contraposition, etc.) we get
.
In Q2, on the other hand, the use of
$\forall ^{1}$
is essential: the analog with
$\forall ^{\omega }$
is weaker than we want, and the analog with
$\forall ^{0}$
isn’t sound. For suppose
$B(n)$
is
$R\rightarrow ^{n}\bot $
, where R is the Restall sentence. (More exactly:
$B(n)$
is “either n isn’t a natural number, or n is a natural number and the result of prefixing
$\bot $
with ‘
$R\rightarrow $
’ n times is true”.) Then for each n, and each limit ordinal
$\lambda $
and natural numbers k,
$|B(n)|_{\lambda +k}$
is 1 iff
$n\geq k\geq 1$
, 0 otherwise. Then
$|\exists nB(n)|_{\lambda +k}$
is 1 for all
$k\geq 1$
(since
$|B(k)|_{\lambda +k}$
is 1 when
$k\neq 0$
), and so
$|\exists nB(n)\rightarrow \bot |_{\lambda +k}$
is 0 for all
$k\geq 1$
, and so
$|\exists nB(n)\rightarrow \bot |_{\lambda +\omega }$
is 0. But for each n,
$|B(n)\rightarrow \bot |_{\lambda +k+1}$
is 1 if
$|B(n)\rightarrow \bot |_{\lambda +k}=0$
, that is, if
$k>n$
; so for each n,
$|B(n)\rightarrow \bot |_{\lambda +\omega }$
is 1, so
$|\forall ^{0}n(B(n)\rightarrow \bot )|_{\lambda +\omega }$
is 1.
That Q2 as stated is sound in the semantics is proved in Appendix A. (It wouldn’t be sound in a semantics based on (SYM), since it implies the K-law for $.) The soundness of the others is obvious.
Of course we don’t need
$\forall ^{1}$
for the rule forms of Q2 and Q2a:
$\forall ^{0}x(Bx\rightarrow C)\vdash \exists xBx\rightarrow C$
follows from Q2, because of $-introduction, and similarly for Q2a.
A crucial feature of
$\forall ^{\omega }$
(not shared by
$\forall ^{1}$
or
$\forall ^{0}$
) is
-
(RQ
$_{\rightarrow }$ )
$\forall ^{\omega }x(\neg Ax\vee Bx)\rightarrow \forall ^{\omega }x(Ax\rightarrow Bx)$ .
That this is not only valid, but provable in our system, is shown in Appendix B. I’ll discuss its ramifications in §5.
It’s also useful to introduce n-ary quantifiers, that are candidates for understanding “for all
$x_{1}$
,…,
$x_{n}$
,
$A(x_{1},{\ldots},x_{n})$
”. I take
$(\forall ^{0}x_{1},{\ldots},x_{n})$
to be defined as a string of n unary
$\forall ^{0}$
s:
$(\forall ^{0}x_{1},{\ldots},x_{n})A(x_{1},{\ldots},x_{n})$
is just
$(\forall ^{0} x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$
. (Similarly for
$(\exists x_{1},{\ldots},x_{n})$
.) But I take
$(\forall ^{1}x_{1},{\ldots},x_{n})$
to be defined as
: that is,
$(\forall ^{1}x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$
, which for
$n>1$
is weaker than a string of
$\forall ^{1}$
s. Strings of more than one
$\forall ^{1}$
s will turn out to be rather less useful. Similarly,
$(\forall ^{\omega }x_{1},{\ldots},x_{n})$
is to be defined as
: that is,
$(\forall ^{\omega }x_{1})(\forall ^{0}x_{2}){\ldots}(\forall ^{0}x_{n})A(x_{1},x_{2},{\ldots},x_{n})$
.
Here are some salient facts about what’s derivable in the system, even without using the Regularity Schema. (More detail, with proofs, in Appendix B.) And about laws you might expect but that aren’t sound.
-
Vacuous quantification: If x isn’t free in C, then
$\exists xC$ and
$\forall ^{0}xC$ are each equivalent to C. But
$\forall ^{1}xC$ is equivalent to $
$C$ , and
$\forall ^{\omega }xC$ is equivalent to
. When
$n>1$ and
$x_{1}$ isn’t free in A,
$(\forall ^{1}x_{1},{\ldots},x_{n})A$ is equivalent to
$(\forall ^{1}x_{2},{\ldots},x_{n})A$ (and analogously for
$x_{i}$ when
$i>1$ ), since the one occurrence of $ remains when the
$x_{1}$ is deleted; analogously for
$\forall ^{\omega }$ .
-
Change of bound variables: Works just as expected.
-
vs.
: We get the analog of the converse Barcan formula for
$\forall ^{0}$ , that is,
; but only the rule form
of the Barcan formula.Footnote 31
-
vs.
, and
vs.
: Analogously, the converse Barcan holds with
$\forall ^{1}$ and
$\forall ^{\omega }$ , but the Barcan only in rule form.
-
Commutation of quantifiers: We have
$\exists x\exists yAxy\rightarrow \exists y\exists xAxy$ , and similarly when all the
$\exists $ ’s are replaced by
$\forall ^{0}$ . With
$\forall ^{1}$ and
$\forall ^{\omega }$ we have only the rule forms, which is part of the reason why the n-ary quantifiers
$(\forall ^{1}x_{1},{\ldots},x_{n})$ and
$(\forall ^{\omega }x_{1},{\ldots},x_{n})$ are more useful than strings of the corresponding unary quantifiers. With the n-ary quantifiers, permuting the variables in the quantifier is unproblematic.
-
Universal analogs of Q2 (and Q2a): We have
$\forall ^{1}x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{0}xBx)$ (and
$\forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{0}xAx\rightarrow \forall ^{0}xBx)$ ). (We can’t replace
$\forall ^{0}$ in the consequent with
$\forall ^{1}$ : see Appendix B.) Similarly
$\forall ^{\omega }x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{n}xBx)$ , for any n (where
$\forall ^{n}$ is
); and
-
$\exists $ -Importation: If C doesn’t contain x free,
$\vdash \exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ .
-
Failure of
$\exists $ -Exportation Rule: The rule
$C\rightarrow \exists xBx\vdash \exists x(C\rightarrow Bx)$ (for C not containing free x) is unsound in the semantics.
That
$\exists $
-Exportation fails is a good thing! For (as noted in [Reference Bacon1]) its validity in Łukasiewicz logic is the source of the inconsistency mentioned in n. 26. (Taking C to be the Restall sentence R and
$B(n)$
to be
$R\rightarrow ^{n}\bot $
,
$\exists -$
Exportation (together with transparency) immediately leads to the proof of R; and with that and the compositional rule for
$\rightarrow $
,
$\neg R$
is easily proved as well.)Footnote
32
These give enough of the flavor of the system to enable us to move to restricted quantification.
5 Restricted quantification
I mentioned early on that my concern in this paper was with a conditional
$\rightarrow $
adequate to defining universal restricted quantification: “All A are B” was to be understood as ‘
$\forall x(Ax\rightarrow Bx)$
’. But we’ve seen at least three candidates for ‘
$\forall $
’, so three candidates for “All A are B”: we could take it as
$\forall ^{0}x(Ax\rightarrow Bx)$
, as
$\forall ^{1}x(Ax\rightarrow Bx)$
, or as
$\forall ^{\omega }x(Ax\rightarrow Bx)$
. Similarly for multi-place restricted: we could take “All
$x_{1}$
…,
$x_{n}$
that stand in relation A stand in relation B” as
$(\forall ^{0}x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$
, or as
$(\forall ^{1}x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$
, or as
$(\forall ^{\omega }x_{1}{\ldots},x_{n})[A(x_{1}{\ldots},x_{n})\rightarrow B(x_{1}{\ldots},x_{n})]$
. I’ll use the notations
$\forall ^{0}x_{Ax}Bx$
,
$\forall ^{1}x_{Ax}Bx$
and
$\forall ^{\omega }x_{Ax}Bx$
for the unary restricted quantifiers (and analogously for the n-ary).
I think there’s a big advantage to the
$\forall ^{\omega }$
readings for both the restricted and unrestricted quantifiers: together these validate the obviously desirable law “If everything is either not-A or B then all A are B”:
-
(RQ)
$\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}Bx$ .
(Using an ordinal bigger than
$\omega $
for both unrestricted and restricted (see n. 29) would have the same advantage, but there seems little point to that.)
I don’t of course mean this as a claim about ordinary English: ordinary speakers don’t have coherent ways of dealing with truth-theoretic paradoxes, and certainly not of the arcane paradoxes involving Restall-like sentences; and only for such sentences is there any real distinction between
$\forall ^{0}$
,
$\forall ^{1}$
and
$\forall ^{\omega }$
. What I mean, rather, is that by and large, ordinary reasoning is best preserved even for arcanely paradoxical sentences if we understand both the unrestricted and the restricted ‘all’ as
$\forall ^{\omega }$
.
If we were to understand
$\forall $
in (RQ) as either
$\forall ^{0}$
or
$\forall ^{1}$
in both occurrences, (RQ) wouldn’t be valid. We could get it by an ad hoc adjustment to the definition of restricted quantification, e.g., taking
$\forall x_{Ax}Bx$
not as
$\forall x(Ax\rightarrow Bx)$
but as
$\forall x[(Ax\rightarrow Bx)\vee (\neg Ax\vee Bx)]$
; but while that would deliver (RQ) it would fail to deliver related laws, such as
-
(RQ*)
$\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}(Ax\wedge Bx)$ .
(For instance, suppose that for any object c,
$|Ac|_{\alpha }$
is
$\frac {1}{2}$
for all
$\alpha $
and
$|Bc|_{\alpha }$
is 1 for odd
$\alpha $
and 0 for even
$\alpha $
. On the proposed ad hoc reading, (RQ*) then amounts to
$(\neg Ac\vee Bc)\rightarrow [(Ac\rightarrow Ac\wedge Bc)\vee \neg Ac\vee (Ac\wedge Bc)]$
. But if
$\alpha +1$
is odd, then
$|\neg Ac\vee Bc|_{\alpha +1}$
is 1 while
$|Ac\rightarrow Ac\wedge Bc|_{\alpha +1}$
and
$|\neg Ac|_{\alpha +1}$
are
$\frac {1}{2}$
and
$|Ac\wedge Bc|_{\alpha +1}$
is 0; so this instance of (RQ*) will get value
$\frac {1}{2}$
at
$\alpha +2$
.) To get laws (RQ) and (RQ*) without going to infinite iterations of $, we need one more $ on the left than on the right: e.g.,
$\forall ^{1}$
on the left and
$\forall ^{0}$
on the right. If we want a single unrestricted ‘
$\forall $
’, and define restricted quantification in terms of it, we need to use
$\forall ^{\omega }$
(or
$\forall ^{\alpha }$
for some bigger
$\alpha $
).
Might we do better by taking the restricted quantifier as primitive, and defining the unrestricted in terms of it (
$\forall xBx$
as
$\forall x_{x=x}Bx$
)? We’d need valuation rules for the restricted quantifier, but the obvious idea would mimic the revision construction here, evaluating
$\forall x_{Ax}Bx$
at an ordinal directly in terms of the values of
$Ac$
and
$Bc$
at prior ordinals. But in that case, explaining
$\forall xBx$
as
$\forall x_{x=x}Bx$
would be unacceptable: if the
$Bc$
’s all have high values at ordinals prior to
$\alpha $
,
$\forall xBx$
would automatically get a high value at
$\alpha $
, even if some or all of its instances
$Bc$
have low values at
$\alpha $
. This can be so even when
$\alpha $
is past the critical ordinal; so universal instantiation would fail.
Of course we could get the equivalence of “everything is B” to “every self-identical thing is B” by reading the former as
$\forall ^{1}xBx$
and the latter as
$\forall ^{0}x_{Ax}Bx$
(or more generally, the former as
$\forall ^{n+1}xBx$
and the latter as
$\forall ^{n}x_{Ax}Bx$
, for finite n). But that seems rather unnatural, which is why I prefer going to
$\forall ^{\omega }$
.
6 Final remarks
I expect that the derivation system presented in §2 and §4 could usefully be expanded (even beyond the added regularity principle suggested at the end of §3; and I expect that that principle too could usefully be extended to a wider class of sentences). A system that is fully complete on this semantics is not in the cards.Footnote 33 But it would certainly be desirable to strengthen the one here; this paper is only a first step.
But I think it is an important step in a number of ways. Probably most centrally, it shows that we can have a theory of transparent truth that has a conditional obeying natural laws and where in particular we can have the law
-
(RQ)
$\forall x(\neg Ax\vee Bx)\rightarrow \forall x_{Ax}Bx$ ,
reading
$\forall x_{Ax}Bx$
as “All A are B”, and taking it to abbreviate
$\forall x(Ax\rightarrow Bx)$
. Admittedly, we get (RQ) only by treating the universal quantifier in an unusual way, not as
$\neg \exists \neg $
; but non-classical theories already pretty much required us to treat the universal restricted quantifier as different from
$\neg \exists x_{Ax}\neg Bx$
(where
$\exists x_{Ax}Cx$
is the restricted existential quantifier
$\exists x(Ax\wedge Cx)$
), so treating the universal unrestricted quantifier as different from
$\neg \exists \neg $
doesn’t seem totally outlandish.
Previously, it was reasonable to worry that a theory of transparent truth with (RQ) might be unobtainable (except in trivial ways, like adding ’ as a disjunct, that don’t deliver related laws like
$\forall x(\neg Ax\vee Bx)\rightarrow \forall x(Ax\rightarrow Ax\wedge Bx)$
.) Brady’s pioneering papers on conditionals in transparent truth theories, and his book [Reference Brady4], don’t validate even the Weakening Rule
-
(WR)
$\forall xBx\vdash \forall x(Ax\rightarrow Bx)$
(and though he didn’t consider varying the semantics of the quantifier, it is pretty clear that there is no remotely acceptable treatment of quantifiers that would give this rule given his treatment of conditionals). In [Reference Bacon2] Bacon modifies Brady’s construction, and says that with his modification, we get not only that rule but a conditional form of it:
-
(WC)
$\vdash \forall xBx\rightarrow \forall x(Ax\rightarrow Bx)$ .
But in the main body of his paper (all but the last section), Bacon gets (WC) only in a very limited language, with no proper negation operator—so for instance, it has no means to express ordinary Liar sentences.Footnote 34 The final section of his paper does contain interesting though quite undeveloped suggestions about how to extend his account to a language with a real negation, with a semantics that uses what is essentially the “Routley star” from relevance logic. But though this delivers (WC), it does not deliver (RQ) (which doesn’t follow from (WC) in its logic since its conditional is non-contraposable). It doesn’t deliver even the rule
-
(WCR)
$\forall x\neg Ax\vdash \forall x(Ax\rightarrow Bx)$ ;
it takes it as consistent that there are no A’s and yet not all A’s are B’s (e.g., in his preferred version, when
$Bx$
is
$x\neq x$
, and
$Ax$
is
$x=x\wedge Q_{2}$
, with
$Q_{2}$
as in n. 14).
In earlier work, e.g., [Reference Field7], I made a different sort of attempt to get a law “If everything is either not A or B then all A are B”: I read it not as RQ but rather as
-
(RQ*)
$\forall x(\neg Ax\vee Bx)\triangleright \forall x(Ax\rightarrow Bx)$ ,
where
$\triangleright $
was a different sort of conditional from
$\rightarrow $
:
$\rightarrow $
is a restricted quantifier conditional,
$\triangleright $
is an ordinary English conditional (perhaps something roughly like the epistemic conditionals in the tradition of Stalnaker 1968). But, while I don’t reject the idea that some pre-theoretic laws about conditionals depend on reading some ‘if…then’s as ordinary English conditionals, the logic that I managed to get by these means was rather weak.
In this paper, I think I’ve achieved much more satisfactory results, and done so without need of a separate conditional
$\triangleright $
. Part of this is due to my suggestion of a different reading for ‘all’. But that wouldn’t have led to useful results in a theory based on the revision-theoretic semantics in my earlier papers, because those papers made a poor choice for how the revision procedure behaved at limit ordinals. The earlier part of the present paper shows, I think, that even independent of quantification, the new limit rule (LimInf) is far preferable to the old one; and the treatment of quantifiers in later sections of this paper wouldn’t have made any sense on the old limit rule.
I remarked at the beginning of the paper that we have a choice in how to deal with the paradoxes: we can keep the logic classical and complicate the truth theory, or keep the truth theory simple and complicate the logic. We can’t know which is the better package of logic plus truth theory until we have developed both packages as best we can. The present paper is part of an ongoing effort to do this, focusing especially on the non-classical side. While just a first step, I think that by getting a closer approximation to classical laws, especially laws of restricted quantification, it significantly improves the competitive advantages of the combination of transparent truth and non-classical logic over classical logic with restrictions on transparency. A further step in improving the competitive advantage would involve the addition of predicates “well-behaved” (guaranteeing classical behavior) and “minimally well-behaved” (guaranteeing regularity), as in [Reference Field9].Footnote 35 But the improvement achieved by adding such predicates is somewhat orthogonal to the issues discussed in this paper.
Appendix A: Soundness
The soundness of many of the axioms and rules, with respect to the [0,1]-based revision theory with LimInf, is obvious. I’ll give proofs for the less obvious ones: S6, S7a (S7b is analogous), S7c, S8 (though it too is pretty obvious), S9, S11, R4a (R4b is analogous) and Q2. The system is also sound with respect to the revision theory based on three values (and on 2n+1 values more generally), and the proofs are simpler there. Of course the assumption of regularity for quantifier-independent sentences requires the continuum-based version.
In many of the proofs below, I’ll tacitly rely on a fact noted in the main text, that the value of a conditional
$A\rightarrow B$
at a limit ordinal can be characterized either as the
$\text {liminf}$
of the values of that conditional at prior ordinals, or as the
$\text {liminf}$
of the values
at prior ordinals (where these are the values obtained from
$|A|_{\gamma }$
and
$|B|_{\gamma }$
by the Łukasiewicz evaluation rule). The former “continuity formulation” is what was built into the LimInf rule as I originally stated it; the latter “general rule” has the advantage of holding for successor ordinals as well as for limits. An easy transfinite induction shows their equivalence for limit ordinals. The general rule can be written more explicitly as
-
(GR) For any
$\alpha>0$ and
$r<1$ ,
$h_{\alpha }(A\rightarrow B)\geq 1-r$ iff
$(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(|A|_{\gamma }-|B|_{\gamma }\leq r+\epsilon )$ .
The equivalence of the two formulations leads immediately to the following lemma.
Lemma A1.
$\top \rightarrow (A\rightarrow B)$
has the same value as
$A\rightarrow B$
at limit ordinals. (So
$|\top \rightarrow (A\rightarrow B)|_{\lambda +1}=|\top \rightarrow (A\rightarrow B)|_{\lambda }$
.)
Proof. By the “general rule”,
$|\top \rightarrow (A\rightarrow B)|_{\lambda }$
is
, which is
$liminf_{\lambda }\{|A\rightarrow B|_{\gamma }\}$
, which by “continuity formulation” is
$|A\rightarrow B|_{\lambda }$
.
Lemma A2. For any r: if for some
$\beta $
,
$|A\rightarrow B|_{\alpha }\geq r$
for all successor ordinals greater than
$\beta $
, then for all final ordinals (successor or limit),
$|A\rightarrow B|_{\alpha }\geq r$
(where a final ordinal is an ordinal greater than or equal to the critical ordinal).
Proof. The premise guarantees that for all final ordinals
$\beta $
,
$|A|_{\beta }\leq |B|_{\beta }$
and hence
; so by the “general rule” for evaluating conditionals, the conclusion follows.
-
Soundness of S6:
$(A\rightarrow B)\wedge (A\rightarrow C)\rightarrow (A\rightarrow B\wedge C).$
Proof. By Lemma A2, we need only establish this for successor ordinals
$\geq \alpha _{0}+1$
(where
$\alpha _{0}$
is the critical ordinal), and that amounts to showing that for all final
$\alpha $
, and all reals
$r<1 $
, if
$|A\rightarrow B|_{\alpha }\geq 1-r$
and
$|A\rightarrow C|_{\alpha }\geq 1-r$
then
$|A\rightarrow B\wedge C|_{\alpha }\geq 1-r$
. By (GR), the premises tell us that for any
$\epsilon>0$
, there are
$\beta _{1}$
and
$\beta _{2}$
such that
$(\forall \gamma \in [\beta _{1},\alpha ))(|A|_{\gamma }-|B|_{\gamma }\leq r+\epsilon )$
and
$(\forall \gamma \in [\beta _{2},\alpha )) (|A|_{\gamma }-|C|_{\gamma }\leq r+\epsilon )$
. Letting
$\beta $
be their maximum, we get that
$(\forall \gamma \in [\beta _{,}\alpha ))(|A|_{\gamma }-min\{|B|_{\gamma },|C|_{\gamma }\}\leq r+\epsilon )$
(since
$min\{|B|_{\gamma },|C|_{\gamma }\}$
is either
$|B|_{\gamma }$
or
$|C|_{\gamma }$
). So
$(\forall \gamma \in [\beta _{,}\alpha ))(|A|_{\gamma }-|B\wedge C|_{\gamma })\leq r+\epsilon $
, and so by (GR) again,
$|A\rightarrow B\wedge C|_{\lambda }\geq 1-r$
.
Lemma A3. For any
$\alpha>0$
,
$|A\rightarrow B|_{\alpha }+|B\rightarrow C|_{\alpha }\leq 1+|A\rightarrow C|_{\alpha }$
. (So if
$|B\rightarrow C|_{\alpha }=1$
,
$|A\rightarrow B|_{\alpha }\leq |A\rightarrow C|_{\alpha }$
.)
[We obviously need the ‘
$1+$
’: e.g., if A implies B and B implies C, both terms on left are 1.]
Proof. Using the well-known inequality
$liminf\{s_{n}+t_{n}\}\geq liminf\{s_{n}\}+ liminf\{t_{n}\}$
, we get that the left-hand side is
. And
is easily seen to be
(they’re equal when neither summand is 1). So the LHS is
, and that’s the RHS.
Lemma A4. For any
$\alpha>0$
,
$|A\sqcap C|_{\alpha }+|A\rightarrow B|_{\alpha }\leq 1+|B\sqcap C|_{\alpha }$
.
Proof. Writing D for
$\neg C$
, Lemma A3 gives
$|A\rightarrow B|_{\alpha }+|B\rightarrow D|_{\alpha }\leq 1+|A\rightarrow D|_{\alpha }$
. But
$|A\sqcap C|_{\alpha }$
is
$1-|A\rightarrow D|_{\alpha }$
, and similarly for
$|B\sqcap C|_{\alpha }$
, so
$|A\rightarrow B|_{\alpha }-|B\sqcap C|_{\alpha }\leq 1-|A\sqcap C|_{\alpha }$
, which is equivalent to the corollary.
-
Soundness of S7a:
$[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)].$
Proof. Obviously
$|A\rightarrow B|_{\alpha }\leq 1$
, and so by Lemma A3, when
$\alpha>0$
,
$|A\rightarrow B|_{\alpha }\leq min\{1,1-|B\rightarrow C|_{\alpha }+|A\rightarrow C|_{\alpha }\}$
. So (for any
$\alpha>0$
)
$|\top \rightarrow (A\rightarrow B)|_{\alpha +1}\leq |(B\rightarrow C)\rightarrow (A\rightarrow C)|_{\alpha +1}$
. So by Lemma A2, we get that for any final
$\alpha $
,
$|\top \rightarrow (A\rightarrow B)|_{\alpha }\leq |(B\rightarrow C)\rightarrow (A\rightarrow C)|_{\alpha }$
. So for any final successor ordinal,
$|[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)]|_{\alpha }=1$
. So by Lemma A2 again,
$|[\top \rightarrow (A\rightarrow B)]\rightarrow [(B\rightarrow C)\rightarrow (A\rightarrow C)]|_{\alpha }=1$
for any final ordinal.
-
Soundness of S7c: Analogous to proof for S7a, but using Lemma A4 instead of A3.
-
Soundness of S8:
$[\top \rightarrow (\neg A\vee B)]\rightarrow [A\rightarrow B].$
Proof. For any
$\alpha $
,
$|\top \rightarrow (\neg A\vee B)|_{\alpha +1}$
is
$max\{1-|A|_{\alpha },|B|_{\alpha }\}$
, which is
$\leq min\{1,1-|A|_{\alpha }+|B|_{\alpha }\}$
, which is
$|A\rightarrow B|_{\alpha +1}$
. Using Lemma A2, we get the result.
-
Soundness of S9:
$[(\top \sqcap A)\rightarrow (B\rightarrow C)]\rightarrow [(\top \rightarrow B)\rightarrow (A\rightarrow C)].$
Proof. We need that for all final
$\alpha $
,
$|(\top \sqcap A)\rightarrow (B\rightarrow C)|_{\alpha }\leq |(\top \rightarrow B)\rightarrow (A\rightarrow C)|_{\alpha }$
. By Lemma A2, it suffices to show this when
$\alpha $
is a successor
$\beta +1$
. In that case, what we need is that
-
(*)
$|\top \rightarrow B|_{\beta }-|A\rightarrow C|_{\beta }\leq max\{0,|\top \sqcap A|_{\beta }-|B\rightarrow C|_{\beta }\}$ .
We prove (*) by induction on
$\beta>0.$
When
$\beta $
is itself a successor
$\delta +1$
, the LHS of (*) is
$B_{\delta }-min\{1,1-A_{\delta }+C_{\delta }\}$
, i.e.,
$max\{B_{\delta }-1,B_{\delta }-1+A_{\delta }-C_{\delta }\}$
(which may be negative).
And the RHS is
$max\{0,A_{\delta }-min\{1,1-B_{\delta }+C_{\delta }\}\}$
, i.e.,
$max\{0,A_{\delta }-1,A_{\delta }-1+B_{\delta }-C_{\delta }\}$
.
If the LHS is
$B_{\delta }-1$
, then it’s
$\leq 0$
, so
$\leq $
RHS. And if it’s
$\leq B_{\delta }-1+A_{\delta }-C_{\delta }$
then its also
$\leq $
RHS, since that’s equivalent to the last term on right-hand side by rearrangement.
When
$\beta $
is a limit
$\lambda $
, we invoke induction hypothesis: for all
$\gamma <\lambda $
, (*) holds, or rewriting it,
$|A\rightarrow C|_{\gamma }\geq |\top \rightarrow B|_{\gamma }-max\{0,|\top \sqcap A|_{\gamma }-|B\rightarrow C|_{\gamma }\}$
, i.e.,
$|A\rightarrow C|_{\gamma }\geq min\{|\top \rightarrow B|_{\gamma },|\top \rightarrow B|_{\gamma }-|\top \sqcap A|_{\gamma }+|B\rightarrow C|_{\gamma }\}$
.
Taking the
$liminf_{\lambda }$
of both sides, we get

and since the
$liminf$
of the minimum of two sequences is the minimum of their
$liminf$
s,

But since
$liminf(-y_{\gamma })=-limsup(y_{\gamma })$
, we have

so the above yields

which means

Taking the negatives of both sides, we get (*) for
$\lambda $
.
-
Soundness of S11:
$[(A\rightarrow B)\rightarrow (\top \rightarrow B)]\rightarrow [\top \rightarrow ((\top \sqcap A)\vee (\top \sqcap B))].$
Proof. By Lemma A2, it suffices to show that at successor ordinals, the LHS has value
$\leq $
the RHS; that is, that for all
$\beta $
,

(The fact that
$|A\rightarrow B|_{\beta }\geq |\top \rightarrow B|_{\beta }$
avoids any need to minimize the LHS with 1. On the RHS, I’ve used
$\top \sqcap (A\vee B)$
instead of the equivalent
$(\top \sqcap A)\vee (\top \sqcap B)$
, to simplify the proof of the limit stage of the induction below.)
If
$\beta $
is itself a successor
$\delta +1$
, the LHS is
$1-min\{1,1-|A|_{\delta }+|B|_{\delta }\}+|B|_{\delta }$
; that is,
$max\{B_{\delta },A_{\delta }$
}. That’s the value of the RHS.
When
$\beta $
is a limit
$\lambda $
we use induction. We can rewrite the inequality as
$1-|\top \sqcap (A\vee B)|_{\beta }+|\top \rightarrow B|_{\beta }\leq |A\rightarrow B|_{\beta }$
. Suppose this holds prior to
$\lambda $
. Taking
$liminf$
s (at
$\lambda $
), we have
$liminf\{1-|\top \sqcap (A\vee B)|_{\beta }+|\top \rightarrow B|_{\beta }\}\leq |A\rightarrow B|_{\lambda }$
. But by the same liminf law used in proving S9, this yields
$1-limsup\{|\top \sqcap (A\vee B)|_{\beta }\}+liminf\{|\top \rightarrow B|_{\beta }\}\leq |A\rightarrow B|_{\lambda }$
; that is,
$1-|\top \sqcap (A\vee B)|_{\lambda }+|\top \rightarrow B|_{\lambda }\leq |A\rightarrow B|_{\lambda }$
, as desired.
-
Soundness of R4a:
$(\top \rightarrow B)\rightarrow B\vdash B\rightarrow (\top \rightarrow B).$
Proof. If
$|(\top \rightarrow B)\rightarrow B|_{\alpha }$
is 1 for all final
$\alpha $
then for all final
$\beta $
,
$|B|_{\beta }\geq |\top \rightarrow B|_{\beta }$
. From this, it’s easily seen by induction that the sequence
$|B|_{\beta }$
is non-decreasing once the critical ordinal is reached. But then it must be constant: if it strictly increased then the valuations at earlier stages wouldn’t be recurrent. So
$|B\rightarrow (\top \rightarrow B)|_{\alpha }$
is 1 for all final
$\alpha $
.
-
Soundness of Q2:
$\forall ^{1}x(Bx\rightarrow C)\rightarrow (\exists xBx\rightarrow C)$ when x not free in
$C.$
Proof. It suffices to prove
(1): For successor
$\alpha +1$
,
$|\forall ^{0}x(Bx\rightarrow C)|_{\alpha +1}\leq |\exists xBx\rightarrow C|_{\alpha +1}$
, and
(2): For limit
$\lambda $
,
$|\top \rightarrow \forall ^{0}x(Bx\rightarrow C)|_{\lambda }\leq |\exists xBx\rightarrow C|_{\lambda }$
.
(1) is obvious: If
$|\exists xBx\rightarrow C|_{\alpha +1}<r$
then
$|\exists xBx|_{\alpha }-|C|_{\alpha }>1-r$
, so there’s a c such that
$|Bc|_{\alpha }-|C|_{\alpha }>1-r$
, and so for which
$|Bc\rightarrow C|_{\alpha +1}<r$
. So
$|\forall ^{0}x(Bx\rightarrow C)|_{\alpha +1}<r$
.
(2): If
$|\top \rightarrow \forall ^{0}x(Bx\rightarrow C)|_{\lambda }\geq r$
then
$(\forall \epsilon>0)(\exists \beta <\lambda )(\forall \gamma \in [\beta ,\lambda ))(\forall c\in |M|)(|Bx\rightarrow C|_{\gamma }\geq r-\epsilon /2)$
; this holds in particular at successor
$\gamma $
, so

For any
$\delta $
, there’s a c such that

so
$(\forall \epsilon>0)(\forall \delta \in [\beta _{\epsilon},\lambda ))(1-|\exists xBx|_{\delta }+|C|_{\delta }\geq r-\epsilon )$
. So
$|\exists xBx\rightarrow C|_{\lambda }\geq r$
.
Appendix B: Quantifier derivations and invalidities
Proofs of most of the sentential laws mentioned in §2 were sketched there, so I’ll just sketch derivations of some of the quantifier laws mentioned but not proved in §4, and a few others. I’ll also elaborate on a few of the invalidities mentioned there.
-
K-law for
:
.
Proof. This is the one place where we need to use the generalized form of the laws, mentioned early in §4, rather than the schemas. That enables us to turn the schematic proof of the K-law for $, given in §2, into a proof of
(A) .
The next step is to generalize this to
(B) .
Since we have induction in the theory (see n. 6), and the
$n=0$
case reduces to the composition law for
$\rightarrow $
, it suffices for (B) that we prove
(C)
$[B(n)\rightarrow (C(n)\rightarrow D(n))]\rightarrow [B(n+1)\rightarrow (C(n+1)\rightarrow D(n+1))]$
,
where
$B(n)$
is
,
$C(n)$
is
, and
$D(n)$
is
. Using contraposition, this amounts to
. But using $-introduction followed by three applications of (A), we get (C) and hence (B).
Using UnivGen on (B) (we only need the weaker form with
$\forall ^{1}$
) and Q2a, we get
(D) .
By $-Introduction and (A), this gives
(E) .
And then an application of Q2 to the consequent (together with the Suffixing Rule) yields
(F) .
In other words, .
The K-laws obviously entail the following rules (given -Introduction and
-Introduction, both proved in the text).
-
K-rules:
and
.
N.B.: We can’t strengthen the K-law for to
. For take A to be
$\top $
and B to be the Restall sentence R. Since
$|R|_{\alpha }$
is 0 whenever
$\alpha $
is a limit ordinal,
is 0 for all
$\alpha $
; so
is as well. But since
$|R|_{\beta +n}$
is 1 for any
$\beta $
and positive integer n,
$|\top \rightarrow R|_{\beta +n}$
is 1 for any
$\beta $
and
$n\geq 2$
, so
has value 1 at any
$\beta +\omega $
(that is, at any limit ordinal that isn’t a multiple of
$\omega ^{2}$
).
-
Crucial feature of
and
$\forall ^{\omega }$ : (i)
and (ii)
.
Proof. By definition, the consequent of (i) is
$(\forall n)True(\top \rightarrow ^{n}(\top \rightarrow B))$
, which is obviously equivalent to
$(\forall n>0)True(\top \rightarrow ^{n}B)$
; which follows from
$(\forall n)True(\top \rightarrow ^{n}B)$
, which is the antecedent. (ii) easily follows from (i).
-
Vacuous quantification: If x isn’t free in C,
$\vdash \exists x\neg C\leftrightarrow \neg C$ , by Q2b and Q1, from which we derive
$\vdash \forall ^{0}xC\leftrightarrow C$ using contraposition. The K-rules for $ and
then yield
and
.
-
Change of variables:
$\vdash \exists xAx\rightarrow \exists yAy$ (when no y in
$Ay$ is in the scope of an
$\exists x$ ), and analogously for
$\forall ^{0}$ , and
$\forall ^{1}$ and
$\forall ^{\omega }$ .
Proof.
$\vdash Ax\rightarrow \exists yAy$
, so by UnivGen and Q2,
$\vdash \exists xAx\rightarrow \exists yAy$
. (Recall that UnivGen has been stated in terms of
$\forall ^{1}$
.) The result for
$\forall ^{0}$
follows from this using contraposition, and the results for
$\forall ^{1}$
and
$\forall ^{\omega }$
then follow by the K-rules for $ and
.
-
Commutation of unary
$\exists $ and
$\forall ^{0}$ quantifiers:
$\vdash \exists x\exists yAxy\rightarrow \exists y\exists xAxy$ , and similarly when all the
$\exists $ ’s are replaced by
$\forall ^{0}$ .
Proof.
$\vdash Axy\rightarrow \exists xAxy$
by Q1. So
$\vdash \exists yAxy\rightarrow \exists y\exists xAxy$
by UnivGen on y plus Q2a. So
$\vdash \exists x\exists yAxy\rightarrow \exists y\exists xAxy$
by UnivGen on x plus Q2. The result for
$\forall ^{0}$
follows from this using contraposition.
-
Commutation within an n-ary
$\forall ^{1}$ quantifier or
$\forall ^{\omega }$ quantifier:
$\vdash (\forall ^{1}x_{1},x_{2},{\ldots}x_{n})A\rightarrow (\forall ^{1}x_{\pi (1)},x_{\pi (2)},{\ldots}x_{\pi (n)})A,$ where
$\pi $ is any permutation of
$\{1,{\ldots},n\}$ . [Equivalently,
$\forall ^{1}x_{1}\forall ^{0}x_{2}{\ldots}\forall ^{0}x_{n}A\rightarrow \forall ^{1}x_{\pi (1)}\forall ^{0}x_{\pi (2)},{\ldots},\forall ^{0}x_{\pi (n)})A.$ ] Analogously for
$\forall ^{\omega }$ .
Proof. An n-ary
$\forall ^{0}$
quantifier is obviously equivalent to a string of n unary
$\forall ^{0}$
quantifiers, so commutation in the string yields commutation within the n-ary
$\forall ^{0}$
. The result for the n-ary
$\forall ^{1}$
quantifier is then immediate by $-introduction and the K-law for $.
-
Commutation rule for unary
$\forall ^{1}$ quantifiers or
$\forall ^{\omega }$ quantifiers:
$\forall ^{1}x\forall ^{1}yAxy\vdash \forall ^{1}y\forall ^{1}xAxy$ , and analogously for
$\forall ^{\omega }$ .
Proof.
$\forall ^{1}x\forall ^{1}yAxy\vdash Axy$
. Apply UnivGen first to x, then to y.
N.B.: We can’t strengthen these to conditional form, e.g., to
$\forall ^{1}x\forall ^{1}yA(x,y)\rightarrow \forall ^{1}y\forall ^{1}xA(x,y)$
. For let
$Axy$
be of form
$Bx\wedge y=y$
. Clearly
$\forall ^{0}yAxy$
is then equivalent to
$\forall ^{0}yBx$
, which is equivalent to
$Bx$
, so
$\forall ^{1}yAxy$
is equivalent to
, so
$\forall ^{1}x\forall ^{1}yA(x,y)$
is equivalent to
. And clearly
$\forall ^{1}xAxy$
is equivalent to
$\forall {}^{1}xBx$
, so
$\forall ^{1}y\forall ^{1}xA(x,y)$
is equivalent to
. So this instance of commutation for
$\forall ^{1}$
is equivalent to the Barcan conditional for
$\forall ^{1}$
, ruled out in n. 31.
The one-way interchange laws
$\forall ^{1}\forall ^{0}\rightarrow \forall ^{0}\forall ^{1}$
and
$\forall ^{\omega }\forall ^{0}\rightarrow \forall ^{0}\forall ^{\omega }$
will be derived below.
-
Universal analogs of Q2a:
(1)
$\vdash \forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{0}xAx\rightarrow \forall ^{0}xBx)$
(1*)
$\forall ^{1}x(C\rightarrow Bx)\rightarrow (C\rightarrow \forall ^{0}xBx)$
(2)
(2*)
(3)
(3*) .
Proofs. (1) is immediate from Q2a, by contraposition of its consequent and the prefixing rule. From (1), the K-rule for $ yields ; the conditional form of K-law for $ yields
, and by Transitivity Rule, we get (2). For (3), we use the K-rule for $
$^{\omega }$
followed by that for $ to get
; the “crucial feature” allows replacing the antecedent by
$\forall ^{\omega +1}x(Ax\rightarrow Bx)$
, and the conditional form of K-law for
then allows replacing the consequent by
$\forall ^{\omega }xAx\rightarrow \forall ^{\omega }xBx$
. (1*)–(3*) follow from (1) to (3) by vacuous quantifier rules.
N.B.: We can’t strengthen (2) to
$\forall ^{1}x(Ax\rightarrow Bx)\rightarrow (\forall ^{1}xAx\rightarrow \forall ^{1}xBx)$
. This is evident, since when
$Ax$
is
$\top $
this is equivalent to
$\forall ^{1}x(\top \rightarrow Bx)\rightarrow (\top \rightarrow \forall ^{1}xBx)$
, and that implies the Barcan formula for
$\forall ^{1}$
which we’ve seen invalid (n. 31). [The antecedent of that Barcan formula,
, is
$\forall ^{1}x[Bx\wedge (\top \rightarrow Bx)]$
, which implies
$\forall ^{1}xBx\wedge \forall ^{1}x(\top \rightarrow Bx)$
, which by the bad strengthening of (2) would imply
$\forall ^{1}xBx\wedge (\top \rightarrow \forall ^{1}xBx)$
, which is the consequent
of that Barcan formula.] Similarly for (3).
-
Partial converse of Q2:
$\vdash (\exists xBx\rightarrow C)\rightarrow \forall ^{0}x(Bx\rightarrow C)$ , when x not free in
$C.$
Proof.
$\vdash Bx\rightarrow \exists xBx$
, so by Suffixing Rule,
$\vdash (\exists xBx\rightarrow C)\rightarrow (Bx\rightarrow C)$
, so by UnivGen and (1*),
$\vdash (\exists xBx\rightarrow C)\rightarrow \forall ^{0}x(Bx\rightarrow C)$
.
Corollary.
$\exists xBx\rightarrow C\vdash \forall ^{1}x(C\rightarrow Bx)$
.
-
Quantifiers with disjunction and conjunction:
(4)
$\vdash \exists x(Ax\vee Bx)\leftrightarrow \exists xAx\vee \exists xBx$
(5)
$\vdash (\forall ^{0}xAx\wedge \forall ^{0}xBx)\rightarrow \forall ^{0}x(Ax\wedge Bx)$
(6)
$\vdash (\forall ^{1}xAx\wedge \forall ^{1}xBx)\rightarrow \forall ^{1}x(Ax\wedge Bx)$
(7)
$\vdash \forall ^{0}x(C\vee Bx)\rightarrow (C\vee \forall ^{0}xBx)$
, when x isn’t free in
$C.$
Proofs. (4) The R to L is easy, using Reverse Adjunction. L to R:
$\vdash Ax\rightarrow (\exists xAx\vee \exists xBx)$
, using Q1,
$\vee $
-Introduction and transitivity; similarly
$\vdash Bx\rightarrow (\exists xAx\vee \exists xBx)$
. So by Reverse Adjunction,
$\vdash (Ax\vee Bx)\rightarrow (\exists xAx\vee \exists xBx)$
. So by UnivGen and Q2,
$\vdash \exists x(Ax\vee Bx)\rightarrow (\exists xAx\vee \exists xBx)$
.
(5): By (4) plus contraposition.
(6): From (5), use $-Introduction plus K-law for $ to get . Using
, conclusion follows.
(7) By Q3,
$\vdash \neg C\wedge \exists x\neg Bx\rightarrow \exists x(\neg C\wedge \neg Bx)$
. From this, we easily get
$\neg (C\vee \neg \exists x\neg Bx)\rightarrow \exists x\neg (C\vee Bx)$
, i.e.,
$\neg (C\vee \forall ^{0}xBx)\rightarrow \exists x\neg (C\vee Bx)$
. The result follows by contraposition (plus definition of
$\forall ^{0}$
and double negation laws).
-
Three analogs of converse Barcan formula:
(8a)
$\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow \forall ^{0}x(\top \rightarrow Bx)$
(8b) (that is,
)
(8c) (that is,
).
Proof of (8a):
$\vdash \forall ^{0}xBx\rightarrow Bx$
, so by prefixing rule,
$\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow (\top \rightarrow Bx)$
. By UnivGen
$\vdash \forall ^{1}x[(\top \rightarrow \forall ^{0}xBx)\rightarrow (\top \rightarrow Bx)]$
, and by (1
$\star$
),
$\vdash (\top \rightarrow \forall ^{0}xBx)\rightarrow \forall ^{0}x(\top \rightarrow Bx)$
.
The proofs of (8b) and (8c) are analogous, using the K-rules for $ and instead of the prefixing rule.
Rule versions of converses of these (that is, rule-analogs of the Barcan formula) are trivial.
-
One-way interchange laws:
$\vdash \forall ^{1}x\forall ^{0}yBxy\rightarrow \forall ^{0}x\forall ^{1}yBxy$ , and analogously with
$\forall ^{\omega }$ for
$\forall ^{1}$ in both occurrences.
Proof. Immediate from (8b) and (8c), taking
$Bx$
to be
$\forall ^{0}yBxy$
.
-
Further analogs of converse Barcan:
(9)
(10)
Proof of (9): use $-Introduction and K-law for $ on (8b).
Proof of (10): is
, which is equivalent to
, i.e.,
. But by (8b) that entails
, which is
.
-
Importation:
$\vdash \exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$ when x isn’t free in C.
Proof.
$\vdash Bx\rightarrow \exists xBx$
by Q1, so
$(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$
by Prefixing rule; then by UnivGen and Q2,
$\exists x(C\rightarrow Bx)\rightarrow (C\rightarrow \exists xBx)$
.
-
(#):
.
Proof. S8 gives
$\vdash (\top \rightarrow \neg A\vee B)\rightarrow (A\rightarrow B)$
. By K-rule for
,
, so by the “crucial feature”, we get the result.
-
(RQ):
$\forall ^{\omega }x(\neg Ax\vee Bx)\rightarrow \forall ^{\omega }x(Ax\rightarrow Bx)$ .
Proof. S8 gives
$\vdash (\top \rightarrow \neg Ax\vee Bx)\rightarrow (Ax\rightarrow Bx)$
. By UnivGen,
$\vdash \forall ^{1}x[(\top \rightarrow \neg Ax\vee Bx)\rightarrow (Ax\rightarrow Bx)]$
, and so by (1),
$\vdash \forall ^{0}x(\top \rightarrow \neg Ax\vee Bx)\rightarrow \forall ^{0}x(Ax\rightarrow Bx)$
. By the K-rule for
and the “crucial feature”, we get the result.
Acknowledgements
I’m grateful to two referees for suggesting useful clarifications.