1 Non-classical Logic

There are no non-experienced truths – L.E.J. Brouwer

1.1 Intuitionistic Logic

Idea: a proof of φψ is a “procedure” that comments a proof of φ into a proof of ψ.

In particular, ¬¬φ is not always the same as φ.

Fact: The law of excluded middle (φ¬φ) is not generally intuitionistically valid.

Moreover, the Axiom of Choice is incompatible with intuitionistic set theory.

We take choice to mean that any family of inhabited sets admits a choice function.

Theorem 1.1.1 (Diaconescu). The law of excluded middle can be intuitionistically deduced from the Axiom of Choice.

Proof. Let φ be a proposition. By the Axiom of Separation, the following are sets (i.e. we can construct a proof that they are sets):

A:={x{0,1}:φ(x=0)}B:={x{0,1}:φ(x=1)}.

As 0A and 1B, we have that {A,B} is a family of inhabited sets, thus admits a choice function f:{A,B}AB by the Axiom of Choice. This satisfies f(A)A and f(B)B by definition.

Thus we have

(f(A)=0φ)(f(B)=1φ)

and f(A),f(B){0,1}. Now f(A){0,1} means that (f(A)=0)(f(A)=1) and similarly for f(B).

We can have the following:

  • (1) We have a proof of f(A)=1, so φ(1=0) has a proof, so we must have a proof of φ.
  • (2) We have a proof of f(B)=0, which similarly gives a proof of φ.
  • (3) We have f(A)=0 and f(B)=1, in which case we can prove ¬φ: given a proof of ϕ, we can prove that A=B (by Extensionality), in which case 0=f(A)=f(B)=1, a contradiction.

So we can always specify a proof of φ or a proof of φ or a proof of ¬φ.

Why bother?

Let’s try to formalise the BHK interpretation of logic.

We will inductively define a provability relation by enforcing rules that implement the BHK interpretation.

We will use the notation Γφ to mean that φ is a consequence of the formulae in the set Γ.

Rules for Intuitionistic Propositional Calculus (IPC)

We obtain classical propositional logic (CPC) by adding either:

By

[A]X[B]YC(A,B)

we mean ‘if we can prove X assuming A and we can prove Y assuming B, then we can infer C by “discharging / closing” the open assumptions A and B’.

In particular, the (-I)-rule can be written as

Γ,[A]BΓAB(A).

We obtain intiuitionistic first-order logic (IQC) by adding rules for quantification:

Example 1.1.2. Let’s give a natural deduction proof of ABBA.

[AB]A[AB]BBAABBA(AB).

Example 1.1.3. Let’s prove the Hilbert-style axioms φ(ψφ) and (φ(ψχ))((φpsi)(φχ)).

[φ]  [ψ]ψφ(ψ)φ(ψφ)(φ)

[φ(ψχ)][φψ][φ]   (toE)


ψχψ   (toE)


χ   (toI,ψ)


φχ   (toI, φψ)


(φψ)(φχ)   (toI, (φ(ψχ)))


(φ(ψχ))((φψ)(φχ))

If Γ is a set of propositions in the language and φ is a proposition, we write ΓIPCφ, ΓIQCφ, ΓCPCφ, ΓCQCφ, if there is a proof of φ from Γ in the respective logic.

Lemma 1.1.4. Assuming that:

  • ΓIPCφ

  • ψ a proposition

Then Γ,ψIPCφ. Moreover, if p is a primitive proposition and ψ is any proposition, then
Γ[p:=ψ]IPCφ[p:=ψ].

Proof. Induction over the size of proofs.

1.2 The simply typed λ-calculus

For now we assume given a set Π of simple types generated by a grammar

Π:=U|ΠΠ,

where U is a countable set of type variables, as well as an infinite set V of variables.

Definition 1.2.1 (Simply typed lambda-term). The set ΛΠ of simply typed λ-terms is defined by the grammar

ΛΠ:=Vvariables|λV:Π.ΛΠλ-abstraction|ΛΠΛΠλ-application.

A context is a set of pairs {x1:τ1,,xn:τn} where the xi are (distinct) variables and each τiΠ. We write C for the set of all possible contexts. Given a context ΓC, we also write Γ,x:τ for the context Γ{x:τ} (if x does not appear in Γ).

The domain of Γ is the set of variables that occur in it, and the range |Γ| is the set of types that it manifests.

Definition 1.2.2 (Typability relation). We define the typability relation C×ΛΠ×Π via:

  • (1)
    For every context Γ, and variable x not occurring in Γ, and type τ, we have Γ,x:τx:τ.
  • (2)
    Let Γ be a context, x a variable not occurring in Γ, and let σ,τΠ be types, and M be a λ-term. If Γ,x:σM:τ, then Γ(λx:σ.M):(στ).
  • (3)
    Let Γ be a context, σ,τΠ be types, and M,NΛΠ be terms. If ΓM:(στ) and ΓN:σ, then Γ(MN):τ.

Notation. We will refer to the λ-calculus of ΛΠ with this typability relation as λ().

A variable x occurring in a λ-abstraction λx̲:σ.M is bound, and it is free otherwise. We say that terms M and N are α-equivalent if they differ only in the names of the bound variables.

If M and N are λ-terms and x is a variable, then we define the substitution of N for x in M by:

Definition 1.2.3 (beta-reduction). The β-reduction relation is the smallest relation β on ΛΠ closed under the following rules:

  • (λx:σ.P)QβP[x:=Q],

  • if PβP, then for all variables x and types σΠ, we have λx:σ.Pβλx:σ.P,

  • PβP and z as a λ-term, then PZβPZ and ZPβZP.

We also define β-equivalence β as the smallest equivalence relation containing β.

Example 1.2.4 (Informal). We have (λx:.(λy:τ.x))Zβ(λy:τ.Z).

When we reduce (λx:σ.P)Q, the term being reduced is called a β-redex, and the result is its β-contraction.

Lemma 1.2.5 (Free variables lemma). Assuming that:

  • ΓM:σ

Then
  • (1) If ΓΓ, then ΓM:σ.
  • (2) The free variables of M occur in Γ.
  • (3) There is a context ΓΓ comprising exactly the free variables in M, with ΓM:σ.

Proof. Exercise.

Lemma 1.2.6 (Generation Lemma).

  • (1)
    For every variable x, context Γ, and type σ, if Γx:σ, then x:σΓ;
  • (2)
    If Γ(MN):σ, then there is a type τ such that ΓM:τσ and ΓN:τ;
  • (3)
    If Γ(λx.M):σ, then there are types τ and ρ such that Γ,x:τM:ρ and σ=(τρ).

Lemma 1.2.7 (Substitution Lemma).

  • (1)
    If ΓM:σ and α is a type variable, then Γ[α:=τ]M:σ[α:=τ];
  • (2)
    If Γ,x:τM:σ and ΓN:τ, then ΓM[x:=N]:σ.

Proposition 1.2.8 (Subject reduction). Assuming that:

  • ΓM:σ

  • MβN

Then ΓN:σ.

Proof. By induction on the derivation of MβN, using Lemma 1.2.6 and Lemma 1.2.7.

Notation. We will write MβN if M reduces to N after (potentially multiple) β-reductions.

Theorem 1.2.9 (Church-Rosser for lambda(->)). Assuming that:

  • ΓM:σ

  • MβN1

  • MβN2

Then there is a λ-term L such that N1βL, N2βL, and ΓL:σ.

Pictorially:

          M


  N1               N2


ββββ         L

Definition (beta-normal form). A λ-term M is in β-normal form if there is no term N such that MβN.

Corollary 1.2.10 (Uniqueness of normal form). If a simply typed λ-term admits a β-normal form, then it is unique.

Proposition 1.2.11 (Uniqueness of types).

  • (1)
    If ΓM:σ and ΓM:τ, then σ=τ.
  • (2)
    If ΓM:σ, ΓN:τ, and MβN, then σ=τ.

Proof.

Example 1.2.12. There is no way to assign a type to λx:x.x. If x is of type τ, then in order to apply x to x, it has to be of type τσ for some σ. But ττσ.

PIC

Definition 1.2.13 (Height). The height function is the recursively defined map h:Π that maps a type variable to 0, and a function type στ to 1+max(h(σ),h(τ)).

We extend the height function from types to β-redexes by taking the height of its λ-abstraction.

Not.: (λx:σ.Pτ)στRσ.

Theorem 1.2.14 (Weak normalisation for lambda(->)). Assuming that:

  • ΓM:σ

Then there is a finite reduction path M:=M0βM1βM2ββMn, where Mn is in β-normal form.

Proof (“Taming the Hydra”). The idea is to apply induction on the complexity of M. Define a function m:ΛΠ× by

m(M)={(0,0)if M is in β-normal form(h(M),redex(M))otherwise,

where h(M) is the greatest height of a redex in M, and redex(M) is the number of redexes in M of that height.

We will use induction over ω×ω to show that if M is typable, then it admits a reduction to β-normal form.

Problem: reductions can copy redexes or create new ones.

Strategy: always reduce the right most redex of maximum height.

We will argue that by following this strategy, any new redexes we generate have to be lower than the height of the redex we picked to reduce.

PIC

If ΓM:σ and M is already in β-normal form, then claim is trivially true. If M is not in β-normal form, let Δ be the rightmost redex of maximal height h.

By reducing Δ, we may introduce copies of existing redexes, or create new ones. Creation of new redexes of Δ has to happen in one of the following ways:

  • (1) If Δ is of the form (λx:(ρμ)xPρ)(λy:ρ.Qμ)Pμ, then it reduces to (λy:ρ.Qμ)ρμPμ, in which case there is a new redex of height h(ρμ)<h.
  • (2) We have Δ=(λx:τ.(λy:ρ.Rμ))Pτ occurring in M in the scenario ΔρμQρ. Say Δ reduces to λy:ρ.R1μ. Then we create a new redex of height h(ρμ)<h(τ(ρμ))=h.
  • (3) The last possibility is that Δ=(λx:(ρμ).x)(λy:ρ.Pμ), and that it occurs in M as ΔρμQρ. Reduction then gives the redex (λy:ρ.Pμ)ρμQρ of height h(ρμ)<h.

Now Δ itself is gone (lowering the count by 1), and we just showed that any newly created redexes have height <h.

If we have Δ=(λx:τ.Pρ)Qτ and P contains multiple free occurrences of x, then all the redexes in Q are multiplied when performing β-reduction.

However, our choice of Δ ensures that the height of any such redex in Q has height <h, as they occur to the right of Δ in M. It is this always the case that m(M)<m(M) (in the lexicographic order), so by the induction hypothesis, M can be reduced to β-normal form (and thus so can M).

Theorem 1.2.15 (Strong Normalisation for lambda(->)). Assuming that:

  • ΓM:σ

Then there is no infinite reduction sequence MβM1β.

Proof. See Example Sheet 1.

1.3 The Curry-Howard Correspondence

Propositions-as-types: idea is to think of φ as the “type of its proofs”.

The properties of the STλC match the rules of IPC rather precisely.

First we will show a correspondence between λ() and the implication fragment IPC() of IPC that includes only the connective, the axiom scheme, and the (I) and (E) rules. We will later extend this to the whole of IPC by introducing more complex types to λ().

Start with IPC() and build a STλC out of it whose set of type variables U is precisely the set of primitive propositions of the logic.

Clearly, the set Π of types then matches the set of propositions in the logic.

Comment: λx:σ.(Mx)ηM if x is not free in M.

Proposition 1.3.1 (Curry-Howard for IPC(->)). Assuming that:

  • Γ is a context for λ()

  • φ a proposition

Then
  • (1) If ΓM:φ, then |Γ|={τΠ:(x:τ)Γ for some x}IPC()φ
  • (2) If |Γ|IPC()φ, then there is a simply typed λ-term Mλ() such that {(xψ:ψ)|ψΓ}M:φ.

Proof.

  • (1) We induct over the derivation of ΓM:φ.

    If x is a variable not occurring in Γ and the derivation is of the form Γ,x:φx:φ, then we’re supposed to prove that |Γ,x:φ|φ. But that follows from φφ as |Γ,x:φ|=|Γ|{φ}.

    If the derivation has M of the form λx:σ.N and φ=στ, then we must have Γ,x:σN:τ. By the induction hypothesis, we have that |Γ,x:σ|τ, i.e. |Γ|,στ. But then |Γ|στ by (-I).

    If the derivation has the form Γ(PQ):φ, then we must have ΓP:(σφ) and ΓQ:σ. By the induction hypothesis, we have that |Γ|σφ and |Γ|σ, so |Γ|φ by (-E).

  • (2) Again, we induct over the derivation of Γφ. Write Δ={(xψ:ψ)|ψΓ}. Then we only have a few ways to construct a proof at a given stage. Say the derivation is of the form Γ,φφ. If φΓ, then clearly Δxφ:φ, and if φΓ then Δ,xφ:φxφ:φ.

    Suppose the derivation is at a stage of the form

    Γφψ  ΓφΓψ.

    Then by the induction hypothesis, there are λ-terms M and N such that ΔM:(φψ) and ΔN:φ, from which Δ(MN):φ.

    Finally, if the stage is given by

    Γ,φψΓφψ,

    then we have two sub-cases:

    • If φΓ, then the induction hypothesis gives ΔM:ψ for some term M. By weakening, we have Δ,x:φM:ψ, where x does not occur in Δ. But then Δ(λx:φ.M):(φψ) as needed.

    • If φΓ, then the induction hypothesis gives Δ,xφ:φM:ψ for some M, thus Δ(λxφ:φ.M):(φψ) as needed.

Example 1.3.2. Let φ,ψ be primitive propositions. The λ-term

λf:(φψ)φ.λ:φψ.g(fgφ)ψ

has type ((φψ)φ)((φψ)ψ), and therefore encodes a proof of that proposition in IPC().

g:φψ, f:(φψ)φ.

g:[φψ]f:[(φψ)φ]   (toE)


fg:φg:[φψ]   (toE)


g(fg):ψ   (toI, φψ)


λg.g(fg):(φψ)ψ   (toI, (φψ) φ)


λf.λg.g(fg):((φψ)φ)((φψ)ψ)

Definition 1.3.3 (Full STlambdaC). The types of the full simply typed λ-calculus are generated by the following grammar:

Π:=U|ΠΠ|Π×Π|Π+Π|0|1,

where U is a set of type variables (usually countable).

Its terms are given by ΛΠ given by:

ΛΠ:=V|λV:Π.ΛΠ|ΛΠΛΠ|Π1(ΛΠ)|Π2(ΛΠ)|ι1(ΛΠ)|ι2(ΛΠ)|case(ΛΠ;V.ΛΠ;V.ΛΠ)||!ΠΛΠ,

where V is an infinite set of variables, and is a constant.

We have new typing rules:

They come with new reduction rules:

When setting up Curry-Howard with these new types, we let:

Example 1.3.4. Consider the following proof of (φχ)(ψφ):

[φχ]φ[ψ]   ()


ψφ   ()


(φχ)(ψφ)

We decorate this proof by turning the assumptions into variables and following the Curry-Howard correspondence:

[φχ]:pφ:π1(p)[ψ]:b   ()


ψφ:λb:ψ.π1(p)   ()


(φχ)(ψφ)

STλC IPC


(primitive) types (primitive) propositions
variable hypothesis
STλ-term proof
type constructor logical connective
term inhabitation provability
term reduction proof normalisation

1.4 Semantics for IPC

Definition 1.4.1 (Lattice). A lattice is a set L equipped with binary commutative and associative operations and that satisfy the absorption laws:

a(ab)=a;a(ab)=a,

for all a,bL.

A lattice is:

  • Distributive if a(bc)=(ab)(ac) for all a,b,cL.

  • Bounded if there are elements ,L such that a=a and a=a.

  • Complemented if it is bounded and for every aL there is aL such that aa= and aa=.

A Boolean algebra is a complemented distributive lattice.

Note that and are idempotent in any lattice. Moreover, we can define an ordering on L by setting ab if ab=a.

Example 1.4.2.

Proposition 1.4.3. Assuming that:

  • is the order induced by the operations in L (ab if ab=a)

Then is a partial order with least element , greatest element , and for any a,bL, we have ab=inf{a,b} and ab=sup{a,b}. Conversely, every partial order with all finite infs and sups is a bounded lattice.

Proof. Exercise.

Classically, we say that Γt if for every valuation v:L{0,1} with v(p)=1 for all pΓ we have v(t)=1.

We might want to replace {0,1} with some other Boolean algebra to get a semantics for IPC, with an accompanying Completeness Theorem. But Boolean algebras believe in the Law of Excluded Middle!

Definition 1.4.4 (Heyting algebra). A Heyting algebra is a bounded lattice equipped with a binary operation :H×HH such that

abca(bc)

for all a,b,cL.

A morphism of Heyting algebras is a function that preserves all finite meets, finite joins, and .

Example 1.4.5.

Definition 1.4.6 (Valuation in Heyting algebras). Let H be a Heyting algebra and L be a propositional language with a set P of primitive propositions. An H-valuation is a function v:PH, extended to the whole of L recursively by setting:

  • v()=,

  • v(AB)=v(A)v(B),

  • v(AB)=v(A)v(B),

  • v(AB)=v(A)v(B).

A proposition A is H-valid if v(A)= for all H-valuations v, and is an H-consequence of a (finite) set of propositions Γ if v(Γ)v(A) for all H-valuations v (written ΓHA).

Lemma 1.4.7 (Soundness of Heyting semantics). Assuming that:

Then ΓIPCA implies ΓHA.

Proof. By induction over the structure of the proof ΓA.

  • (Ax) As v((Γ)A)=v()v(A)v(A) for any Γ and A.
  • (-I) A=BC and we have derivations Γ1B, Γ2C, with Γ1,Γ2Γ. By the induction hypothesis, we have v(Γ)v(Γ1)v(Γ2)v(B)v(C)=v(BC)=v(A), i.e. ΓHA.
  • (-I) A=BC and so we must have Γ{B}C. By induction hypothesis, we have v(Γ)v(B)=v(γB)v(C). By the definition of , this implies v(Γ)[v(B)v(C)]=v(BC)=v(A), i.e. ΓHA.
  • (-I) A=BC and without loss of generality we have a derivation ΓB. By the induction hypothesis we have v(Γ)v(B), but v(BC)=v(B)v(C), and hence v(B)v(BC)=v(A).
  • (-E) By the induction hypothesis, we have v(Γ)v(BC)=v(B)v(C)v(B),v(B).
  • (-E) We know that v(AB)=(v(A)v(B)). From v(AB)v(A)v(B), we derive v(A)v(AB)v(B) by definition of . So if v(Γ)v(AB) and v(Γ)v(A), then v(Γ)v(B), as needed.
  • (-E) By induction hypothesis: v(AΓ)v(C), v(BΓ)v(C) and v(Γ)v(AB)=v(A)v(B). This last fact means that v(Γ)(v(A)v(B))=v(Γ). Now this is the same as (v(Γ)v(A))(v(Γ)v(B)) as Heyting algebras are distributive lattices (see Example Sheet 2), and this is v(C) by the first two inequalities of this paragraph.
  • (-E) If v(Γ)v()=, then v(Γ)=, in which case v(Γ)v(A) for any A by minimality of in H.

Example 1.4.8. The Law of Excluded Middle is not intuitionistically valid. Let p be a primitive proposition and consider the Heyting algebra given by the topology {,{1},{1,2}} on {1,2}.

We can define a valuation v with v(p)={1}, in which case v(¬p)=¬{1}=int(X{1})=.

So v(p¬p)={1}={1}. Thus Soundness of Heyting semantics implies that IPCp¬p.

Example 1.4.9. Peirce’s Law ((pq)p)p is not intuitionistically valid.

Take the valuation on the usual topology of 2 that maps p to 2{(0,0)} and q to .

Classical completeness: ΓCPCA if and only if Γ2A.

Intuitionistic completeness: no single finite replacement for 2.

Definition (Lindenbaum-Tarski algebra). Let Q be a logical doctrine (CPC, IPC, etc), L be a propositional language, and T be an L-theory. The Lindenbaum-Tarski algebra FQ(T) is built in the following way:

  • The underlying set of FQ(T) is the set of equivalence classes [φ] of propositions φ, where φψ when T,φQψ and T,ψQφ;

  • If is a logical connective in the fragment Q, we set [φ][ψ]:=[φψ] (should check well-defined: exercise).

We’ll be interested in the case Q=CPC, Q=IPC, and Q=IPC{}.

Proposition 1.4.10. The Lindenbaum-Tarski algebra of any theory in IPC{} is a distributive lattice.

Proof. Clearly, and inherit associativity and commutativity, so in order for FIPC{}(T) to be a lattice we need only to check the absorption laws:

[φ][φψ]=[φ](α)[φ][φψ]=[φ](β)

Equation (α) is true since T,φIPC{}φ(φψ) by (-I), and also T,φ(φψ)IPC{}φ by (-E). Equation (β) is similar.

Now, for distributivity: T,φ(ψχ)(φψ)(φχ) by (-E) followed by (-E):

φ(ψχ)   (-E)


φψχ   (-E)


(φψ)(φχ)

Conversely, T,((φψ)(φχ))φ(ψχ) by (-E) followed by (-I).

Lemma 1.4.11. The Lindenbaum-Tarski algebra of any theory relative to IPC is a Heyting algebra.

Proof. We already saw that FIPC(T) is a distributive lattice, so it remains to show that [φ][ψ]:=[φψ] gives a Heyting implication, and that FIPC(T) is bounded.

Suppose that [φ[ψ][χ], i.e. τ,φψIPCχ. We want to show that [φ][ψχ], i.e. τ,φ(ψχ). But that is clear:

φ[ψ]  


φψ   (hyp)


χ   (-I, ψ)


ψχ

Conversely, if τ,φ(ψχ), then we can prove τ,φψχ:

φψ   (-E)


φψ   (hyp)


ψχψ   (-E)


χ

So defining [φ][ψ]:=[φψ] provides a Heyting .

The bottom element of FIPC(T) is just []: if [φ] is any element, then T,IPCφ by -E.

The top element is :=[: if φ is any proposition, then [φ][] via

φ[]   (-E)


 


Theorem 1.4.12 (Completeness of the Heyting semantics). A proposition is provable in IPC if and only if it is H-valid for every Heyting algebra H.

Proof. One direction is easy: if IPCφ, then there is a derivation in IPC, thus v(φ) for any Heyting algebra H and valuation v, by Soundness of Heyting semantics. But then v(φ)= and φ is H-valid.

For the other direction, consider the Lindenbaum-Tarski algebra F(L) of the empty theory relative to IPC, which is a Heyting algebra by Lemma 1.4.11. We can define a valuation v by extending PF(L), p[p] to all propositions.

As v is a valuation, it follows by induction (and the construction of F(L)) that v(φ)=[φ] for all propositions.

Now φ is valid in every Heyting algebra, and so is valid in F(L) in particular. So v(φ)==[φ], hence IPCφ, hence IPCφ.

Given a poset S, we can construct sets a:={sS:as} called principal up-sets.

Recall that US is a terminal segment if aU for each aU.

Proposition 1.4.13. If S is a poset, then the set T(S)={US:U is a terminal segment of S} can be made into a Heyting algebra.

Proof. Order the terminal segments by . Meets and joins are and , so we just need to define . If U,VT(S), define (UV):={sS:(s)UV}.

If U,V,WT(S), we have

W(UV)(w)UVwW,

which happens if for every wW and uU we have wuuV. But W is a terminal segment, so this is the same as saying that WUV.

Definition 1.4.14 (Kripke model). Let P be a set of primitive propositions. A Kripke model is a tuple (S,,) where (S,) is a poset (whose elements are called “worlds” or “states”, and whose ordering is called the “accessibility relation”) and S×P is a binary relation (“forcing”) satisfying the persistence property: if pP is such that sp and ss, then sp.

Every valuation v on T(S) induces a Kripke model by setting sp is sv(p).

Definition 1.4.15 (Forcing relation). Let (S,,) be a Kripke model for a propositional language. We define the extended forcing relation inductively as follows:

  • There is no sS with s;

  • sφψ if and only if sφ and sψ;

  • sφψ if and only if sφ or sψ;

  • s(φψ) if and only if sφ implies sψ for every ss.

It is easy to check that the persistence property extends to arbitrary propositions.

Moreover:

Notation. Sφ for φ a proposition if all worlds in S force φ.

Example 1.4.16. Consider the following Kripke models:

  • (1)
     ′
s ⊩p

s
  • (2)
    s′              s′′⊩p

       s
  • (3)
    s′⊩p,⊩q


s

In (1), we have s¬p, since ss and sp. We also know that sp, thus sp¬p.

It is also the case that s¬¬p, yet sp, so s(¬¬pp) either.

In (2), s¬¬p since ss can’t access a world that forces p. Also s¬p either, as ss forces p. So s¬¬p¬p.

In (3), s(pq)(¬pq). All worlds force pq, and sq. So to check the claim we just need to verify that s¬p. But that is the case, as ss and sp.

Definition 1.4.17 (Filter). A filter F on a lattice L is a subset of L with the following properties:

  • F

  • F is a terminal segment of L (i.e., if fx and fF, then xF)

  • F is closed under finite meets

Example 1.4.18.

  • (1) Given an element jI of a set I, then the family Fj of all subsets of I containing j is a filter on P(I). Such a filter is called a principal filter.
  • (2) The family of all cofinite subsets of I is a filter on P(I), the Fréchet filter.

    Exercise: a maximal proper filter (known as an ultra filter) is not principal if and only if it contains the Fréchet filter.

  • (3) The family of all subsets of [0,1] with Lebesgue measure 1 is a filter.

A filter is proper if FL.

A filter F on a Heyting algebra is prime if it is proper and satisfies: whenever (xy)F, we can conclude that xF or yF.

If F is a proper filter and xF, then there is a prime filter extending F that still doesn’t contain x (by Zorn’s Lemma).

Lemma 1.4.19. Assuming that:

Then there is a Kripke model (S,,) such that vHφ if and only if Sφ, for every proposition φ.

Proof (sketch). Let S be the set of all prime filters of H, ordered by inclusion. We write Fp if and only if v(p)F for primitive propositions p.

We prove by induction that Fφ if and only if v(φ)F for arbitrary propositions.

For the implication case, say that F(ψψ) and v(ψψ)=[v(ψ)v(ψ)]F. Let G be the least filter containing F and v(ψ). Then

G={b:(fF)(fv(ψ)b)}.

Note that v(ψ)G, or else fv(ψ)v(ψ) for some fF, whence fv(ψψ) and so v(ψψ)F (as F is a terminal segment).

In particular, G is proper. So let G be a prime filter extending G that does not contain v(ψ) (exists by Zorn’s lemma).

By the induction hypothesis, Gψ, and since F(ψψ) and G (this G) contains F, we have that Gψ. But then v(ψ)G, contradiction.

This settles that F(ψψ) implies v(ψψ)F.

Conversely, say that v(ψψ)FGψ. By the induction hypothesis, v(ψ)G, and so v(ψ)v(ψ)G (as FG). But then v(ψ)v(ψ)(v(ψ)v(ψ))G, as G is a filter.

So the induction hypothesis gives Gψ, as needed.

The cases for the other connectives are easy ( needs primality). So (S,,) is a Kripke model. Want to show that vHφ if and only if Sφ, for each φ.

Conversely, say Sφ, but vHφ. Since v(φ), there must be a proper filter that does not contain it. We can extend it to a prime filter G that does not contain it, but then Gφ, contradiction.

Theorem 1.4.20 (Completeness of the Kripke semantics). Assuming that:

  • φ a proposition

Then ΓIPCφ if and only if for all Kripke models (S,,), the condition SΓ implies Sφ.

Proof. Soundness: induction over the complexity of φ.

Adequacy: Say ΓIPCφ. Then vHΓ but vHφ for some Heyting algebra H and H-valuation v (Theorem 1.4.12). But then Lemma 1.4.19 applied to Hand v provides a Kripke model (S,,) such that SΓ, but Sφ, contradicting the hypothesis on every Kripke model.

1.5 Negative translations

Definition 1.5.1 (Double-negation translation). We recursively define the ¬¬-translation φN of a proposition φ in the following way:

  • If p is a primitive proposition, then pN:=¬¬p;

  • (φψ)N:=φNψN

  • (φψ)N:=φNψN

  • (φψ)N:=¬(¬φN¬ψN)

  • (¬φ)N:=¬φN

Lemma 1.5.2. Assuming that:

Then the map ¬¬:HH preserves and .

Proof. Example Sheet 2.

Lemma 1.5.3 (Regularisation). Assuming that:

Then

Proof.

  • (1) Give H¬¬:={xH:¬¬x=x} the inherited order, so that , , and (which are preserved by ¬¬) remain the same. We just need to define disjunctions in H¬¬ properly.

    Define a¬¬b:=¬¬(ab) in H. It is easy to show that this gives sup{a,b} in H¬¬ (as ¬¬ preserves order), so H¬¬ is a Heyting algebra.

    As every element of H¬¬ is regular (i.e. ¬¬x=x), it is a Boolean algebra (see Example Sheet 2).

  • (2) Given a Heyting homomorphism g:HB, where B is a Boolean algebra, define g¬¬:HB as gH¬¬. It clearly preserves ,,,, as those operations in H¬¬ are inherited from H.

    But we also have

    g¬¬(a¬¬b)=g|H¬¬(¬¬(ab))=¬¬(g(a)g(b))=g(a)g(b)B is Boolean=g¬¬(a)g¬¬(b)

    Thus g¬¬ is a morphism of Boolean algebras. Note that any xH provides an element ¬¬xH¬¬, since ¬¬¬¬x=¬¬x in H. Additionally,

    g¬¬(¬¬x)=g(¬¬x)=¬¬g(x)=g(x)

    for all xH (as g(x) is in a Boolean algebra).

    Now, if h:H¬¬B is a morphism of Boolean algebras with g(x)=h(¬¬x) for all xH, then h(a)=h(¬¬a)=g(a)=g¬¬(a) for all aH. So g¬¬ is unique with this property.

In particular, if S is a set, then FHeyt(S)¬¬FBool(S).

Theorem 1.5.4 (Glivenko’s Theorem). Assuming that:

  • φ and ψ are propositions

Then CPCφψ if and only if IPC¬¬φ¬¬ψ.

Proof.

  • If CPCφψ, then φψ in FBool(L)=FHeyt(L)¬¬. As the inclusion i:FHeyt(L)¬¬FHeyt(L) strictly preserves and , it follows that i()i(φψ)=φψ=¬¬(φψ)as φψFHeyt(L)¬¬=¬¬φ¬¬ψ

    in FHeyt(L), so IPC¬¬φ¬¬ψ.

  • Obvious.

Corollary 1.5.5. Assuming that:

  • φ a proposition

Then CPCφ if and only if IPCφN.

Proof. Induction over the complexity of formulae.

Corollary 1.5.6. CPC is inconsistent if and only if IPC is inconsistent.

Proof.

  • If CPC is inconsistent, then there is φ such that CPCφ and IPC¬φ. But then IPC¬¬φ and IPC¬φ, so IPC.
  • Obvious.