Natural deduction
From Academic Kids

In mathematical logic, natural deduction is the name given to a class of foundational approaches for two key concepts in logic, propositions and proofs. There is no universal agreement on the proper foundation of these concepts; natural deduction takes the approach of mimicking the mental picture of logical reasoning as closely as possible, though in much greater detail than is usual in published mathematics.
Natural deduction grew out of a context of dissatisfaction with sentential axiomatizations common to the systems of Hilbert, Frege and Russell, most famously used by Russell and Whitehead in their mathematical treatise Principia Mathematica. Spurred on by a series of seminars in Poland in 1926 by Łucasiewicz that advocated a more natural treatment of logic, Jaśkowski made the earliest attempts at defining a more natural deduction, first in 1929 using a diagrammatic notation, and later updating his proposal in a sequence of papers in 1934 and 1935. His proposals, though, did not prove to be popular, and natural deduction in its modern form was independently proposed by the German mathematician Gentzen in 1935, in a dissertation delivered to the faculty of mathematical sciences of the university of Göttingen. The term natural deduction (rather, its German form) was coined in that paper:
 Ich wollte zunächst einmal einen Formalismus aufstellen, der dem wirklichen Schließen möglichst nahe kommt. So ergab sich ein „Kalkül des natürlichen Schließens“. (First I wished to construct a formalism that comes as close as possible to actual reasoning. Thus arose a "calculus of natural deduction".)
Gentzen was motivated by a desire to establish the consistency of number theory, and he found immediate use for his natural deduction calculus. He was nevertheless dissatisfied with the complexity of his proofs, and in 1938 gave a new consistency proof using his sequent calculus. In a series of seminars in 1961 and 1962 Prawitz gave a comprehensive summary of natural deduction calculi, and transported much of Gentzen's work with sequent calculi into the natural deduction framework. His 1965 monograph Natural deduction: a prooftheoretical study was to become the definitive work on natural deduction, and included applications for modal and second order logic.
The system presented in this article is a minor variation of Gentzen's or Prawitz's formulation, but with a closer adherence to MartinLöf's description of logical judgements and connectives (MartinLöf, 1996).
Contents 
Judgements and propositions
A judgement is something that is knowable, that is, an object of knowledge. It is evident if one in fact knows it. Thus "it is raining" is a judgement, which is evident for the one who knows that it is actually raining; in this case one may readily find evidence for the judgement by looking outside the window or stepping out of the house. In mathematical logic however, evidence is often not as directly observable, but rather deduced from more basic evident judgements. The process of deduction is what constitutes a proof; in other words, a judgement is evident if one has a proof for it.
The most important judgements in logic are of the form "A is true", where the letter A stands for any specified proposition. Since one cannot reason about the judgement A is true unless one knows that A is a proposition, we require a second and more basic judgement "A is a proposition". Many other judgements have been studied: for example, "A is false", "A is true at time t" (see temporal logic), "A is necessarily true" (see modal logic), "the program M has type τ" (see programming languages and type theory), "A is true by consuming all hypotheses linearly" (see linear logic), and many others. To start with, we shall concern ourselves with the simplest two judgements "A is a proposition" and "A is true", which are abbreviated as "A prop" and "A true" respectively.
The judgement "A prop" defines the structure of valid proofs of A, which in turn defines the structure of propositions. For this reason, the inference rules for this judgement are sometimes known as formation rules. To illustrate, if we have two propositions A and B (that is, the judgements "A prop" and "B prop" are evident), then we form the compound proposition A and B, written symbolically as "A ∧ B". We can write this in the form of an inference rule:
A prop B prop  ∧F A ∧ B prop 
This inference rule is schematic: A and B can be instantiated with any proposition. The general form of an inference rule is:
J_{1} J_{2} ... J_{n}  name J 
where each J_{i} is a judgement and the inference rule is named name. The judgements above the line are known as premisses, and those below the line are conclusions. Other common logical propositions are disjunction (A ∨ B), negation (¬ A), implication (A ⊃ B), and the logical constants truth (T) and falsehood (⊥). Their formation rules are below.
A prop B prop  ∨F A ∨ B prop 
 TF T prop 
 ⊥F ⊥ prop 
A prop  ¬F ¬ A prop 
A prop B prop  ⊃F A ⊃ B prop 
Introduction and elimination
Now we discuss the "A true" judgement. Inference rules that introduce a logical connective in the conclusion are known as introduction rules. To introduce conjunctions, i.e., to conclude "A and B true" for propositions A and B, one requires evidence for "A true" and "B true". As an inference rule:
A true B true  ∧I A ∧ B true 
It must be understood that in such rules the objects are propositions. That is, the above rule is really an abbreviation for:
A prop B prop A true B true  ∧I A ∧ B true 
This can also be written:
A ∧ B prop A true B true  ∧I A ∧ B true 
In this form, the first premiss can be satisfied by the ∧F formation rule, giving the first two premisses of the previous form. In this article we shall elide the "A prop" judgements where they are understood. In the zeroary case, one can derive truth from no premisses.
 TI T true 
If the truth of a proposition can be established in more than one way, the corresponding connective has multiple introduction rules.
A true  ∨I_{1} A ∨ B true 
B true  ∨I_{2} A ∨ B true 
Note that in the zeroary case, i.e., for falsehood, there are no introduction rules. Thus one can never infer falsehood from simpler judgements.
Dual to introduction rules are elimination rules to describe how to deconstruct information about a compound proposition into information about its constituents. Thus, from "A ∧ B true", we can conclude "A true" and "B true":
A ∧ B true  ∧E_{1} A true 
A ∧ B true  ∧E_{2} B true 
As an example of the use of inference rules, consider commutativity of conjunction. If A ∧ B is true, then B ∧ A is true; This derivation can be drawn by composing inference rules in such a fashion that premisses of a lower inference match the conclusion of the next higher inference.
A ∧ B true A ∧ B true  ∧E_{1}  ∧E_{2} B true A true  ∧I B ∧ A true 
Of course we don't in fact know that "A ∧ B true", so this is an incomplete derivation with two unsatisfied premisses.
To eliminate disjunctions, we require a more complex notion of inference rules, as described in the next section.
Hypothetical derivations
A pervasive operation in mathematical logic is reasoning from assumptions. For example, consider the following derivation:
A ∧ (B ∧ C) true  ∧E_{2} B ∧ C true  ∧E_{1} B true 
This derivation does not established the truth of B as such; rather, it establishes the following fact:
 If A ∧ (B ∧ C) is true then B is true.
In logic, one says "assuming A ∧ (B ∧ C) is true, we show that B is true"; in other words, the judgement "B true" depends on the assumed judgement "A ∧ (B ∧ C) true". This is a hypothetical derivation, which we write as follows:
A ∧ (B ∧ C) true ⋮ B true 
The interpretation is: "B true is derivable from A ∧ (B ∧ C) true". Of course, in this specific example we actually know the derivation of "B true" from "A ∧ (B ∧ C) true", but in general we may not apriori know the derivation. The general form of a hypothetical derivation is:
D_{1} D_{2} ... D_{n} ⋮ J 
Each hypothetical derivation has a collection of antecedent derivations (the D_{i}) written on the top line, and a succedent judgement (J) written on the bottom line. Each of the premisses may itself be a hypothetical derivation. (For simplicity, we treat a judgement as a premissless derivation.)
The notion of hypothetical judgement is internalised as the connective of implication. The introduction and elimination rules are as follows.
 u A true ⋮ B true  ⊃I^{u} A ⊃ B true 
A ⊃ B true A true  ⊃E B true 
In the introduction rule, the antecedent named u is discharged in the conclusion. This is a mechanism for delimiting the scope of the hypothesis: its sole reason for existence is to establish "B true"; it cannot be used for any other purpose, and in particular, it cannot be used below the introduction. As an example, consider the derivation of "A ⊃ (B ⊃ (A ∧ B)) true":
 u  w A true B true  ∧I A ∧ B true  ⊃I^{w} B ⊃ (A ∧ B) true  ⊃I^{u} A ⊃ (B ⊃ (A ∧ B)) true 
This full derivation has no unsatisfied premisses; however, subderivations are hypothetical. For instance, the derivation of "B ⊃ (A ∧ B) true" is hypothetical with antecedent "A true" (named u).
With hypothetical derivations, we can now write the elimination rule for disjunction:
 u  w A true B true ⋮ ⋮ A ∨ B true C true C true  ∨E^{u,w} C true 
In words, if A ∨ B is true, and we can derive C true both from A true and from B true, then C is indeed true. Note that this rule does not commit to either A true or B true. In the zeroary case, i.e. for falsehood, we obtain the following elimination rule:
⊥ true  ⊥E C true 
This is read as: if falsehood is true, then any proposition C is true.
Negation is similar to implication.
 u A true ⋮ p true  ¬I^{u,p} ¬ A true 
¬ A true A true  ¬E C true 
The introduction rule discharges both the name of the hypothesis u, and the succedent p, i.e., the proposition p must not occur in the conclusion ¬ A. Since these rules are schematic, the interpretation of the introduction rule is: if from "A true" we can derive for every proposition p that "p true", then A must be false, i.e., "¬ A true". For the elimination, if both A and ¬ A are shown to be true, then there is a contradiction, in which case every proposition C is true. Because the rules for implication and negation are so similar, it should be fairly easy to see that ¬ A and A ⊃ ⊥ are equivalent, i.e., each is derivable from the other.
Consistency, completeness, and normal forms
A logic is said to be consistent if falsehood is not provable (from no assumptions), and is complete if every theorem is provable using the inference rules of the logic. These are statements about the entire logic, and are usually tied to some notion of a model. However, there are local notions of consistency and completeness that are purely syntactic checks on the inference rules, and require no appeals to models. The first of these is local consistency, also known as local reducibility, which says that any derivation containing an introduction of a connective followed immediately by its elimination can be turned into an equivalent derivation without this detour. It is a check on the strength of elimination rules: they must not be so strong that they include knowledge not already contained in its premisses. As an example, consider conjunctions.
 u w A true B true  ∧I A ∧ B true  ∧E_{1} A true  ⇒ 
 u A true 
Dually, local completeness says that the elimination rules are strong enough to decompose a connective into the forms suitable for its introduction rule. Again for conjunctions:
 u A ∧ B true  ⇒ 
 u  u A ∧ B true A ∧ B true  ∧E_{1}  ∧E_{2} A true B true  ∧I A ∧ B true 
These notions correspond exactly to βreduction and ηexpansion in the lambda calculus, using the CurryHoward isomorphism. By local completeness, we see that every derivation can be converted to an equivalent derivation where the principal connective is introduced. In fact, if the entire derivation obeys this ordering of eliminations followed by introductions, then it is said to be normal. In a normal derivation all eliminations happen above introductions. In most logics, every derivation has an equivalent normal derivation, called a normal form. The existence of normal forms is generally hard to prove using natural deduction alone, though such accounts do exist in the literature, most notably by Dag Prawitz in 1961; see his book Natural deduction: a prooftheoretical study, A&W Stockholm 1965, no ISBN. It is much easier to show this indirectly by means of a cutfree sequent calculus presentation.
First and higher order extensions
First_order_natural_deduction.png
The logic of the earlier section is an example of a singlesorted logic, i.e., a logic with a single kind of object: propositions. Many extensions of this simple framework have been proposed; in this section we will extend it with a second sort of individuals or terms. More precisely, we will add a new kind of judgement, "t is a term" (or "t term") where t is schematic. We shall fix a countable set V of variables, another countable set F of function symbols, and construct terms as follows:
v ∈ V  varF v term 
f ∈ F t_{1} term t_{2} term ... t_{n} term  appF f(t_{1}, t_{2}, ..., t_{n}) term 
For propositions, we consider a third countable set P of predicates, and define atomic predicates over terms with the following formation rule:
φ ∈ P t_{1} term t_{2} term ... t_{n} term  predF φ(t_{1}, t_{2}, ..., t_{n}) prop 
In addition, we add a pair of quantified propositions: universal (∀) and existential (∃):
 u x term ⋮ A prop  ∀F^{u} ∀x. A prop 
 u x term ⋮ A prop  ∃F^{u} ∃x. A prop 
These quantified propositions have the following introduction and elimination rules.
 u a term ⋮ [a/x] A true  ∀I^{u,a} ∀x. A true 
∀x. A true t term  ∀E [t/x] A true  
[t/x] A true  ∃I ∃x. A true 
 u  v a term [a/x] A true ⋮ ∃x. A true C true  ∃E^{a,u,v} C true 
In these rules, the notation [t/x] A stands for the substitution of t for every (visible) instance of x in A, avoiding capture; see the article on lambda calculus for more detail about this standard operation. As before the superscripts on the name stand for the components that are discharged: the term a cannot occur in the conclusion of ∀I (such terms are known as eigenvariables or parameters), and the hypotheses named u and v in ∃E are localised to the second premiss in a hypothetical derivation. Although the propositional logic of earlier sections was decidable, adding the quantifiers makes the logic undecidable.
So far the quantified extensions are firstorder: they distinguish propositions from the kinds of objects quantified over. Higher order logic takes a different approach and has only a single sort of propositions. The quantifiers have as the domain of quantification the very same sort of propositions, as reflected in the formation rules:
 u p prop ⋮ A prop  ∀F^{u} ∀p. A prop 
 u p prop ⋮ A prop  ∃F^{u} ∃p. A prop 
A discussion of the introduction and elimination forms for higher order logic is beyond the scope of this article. It is possible to be in between first order and higher order logics. For example, second order logic has two kinds of propositions, one kind quantifying over terms, and the second kind quantifying over propositions of the first kind.
Proofs and typetheory
The presentation of natural deduction so far has concentrated on the nature of propositions without giving a formal definition of a proof. To formalise the notion of proof, we alter the presentation of hypothetical derivations slightly. We label the antecedents with proof variables (from some countable set V of variables), and decorate the succedent with the actual proof. The antecedents or hypotheses are separated from the succedent by means of a turnstile (⊢). This modification sometimes goes under the name of localised hypotheses. The following diagram summarises the change.
 u_{1}  u_{2} ...  u_{n} J_{1} J_{2} J_{n} ⋮ J  ⇒ 
u_{1}:J_{1}, u_{2}:J_{2}, ..., u_{n}:J_{n} ⊢ J 
The collection of hypotheses will be written as Γ when their exact composition is not relevant. To make proofs explicit, we move from the proofless judgement "A true" to a judgement: "π is a proof of A true", which is written symbolically as "π : A true". Following the standard approach, proofs are specified with their own formation rules for the judgement "π proof". The simplest possible proof is the use of a labelled hypothesis; in this case the evidence is the label itself.
u ∈ V  proofF u proof 
 hyp u:A true ⊢ u : A true 
For brevity, we shall leave off the judgemental label true in the rest of this article, i.e., write "Γ ⊢ π : A". Let us reexamine some of the connectives with explicit proofs. For conjunction, we look at the introduction rule ∧I to discover the form of proofs of conjunction: they must be a pair of proofs of the two conjuncts. Thus:
π_{1} proof π_{2} proof  pairF (π_{1}, π_{2}) proof 
Γ ⊢ π_{1} : A Γ ⊢ π_{2} : B  ∧I Γ ⊢ (π_{1}, π_{2}) : A ∧ B 
The elimination rules ∧E_{1} and ∧E_{2} select either the left or the right conjunct; thus the proofs are a pair of projections — first (fst) and second (snd).
π proof  fstF fst π proof 
Γ ⊢ π : A ∧ B  ∧E_{1} Γ ⊢ fst π : A  
π proof  sndF snd π proof 
Γ ⊢ π : A ∧ B  ∧E_{2} Γ ⊢ snd π : B 
For implication, the introduction form localises or binds the hypothesis, written using a λ; this corresponds to the discharged label. In the rule, "Γ, u:A" stands for the collection of hypotheses Γ, together with the additional hypothesis u.
π proof  λF λu. π proof 
Γ, u:A ⊢ π : B  ⊃I Γ ⊢ λu. π : A ⊃ B  
π_{1} proof π_{2} proof  appF π_{1} π_{2} proof 
Γ ⊢ π_{1} : A ⊃ B Γ ⊢ π_{2} : A  ⊃E Γ ⊢ π_{1} π_{2} : B 
With proofs available explicitly, one can manipulate and reason about proofs. The key operation on proofs is the substitution of one proof for an assumption used in another proof. This is commonly known as a substitution theorem, and can be proved by induction on the depth (or structure) of the second judgement.
 Substitution theorem
 If Γ ⊢ π_{1} : A and Γ, u:A ⊢ π_{2} : B, then Γ ⊢ [π_{1}/u] π_{2} : B.
So far the judgement "Γ ⊢ π : A" has had a purely logical interpretation. In type theory, the logical view is exchanged for a more computational view of objects. Propositions in the logical interpretation are now viewed as types, and proofs as programs in the lambda calculus. Thus the interpretation of "π : A" is "the program π has type A". The logical connectives are also given a different reading: conjunction is viewed as product (×), implication as the function arrow (→), etc. The differences are only cosmetic, however. Type theory has a natural deduction presentation in terms of formation, introduction and elimination rules; in fact, the reader can easily reconstruct what is known as simple type theory from the previous sections.
The difference between logic and type theory is primarily a shift of focus from the types (propositions) to the programs (proofs). Type theory is chiefly interested in the convertibility or reducibility of programs. For every type, there are canonical programs of that type which are irreducible; these are known as canonical forms or values. If every program can be reduced to a canonical form, then the type theory is said to be normalising (or weakly normalising). If the canonical form is unique, then the theory is said to be strongly normalising. Normalisability is a rare feature of most nontrivial type theories, which is a big departure from the logical world. (Recall that every logical derivation has an equivalent normal derivation.) To sketch the reason: in type theories that admit recursive definitions, it is possible to write programs that never reduce to a value; such looping programs can generally be given any type. In particular, the looping program has type ⊥, although there is no logical proof of "⊥ true". For this reason, the propositions as types; proofs as programs paradigm only works in one direction, if at all: interpreting a type theory as a logic generally gives an inconsistent logic.
Like logic, type theory has many extensions and variants, including first order and higher order versions. An interesting branch of type theory, known as dependent type theory, allows quantifiers to range over programs themselves. These quantified types are written as Π and Σ instead of ∀ and ∃, and have the following formation rules:
Γ ⊢ A type Γ, x:A ⊢ B type  ΠF Γ ⊢ Πx:A. B type 
Γ ⊢ A type Γ, x:A ⊢ B type  ΣF Γ ⊢ Σx:A. B type 
These types are generalisations of the arrow and product types, respectively, as witnessed by their introduction and elimination rules.
Γ, x:A ⊢ π : B  ΠI Γ ⊢ λx. π : Πx:A. B 
Γ ⊢ π_{1} : Πx:A. B Γ ⊢ π_{2} : A  ΠE Γ ⊢ π_{1} π_{2} : [π_{2}/x] B 
Γ ⊢ π_{1} : A Γ, x:A ⊢ π_{2} : B  ΣI Γ ⊢ (π_{1}, π_{2}) : Σx:A. B 
Γ ⊢ π : Σx:A. B  ΣE_{1} Γ ⊢ fst π : A 
Γ ⊢ π : Σx:A. B  ΣE_{2} Γ ⊢ snd π : [fst π/x] B 
Dependent type theory in full generality is very powerful: it is able to express almost any conceivable property of programs directly in the types of the program. This generality comes at a steep price — checking that a given program is of a given type is undecidable. For this reason, dependent type theories in practice do not allow quantification over arbitrary programs, but rather restrict to programs of a given decidable index domain, for example integers, strings, or linear programs.
Since dependent type theories allow types to depend on programs, a natural question to ask is whether it is possible for programs to depend on types, or any other combination. There are many kinds of answers to such questions. A popular approach in type theory is to allow programs to be quantified over types, also known as parametric polymorphism; of this there are two main kinds: if types and programs are kept separate, then one obtains a somewhat more wellbehaved system called predicative polymorphism; if the distinction between program and type is blurred, one obtains the typetheoretic analogue of higherorder logic, also known as impredicative polymorphism. Various combinations of dependency and polymorphism have been considered in the literature, the most famous being the lambda cube of Henk Barendregt.
The intersection of logic and type theory is a vast and active research area. New logics are usually formalised in a general type theoretic setting, known as a logical framework. Popular modern logical frameworks such as the calculus of constructions and LF are based on higherorder dependent type theory, with various tradeoffs in terms of decidability and expressive power. These logical frameworks are themselves always specified as natural deduction systems, which is a testament to the versatility of the natural deduction approach.
Classical and modal logics
For simplicity, the logics presented so far have been intuitionistic. Classical logic extends intuitionistic logic with an additional axiom or principle of excluded middle:
 For any proposition p, the proposition p ∨ ¬ p is true.
This statement is not obviously either an introduction or an elimination; indeed, it involves two distinct connectives. Gentzen's original treatment of excluded middle prescribed one of the following three (equivalent) formulations, which were already present in analogous forms in the systems of Hilbert and Heyting:
 XM_{1} A ∨ ¬ A true 
¬ ¬ A true  XM_{2} A true 
 u ¬ A true ⋮ p true  XM_{3}^{u,p} A true 
(XM_{3} is merely XM_{2} expressed in terms of ¬E.) This treatment of excluded middle, in addition to being objectionable from a purist's standpoint, introduces additional complications in the definition of normal forms.
A comparatively more satisfactory treatment of classical natural deduction in terms of introduction and elimination rules alone was first proposed by Parigot in 1992 in the form of a classical lambda calculus called λμ. The key insight of his approach was to replace a truthcentric judgement A true with a more classical notion: in localised form, instead of Γ ⊢ A, he used Γ ⊢ Δ, with Δ a collection of propositions similar to Γ. Γ was treated as a conjunction, and Δ as a disjunction. This structure is essentially lifted directly from classical sequent calculi, but the innovation in λμ was to give a computational meaning to classical natural deduction proofs in terms of a callcc or a throw/catch mechanism seen in LISP and its descendants. (See also: first class control.)
Another important extension was for modal and other logics that need more than just the basic judgement of truth. These were first described in a natural deduction style by Prawitz in 1965, and have since accumulated a large body of related work. To give a simple example, the modal logic of necessity requires one new judgement, "A valid", that is categorical with respect to truth:
 If "A true" under no assumptions of the form "B true", then "A valid".
This categorical judgement is internalised as a unary connective ◻A (read "necessarily A") with the following introduction and elimination rules:
A valid  ◻I ◻ A true 
◻ A true  ◻E A true 
Note that the premiss "A valid" has no defining rules; instead, the categorical definition of validity is used in its place. This mode becomes clearer in the localised form when the hypotheses are explicit. We write "Ω;Γ ⊢ A true" where Γ contains the true hypotheses as before, and Ω contains valid hypotheses. On the right there is just a single judgement "A true"; validity is not needed here since "Ω ⊢ A valid" is by definition the same as "Ω;⋅ ⊢ A true". The introduction and elimination forms are then:
Ω;⋅ ⊢ π : A true  ◻I Ω;⋅ ⊢ box π : ◻ A true 
Ω;Γ ⊢ π : ◻ A true  ◻E Ω;Γ ⊢ unbox π : A true 
The modal hypotheses have their own version of the hypothesis rule and substitution theorem.
 validhyp Ω, u:(A valid) ; Γ ⊢ u : A true 
 Modal substitution theorem
 If Ω;⋅ ⊢ π_{1} : A true and Ω, u:(A valid) ; Γ ⊢ π_{2} : C true, then Ω;Γ ⊢ [π_{1}/u] π_{2} : C true.
This framework of separating judgements into distinct collections of hypotheses, also known as multizoned or polyadic contexts, is very powerful and extensible; it has been applied for many different modal logics [5], and also for linear [6] and other substructural logics, to give a few examples.
Comparison with other foundational approaches
Sequent calculus
Main article: sequent calculus
The sequent calculus is the chief alternative to natural deduction as a foundation of mathematical logic. In natural deduction the flow of information is bidirectional: elimination rules flow information downwards by deconstruction, and introduction rules flow information upwards by assembly. Thus, a natural deduction proof does not have a purely bottomup or topdown reading, making it unsuitable for automation in proof search, or even for proof checking (or typechecking in type theory). To address this fact, Gentzen in 1935 proposed his sequent calculus, though he initially intended it as a technical device for improving the consistency of predicate logic, and it is the reformulation by Kleene in his seminal 1952 book Introduction to Metamathematics (ISBN 0720421039) that is the most commonly followed.
In the sequent calculus all inference rules have a purely bottomup reading. Inference rules can apply to elements on both sides of the turnstile. (To differentiate from natural deduction, this article uses a double arrow ⇒ instead of the right tack ⊢ for sequents.) The introduction rules of natural deduction are viewed as right rules in the sequent calculus, and are structurally very similar. The elimination rules on the other hand turn into left rules in the sequent calculus. To give an example, consider disjunction; the right rules are familiar:
Γ ⇒ A  ∨R_{1} Γ ⇒ A ∨ B 
Γ ⇒ B  ∨R_{2} Γ ⇒ A ∨ B 
On the left:
Γ, u:A ⇒ C Γ, v:B ⇒ C  ∨L Γ, w:(A ∨ B) ⇒ C 
Recall the ∨E rule of natural deduction in localised form:
Γ ⊢ A ∨ B Γ, u:A ⊢ C Γ, v:B ⊢ C  ∨E Γ ⊢ C 
The proposition A ∨ B, which is the succedent of a premiss in ∨E, turns into a hypothesis of the conclusion in the left rule ∨L. Thus, left rules can be seen as a sort of inverted elimination rule. This observation can be illustrated as follows:
natural deduction  sequent calculus  

 hyp   elim. rules  ↓  ↑↓ meet ↑   intro. rules  conclusion  ⇒ 
 init ↑ ↑    left rules  right rules   conclusion 
In the sequent calculus, the left and right rules are performed in lockstep until one reaches the initial sequent, which corresponds to the meeting point of elimination and introduction rules in natural deduction. These initial rules are superficially similar to the hypothesis rule of natural deduction, but in the sequent calculus they describe a transposition or a handshake of a left and a right proposition:
 init Γ, u:A ⇒ A 
The correspondence between the sequent calculus and natural deduction is a pair of soundness and completeness theorems, which are both provable by means of an inductive argument.
 Soundness of ⇒ wrt. ⊢
 If Γ ⇒ A, then Γ ⊢ A.
 Completeness of ⇒ wrt. ⊢
 If Γ ⊢ A, then Γ ⇒ A.
It is clear by these theorems that the sequent calculus does not change the notion of truth, because the same collection of propositions remain true. Thus, one can use the same proof objects as before in sequent calculus derivations. As an example, consider the conjunctions. The right rule is virtually identical to the introduction rule
sequent calculus  natural deduction  

Γ ⇒ π_{1} : A Γ ⇒ π_{2} : B  ∧R Γ ⇒ (π_{1}, π_{2}) : A ∧ B 
Γ ⊢ π_{1} : A Γ ⊢ π_{2} : B  ∧I Γ ⊢ (π_{1}, π_{2}) : A ∧ B 
The left rule, however, performs some additional substitutions that are not performed in the corresponding elimination rules.
sequent calculus  natural deduction  

Γ, u:A ⇒ π : C  ∧L_{1} Γ, v:(A ∧ B) ⇒ [fst v/u] π : C 
Γ ⊢ π : A ∧ B  ∧E_{1} Γ ⊢ fst π : A  
Γ, u:B ⇒ π : C  ∧L_{2} Γ, v:(A ∧ B) ⇒ [snd v/u] π : C 
Γ ⊢ π : A ∧ B  ∧E_{2} Γ ⊢ snd π : B 
The kinds of proofs generated in the sequent calculus are therefore rather different from those of natural deduction. The sequent calculus produces proofs in what is known as the βnormal ηlong form, which corresponds to a canonical representation of the normal form of the natural deduction proof. If one attempts to describe these proofs using natural deduction itself, one obtains what is called the intercalation calculus (first described by John Byrnes [3]), which can be used to formally define the notion of a normal form for natural deduction.
The substitution theorem of natural deduction takes the form of a structural rule or structural theorem known as cut in the sequent calculus.
 Cut (substitution)
 If Γ ⇒ π_{1} : A and Γ, u:A ⇒ π_{2} : C, then Γ ⇒ [π_{1}/u] π_{2} : C.
In most well behaved logics, cut is unnecessary as an inference rule, though it remains provable as a metatheorem; the superfluousness of the cut rule is usually presented as a computational process, known as cut elimination. This has an interesting application for natural deduction; usually it is extremely tedious to prove certain properties directly in natural deduction because of an unbounded number of cases. For example, consider showing that a given proposition is not provable in natural deduction. A simple inductive argument fails because of rules like ∨E or ¬E which can introduce arbitrary propositions. However, we know that the sequent calculus is complete with respect to natural deduction, so it is enough to show this unprovability in the sequent calculus. Now, if cut is not available as an inference rule, then all sequent rules either introduce a connective on the right or the left, so the depth of a sequent derivation is fully bounded by the connectives in the final conclusion. Thus, showing unprovability is much easier, because there are only a finite number of cases to consider, and each case is composed entirely of subpropositions of the conclusion. A simple instance of this is the global consistency theorem: "⋅ ⊢ ⊥ true" is not provable. In the sequent calculus version, this is manifestly true because there is no rule that can have "⋅ ⇒ ⊥" as a conclusion! Proof theorists often prefer to work on cutfree sequent calculus formulations because of such properties.
References
Template:Journal reference issue Lecture notes to a short course at Università degli Studi di Siena, April 1983. Template:Journal reference issue
External links
 Introduction to natural deduction (http://www.danielclemente.com/logica/dn.en.html)
 Domino On Acid (http://www.winterdrache.de/freeware/domino/). Natural deduction visualized as a game of dominoes.de:Systeme natürlichen Schließens