In
In Proceedings
of the Aristotelian Society, New Series, Vol. 69 (1968 - 1969),
pp. 33-50
Blackwell Publishing on behalf of The Aristotelian Society
Stable URL: http://www.jstor.org/stable/4544767
Presented at: Meeting of the Aristotelian Society
at 21 Bedford Square, London, W.C.1,
on Monday, 25th November, 1968, at 7.30 p.m.
EXPLAINING LOGICAL NECESSITY1
Aaron Sloman
(Now at University of Birmingham
http://www.cs.bham.ac.uk/~axs
PLAN:
I: Some facts about logical necessity stated.
II: Not all necessity is logical.
III: The need for an explanation.
IV: Formalists attempt unsuccessfully to reduce logic to syntax.
V: The no-sense theory of Wittgenstein's Tractatus merely
reformulates the problem.
VI: Crude conventionalism is circular.
VII: Extreme conventionalism is
more sophisticated.
VIII: It yields some important insights.
IX: But it ignores the variety of kinds of proof.
X: Proofs show why things must be so, but different proofs
show different things.
Hence there can be no general explanation
of necessity.
An adequate theory of meaning and truth must account for the following
facts, whose explanation is the topic, though not the aim, of the paper.
-
(i) Different signs (e.g., in different languages) may express the same
proposition.2
(ii) The syntactic and semantic rules in virtue of which sentences are
able to express contingent propositions also permit the expression of
necessary propositions and generate necessary
relations between contingent propositions. E.g., although `It snows in
Sydney or it does not snow in Sydney' can be verified empirically (since
showing one disjunct to be true would be an empirical
verification, just as a proposition of the form `p and not-p' can be
falsified empirically), nevertheless the empirical enquiry can be
short-circuited showing by what the result must be.
(iii) At least some such restrictions on truth-values, or
combinations of truth-values (e.g., when two or more contingent
propositions are logically equivalent, or inconsistent, or when one
follows from others), result from purely formal, or logical, or
topic-neutral features of the construction of the relevant propositions,
features which have nothing to do with precisely which concepts
occur, or
which objects are referred to. Hence we call some propositions logically
true, or logically false, and say some inferences are valid in virtue of
their logical form, which prevents simultaneous truth of premisses and
falsity of conclusion.
(iv) The truth-value-restricting logical forms are systematically
inter-related so that the whole infinite class of such forms can be
recursively generated from a relatively small subset, as illustrated in
axiomatisations of logic.
Subsequent discussion will show these
statements to be over-simple. Nevertheless, they will serve to draw
attention to
the range of facts whose need of explanation is the starting point
of this paper. They have deliberately been formulated to allow that
there may be cases of non-logical necessity.
I think some propositions are necessary though their truth-value is not
determined by their logical form alone. For instance,
although one can empirically test the proposition that no solid
object is completely bounded by three flat sides (e.g., by trying to
make one), one can tell in advance what the outcome will be.
To prove that no possible state of affairs can make the proposition
false, first consider the possible ways in which two planes may
be oriented to each other (parallel, or intersecting in a straight line)
and secondly, for each case consider the possible configurations
obtainable by adding a third plane: it is clear that the third
cannot close up a space. Such a proof does not make logical
deductions from explicit or implicit definitions whose logical
structure, but not content, enters into the proof essentially.
Rather, it uses essentially the fact that spatial configurations are
involved. Admittedly, since we are dealing with empirical
concepts allowing borderline cases, we should qualify our theorem
thus: There are no clear (or central) cases of a solid object
bounded by exactly three flat sides. (This sort of qualification
will be discussed below.) Similarly, there are non-logical proofs
of theorems that certain positions cannot be reached in a game of
chess played according to the rules, theorems that certain patterns
of marks cannot be generated by specified formation rules, and
other theorems whose proofs involve not mere logical forms, but
also their contents in an essential way.
For various reasons (including unwillingness to follow Kant in referring
to "intuition") some philosophers try to reduce all necessity to
logical necessity. They would argue that there are relations between the
concepts used in our geometrical theorem without which it is
impossible to identify those concepts, and from which the theorem can
be deduced using purely logical steps. Without the italicised condition
(or something similar) the claim would be trivial, for whatever theorems
we ultimately want to prove could be combined into (perhaps infinitely
many) long conjunctions: and from these our theorems could be deduced
logically. Modern mathematics is dominated by the abstracting and
generalising motive: given a set S of theorems on any topic, one can
look for a more general theory of which this topic and many others can
provide illustrations. This process of abstraction and generalisation
can always be continued until a purely logical structure, whose proofs
use only logical inference, is reached.
As noted already, this is sometimes trivial.
However, if the theorems of S are not finite in number, and especially
if they are not all logically deducible from some finite subset, then
the construction of a finitely specifiable axiom system T containing
analogues of all theorems of S, all logically deducible within T, is
not trivial.
(In fact, it is impossible when S contains all theorems
of general set theory, or arithmetic.) But even this sort of non-trivial
achievement leaves open the question whether there is a way of
identifying the concepts of S (e.g., ostensively) which neither
explicitly nor implicitly uses the axioms of T and which nevertheless
identifies concepts sufficiently definite to make the theorems of S
necessarily true.
It would then be a non-logical truth that the axioms of
T define relations which do hold between these concepts or (less
plausibly) that the axioms implicitly define concepts necessarily
co-extensive with those defined previously. Thus, the
empirical concepts used in classifying objects according to shape can be
grasped and used without even implicitly using the topic-neutral
relations between such concepts which are characterised by
the axioms of formal systems of geometry, or at any rate without using
enough of them to generate by purely logical inferences such theorems as
that no three flat sides can enclose a space.
But my main concern is
with logical necessity. The question whether all kinds of necessity are logical
has been touched on mainly in order to show that it is an open question,
and different from the question of the nature of logical necessity,
now to be discussed.
(Moreover, I think an analysis of non-logical proofs, can shed light on
the nature of logical proofs, though there will not be enough space to
develop this point here.) The facts about logical necessity referred to
in (ii)-(iv) above seem to need some sort
of explanation. How can its logical form prevent a
proposition being falsified by any state of affairs?
How does the form of an inference prevent the premisses being true and
the conclusion false, no matter how things are? It will help to clarify
these and similar questions if we start by examining some attempts
to give general answers, especially those intended to deflate the
questions by showing that there is nothing very remarkable to
explain. The motives for such deflation are related to the motives
for attempting to reduce all necessity to logic: e.g., abhorrence of
special realms of non-contingent fact and peculiar non-empirical
means of discovering such facts, and a reluctance to accept anything
remotely resembling the thin end of a wedge opening a back door for
metaphysics.
The deflationary strategies we shall examine are formalism, the no-sense
theory, and two kinds of conventionalism. My ultimate conclusion is that
no general theory explains all kinds of necessity.
Formalism is the attempt to reduce logic to syntax: the structural
relations and properties of signs.
Its basis is the important fact
that a large number of truth- or validity-guaranteeing logical forms can
be systematically derived from a subclass (cf. (iv), above), and systems
of recursively generated symbols can represent these forms.
Consequently, by purely symbolic procedures (manipulations of
meaningless symbols) we can test for logical truth or validity. The
formalist, however, says that we are testing for nothing but
properties of symbols, and that although we do not explicitly refer to
such properties (e.g., derivability in accordance with rules R from a
set of initial formulae P), nevertheless we presuppose the possibility
of doing so whenever we infer validly or state something as logically
true.3 It is argued that since logical words "do not
refer to anything in the extra-linguistic
world",4 the rules giving them their meanings must be
wholly syntactical, concerned merely with permitted combinations of
symbols. The bare bones of the thesis may be obscured by virtuoso
constructions of artificial "languages" and the definition of
syntactic analogues to such semantic concepts as `true', `denotes',
`analytic', etc.
The irony is that this reduces logic to geometry. That certain patterns
of marks are derivable from certain others by specified manipulations is
a geometrical fact whose necessity is at least as much in need of
philosophical discussion as logical necessity. There is no possibility
of some pattern ceasing to pass the test in a
strong magnetic field, or on Mars: but why not, and what right have we
to be sure? Even if we abstract from notational peculiarities
of particular symbolic systems and generalise to the study of structural
features common to different systems, we are still in the realm of
something more like geometry than logic.
For if the symbols are uninterpreted they cannot say anything true or
false and our problems about validity and truth do not arise.
But if they are interpreted (e.g., correlated by semantic conventions
with objects and their properties and relations) then our symbolic tests
are tests for something non-syntactical. So if the tests are reliable
and never, for instance, select as valid a pattern of inference which
is not truth-preserving, then since truth is not a syntactical
property of sentences this reliability needs explaining.
The explanation will, of course, refer to the connection between
formation-rules and truth-condition-rules for the use of topic-neutral
signs and constructions (i.e., syntactic and semantic rules). But it
will not reduce logic to syntax.
The no-sense theory (sometimes combined with formalism) attempts
to deflate logical necessity by saying that sentences like
`It is now raining outside or it is not now raining outside'
really say nothing, and so cannot be called `true' in the usual sense:
hence there is
no question of
explaining their necessity.
Similarly, in a logically valid
inference there is
no sense, i.e., nothing said,
in the conclusion over and
above what is
already in the
premisses:
so the impossibility of true premisses and a false conclusion is
simply the trivial impossibility of premisses which are all true yet
say something false. There aren't two distinct things related so
as to prevent one being true and the other false. (Notice that
applying this theory to logical relations like contrariety is not
quite so straightforward.) A version of this strategy is in
Wittgenstein's Tractatus (e.g., in 4.46, ff. and 6.1, ff.) along with
traces of formalism (e.g., 6.122 and most of 6.126).
In effect, this arbitrarily equates `being false in no possible states
of affairs' with `having no sense'. It does not explain
why ordinary rules
for constructing significant sentences out
of meaningful components
should sometimes lead to no sense. It rules, apparently without any justification,that our ordinary ways of telling that sentences have different sense and should,for instance,be differently translated into French,simply go wrong when it comes to (e.g.) logical
truths.
However, as
noted in (ii),
these can sometimes be verified by ordinary empirical procedures. How
could one verify something without content?
What verifies
`p or not-p' empirically need not empirically verify `q or not-q':
but how so if they both have the same sense, namely none?
Though unnecessary, this is a perfectly ordinary empirical
verification (showing one disjunct to be true verifies the whole disjunction).
The only difference is that there is a short cut: examination of the
logical form shows in advance what the outcome must be (and
in some cases it is not quite so obvious). Similarly, we can tell in
advance that whatever numbers are substituted for the variables
in `(a + b)(a - b) + b2 - a2' calculation of the result must yield the
value 0: nevertheless the
laborious calculation is still possible, and not merely in a
special sense of `calculation'. So with verification.
The no-sense theory fails to explain the fact that ordinary
empirical procedures of verification sometimes have an outcome
which depends on how things are, and can only be determined by
going through the procedure, and sometimes do not. It fails to
explain how (e.g., with logically related contingent propositions)
the outcomes of different verifications cannot vary independently,
even though each depends on how things are and may be truth or
falsity. It does not explain the difference between cases where
logical forms rule out one or other truth-value (or some combinations
of truth-values) and cases where something prevents the
verification procedure having any outcome at all, as in self-defeating
propositions (sometimes called `category mistakes')
like `Thursdays are three miles long', or `The rear axle of France
is bald'. Some who quote him with approval were not as aware
as Wittgenstein was of all these difficulties, but his only answer
seems to have been that some things cannot be asserted, though
they show themselves (e.g., Tractatus 4.122 and 6.133, f.). We
shall return to a similar suggestion later.
The formalist and no-sense theories really do little more than
restate some of the facts without offering any explanation.
Conventionalism attempts to deflate necessity by explaining it as
all arising out of arbitrary (not necessarily explicit) decisions to use
words in certain ways, or, more vaguely to adopt a certain "logical
framework" or "conceptual scheme".
Thus (in
Philosophical Logic, pp. 11-12) Strawson writes:
"Superfacts are seen to be superfluous; and the meanings of
sentences expressing necessary propositions are acknowledged . . . to be
enough to guarantee their necessity". Guarantee? How?
Although it may be a matter of convention which rules or interpretations
(or logical frameworks, etc.), govern the use of our symbols, this does
not explain how they can have consequences,
or guarantee anything. Is it
a logical truth
that if these rules or conventions or meanings are associated
with symbols then certain combinations of symbols must express true
propositions, or valid
inferences? Are Strawson's "super-facts" after all
lurking in the connexion between rules or conventions and the results
they generate? It seems that no complete,non-circular, explanation of
logical necessity (etc.)
can take the form `All cases of necessity are
consequences of so-and-so' unless `consequence' is used in a way
which does not imply necessity. Extreme conventionalism answers that what
the consequences of a
convention are is a matter of further convention: this
is the most sophisticated deflationary strategy. There are many signs of
it in Wittgenstein's later writings,5 and
perhaps also in Quine,6 though neither is completely unambiguous.
The main idea behind extreme conventionalism is that there is no way of
identifying a rule or convention independently of
explicitly specifying its applications or consequences, yet somehow
inexorably determining them in advance like rails "laid invisibly
to infinity" (Investigations I 128). One can formulate an
expression of the rule in advance of applying it, but what one intends
it to mean, how one understands it, is constituted by the way one
applies it (e.g., in drawing consequences), and therefore cannot explain
or justify the applications. If X (partly) constitutes Y, then the
existence of Y does not explain X. For instance, someone sitting beside
a conveyor belt may have in mind a certain colour or shape and pick up
objects coming along the belt if and only if they have the quality he
has in mind. It would be natural to say he knows in advance of what
actually comes along what procedure or rule he is following, and that
his following this rule (having this quality in mind) explains his
picking up these objects and leaving others. Similarly, we admit that
one can grasp a recursive rule for picking out well-formed formulae, or
proof-sequences, of a formal system, in advance of actually coming
across cases. But our conventionalist (showing behaviouristic and
nominalistic tendencies) regards these ways of speaking as misleading
insofar as they obscure the fact that the applications one makes are a
criterion for one's having this quality, procedure, or rule
in mind rather than some other: they constitute it.
This theory suggests, though it cannot consistently say,
that it is wrong to distinguish sharply between applying a predetermined
rule and deciding, or making up one's mind, exactly which rule one is
following; or between moves which are right because of a pre-existing
decision and moves which we simply decide to call `right'.
It cannot consistently say that we are wrong in our use of `He had
no choice but to pick up this one', `It follows necessarily that no
symbol like this is well-formed', for this would presuppose an ability
to identify what we intend to say independently of what we do say, thus
contradicting a basis of the theory.
To get round this difficulty, Wittgenstein tries to show that such
concepts as `necessarily', `unavoidably', `has to be', etc., have a
rôle in our lives, but not the one philosophers think. For instance,
their use can express a certain attitude to some of the decisions we
make, such as: the attitude of being unwilling in
(almost) any circumstances to revoke the decision, or being resolved to
use the decision as some kind of yardstick or standard of comparison for
other decisions. Part of what characterises the attitude is the
feeling that we have no choice, that it is based on some sort of
compelling external non-human justification (a "super-fact" ?). But
our feeling compelled can be explained by such things as: (i) our
linguistic training and human nature, which cause us to take some
decisions rather than others, (ii) our inability to imagine any sort of
(e.g., empirical) test showing the decision to be wrong, (iii) the
"deep need" that we feel for certain conventions (Remarks, p.23)
without which we could not have the concepts we do have and
communication (with ordinary humans) would break down. Compare someone's
feeling that his disgust at (e.g.) belching after meals is
justified by the "disgusting nature" of the behaviour, when in
fact he has
merely
been trained to react thus in conformity with a mere convention. His
attempts to discuss or argue with someone whose conventions are
different may simply breakdown.
Thus the conventionalist agrees that there are cases of logical
necessity, and that rules can guarantee, or determine,
certain consequences, which we can discover, for he too feels
compelled to take certain decisions and in agreeing that they are not
arbitrary he is expressing attitudes we all share. But he stresses that
there could be strange creatures who follow our usage so far then
suddenly diverge in ways which look wrong to us although they appear to
get along smoothly in their own complex social life. We could not
understand or explain such behaviour, although we can imagine it
happening. We could only draw the conclusion that their concepts
(including `rule', `apply', `meaning', `concept', `same', `correct' and
`language') were different from ours: though to say this is merely to
characterise their behaviour not to explain it.
Despite our present resolves or
attitudes to the contrary, we may ourselves one day come to behave
like them, and look back at our previous behaviour with
incomprehension. (Compare looking back now at outmoded fashions, or
at the code of honour associated with duelling.) Thus there is no
absolute non-human justification for the conclusions we actually
do draw: that we do is just a fact.
If we object that this conventionalist analysis conflicts with what we
mean by `necessarily', `follows logically', etc., the reply is that what
we mean cannot be something over and above the use we actually make of
such expressions: and he does not criticise this use, only certain
philosophical theories about what lies behind it. This manoeuvre makes
it possible for him to agree (in a certain tone of voice) with
anything one
might say in an attempt to refute him, without allowing that it
contradicts anything he is saying. Thus not only does it have the
desired effect of deflating necessity, conventionalism also
has the apparent advantage of not being directly confrontable with
any counter-argument.
We shall see, nevertheless,
that it is inadequate as an explanatory theory, though it does
provide some useful insights.
If we examine closely some cases of discovering necessary truths which
appear to be wholly different from adopting conventions, we shall find
hidden complexities. In particular, where we thought we had completely
identified some rule independently of its consequences, we might
discover a lack of determinateness which could only be removed by taking
something like an arbitrary decision to accept or reject an alleged
consequence. In this way we find that there is something right about the
conventionalist thesis: by getting clear exactly how much is right we
are in a better position to point to a residue of error. Our final
rejection of conventionalism then is based on its inability to explain
adequately the difference between those aspects of (e.g.) logical and
mathematical discovery which involve deciding to modify concepts and
those which do not. A geometrical and a logical example will help to
illustrate all this.
Suppose we have proved (cf. section II above) the theorem that no solid
is bounded by exactly three flat sides and then come across a large, or
small but very thin, steel plate which appears, to the eye, to be
bounded by two large triangular surfaces meeting in two sharp edges(like
an almost squashed paper cone) and a long thin flat surface joining the
remaining two edges. It is tempting
to say "The triangular sides cannot be perfectly flat: there must be a
gradual curve indiscernible to the eye": but how can we be sure, without
adopting conformity to the theorem as a new criterion for the
instantiation of the concepts involved? And how can we be sure that no
other unsuspected actual or potential counter-example to the theorem
will turn up, unless we choose not to let it?
The history of mathematics shows that however compelling a proof may
look it is rash to assume that no counter-example will ever turn up,
unless the concepts involved are redefined as suggested above. But it
does not follow from this that all a proof does is somehow lead us to
such a redefinition: for even if the complex interrelations of our
concepts make it difficult for us to survey all their possible
applications nevertheless the proof may show quite clearly that there is
a range of cases (e.g., solids bounded by flat sides no two of
which are almost superimposed) within which counter-examples are
impossible. Although our concepts may, to start with, have unsuspected
areas of indeterminateness in which putative counter-examples may turn
up, and although within the determinate area there may be configurations
not taken into account in the proof, nevertheless in central cases, of a
type considered in the proof, the concepts, as identified prior to
giving the proof, may be sufficiently determinate to leave no room for
further conventions to govern their application. There is much more to
be said about this, but first let us look at the logical example,the
theorem that all propositions of the form not-(p and not-p) must be
true.
The normal proof starts by assuming that `not' and `and' are
truth-functional connectives defined by the usual truth-tables, and
showing how these automatically guarantee that the truth-table for the
complex proposition not-(p and not-p) contains only T's in its final
column. A conventionalist might comment that there is nothing which
guarantees that every proposition we can express must have a
definite truth-value, and a unique truth-value, both of which are
assumed in the truth-table proof. Examples using borderline cases of
indeterminate concepts, or "category mistakes", or unsuccessful
reference, easily come to mind. Further, cases like `I am saying
something false' , and `The set of
all non-self-containing sets contains itself' might be taken to show
that normal procedures for assignment of truth-value sometimes
assign both truth and falsity to a proposition. The only thing
which can guarantee that these and perhaps other unsuspected
types of case will not refute the theorem is a new decision to
accept the theorem as giving a criterion for application of such
concepts as `true', `false' and `proposition', so that nothing is
described as a counter-example. Thus putative counter-examples
can be dealt with by not calling them `propositions' or by
somehow modifying procedures of truth-value assignment. The
need for some such decision or convention is especially clear if it is
noticed that there are always possibilities of extension of a
language, e.g., by introducing new concepts, or new types of use of
old concepts and linguistic activities (such as embedding assertions
in the context of a new type of ritual); for the only way to be sure
that no such extension will generate a new counter-example is to
decide not to permit any extension unless it preserves the truth
of the theorem.
However, all this again ignores the fact that there is a range of
central cases where p has a determinate truth-value: here there is
no possibility of exception, and the truth-table proof shows why.
Here the central range of cases for which there is no need to
adopt a new convention is clearer than in the geometrical example.
We thus have two important conventionalist insights: (a) the
only way to guarantee that no unsuspected counter-example can
turn up is to adopt the truth of the theorem as giving a new
criterion for applying the concepts used therein, and (b) there
may be great difficulties in specifying precisely the range of
central cases for which the theorem and proof hold in their
original interpretation, unless it is identified as the
range for which
the theorem holds: yet elaborate proofs hardly seem required for
showing that theorems hold where they hold. Despite the importance of
these points, however, the conventionalist fails to account
for our having a clear view of at least some cases which can be
non-circularly characterised and for which the proof demonstrates that
they conform to
the theorem.
The conventionalist oversimplifies: he asks too sweeping a
question (what
can guarantee that nothing will ever turn up to disprove the
theorem?) and gives too sweeping an answer (only a convention to count
nothing as a counter-example). But why should there be infallible
ways of making logical or mathematical discoveries? Why should we always
adopt new conventions that what we appear to have proved is to be called
`true' come what may? (The strains involved in following such a policy
through would probably be intolerable.) Why not accept that we can find,
through further investigations, that we have made mistakes: that our
definitions may lead to borderline cases, or generate inconsistencies;
that our proofs concerning their consequences may fail to take account
of all cases, or fail to distinguish cases for which different sorts of
proof are needed,etc.; that precisely what has been proved is not
accurately stated in the original theorem; or that we have not provided
a non-circular way of identifying the range of cases for which the
original theorem is true? But rarely does a proof turn out to have
proved nothing at all. Moreover, we can give new proofs that our
old proofs or theorems were mistaken, and new proofs of the old theorems
or new formulations of what the old proofs proved. We can construct new
concepts related to the old ones and prove new theorems using them. In
all this we can again make mistakes, and discover them in subsequent
investigations.7
For instance, returning to our logical theorem proved by
truth-table, we may examine the cases not covered by the proof
and as a result try to give a more general definition of
`proposition' which assumes only that every proposition identifies a
set of
truth-conditions and a set of falsity-conditions, allowing that
the method of identification may in some cases result in overlapping
sets, or non-exhaustive sets (i.e., the proposition lacks a
truth-value in some states of affairs). We can then explore different
possible ways of redefining the truth-functional connectives and the
consequences of the
new definitions. Thus `not' could be defined as simply
interchanging T-conditions
and F-conditions, or as producing a new set of
T-conditions containing all possible states of affairs which are not
T-conditions for the original proposition. We could take the
F-conditions of `p and q' as the union of the two original sets of
F-conditions or as including only conditions where
both p and q have definite truth-values. Alternatively, we can take
a narrower definition of
`proposition', ruling out the cases where T-conditions and
F-conditions overlap. General theorems about the consequences
of these various moves could then be proved with the aid of
modified truth-tables. Or we might prove our old theorem for
some restricted class of propositions (e.g., mathematical
propositions) - which would involve proving that for this class the
methods of identifying sets of T-conditions and F-conditions
guarantee that each pair of sets is exhaustive and mutually
exclusive: for this case the truth-table would merely be the last
step in a more complex proof. Thus, what is normally treated as
a simple matter in logic books has hidden complexities. (With
quantifiers the situation is even more complicated.) Similar
explorations, with modified definitions, new sub-ranges, revised
formulations of the theorem, etc., are possible with our geometrical
example. For instance we could explore the possibility of
eliminating the putative counter-example (which must have one
long thin side bounded by two edges) by explicitly defining `flat'
so that two flat surfaces cannot intersect in a curved line and using
a theorem that if two lines meet in two points then at least one
is curved. Or we may look for a more or less approximate
characterisation of the range of cases for which the original proof
does work, without requiring any modifications of the concepts
involved.
In short, although we can adopt new concept-fixing conventions
which make all our theorems true by definition (and thus
insignificant?), we need not: there is a rich realm of
possibilities, not accounted for by
conventionalism, including the possibility of discovering some
non-contingent fact about our concepts as
already identified, such as the fact that some supposed theorem
is false. Thus, Formalism, the no-sense theory, and
conventionalism each fails to explain or explain away all cases of
logical (and non-logical) necessity, though each contains some
important insights.
Then how are we to explain the facts noted in section I, above? Take as
illustration the theorem that if p is a proposition with a definite
truth-value and `not' and `and' are defined by the normal truth-tables
then the proposition not-(p and not-p) must be
true. (This formulation is immune to the objections raised
previously.) If someone asks why, i.e., wants to know the explanation,
what can we do but give him a proof? For instance, we show
how, on the assumption that p has a definite truth-value, T or F,
we can use the definitions of `not' and `and' to construct a
truth-table for the complex proposition in which the final column
contains nothing but `T'. Going through the construction shows
why the rules guarantee this consequence. There is nothing in the
construction which allows the time and place or circumstances in
which it is carried out to make any difference to the outcome.
Hence one can see (though this word requires further discussion)
that no matter when or where the process is repeated, the result
cannot vary in any essential detail.
This shows why even if p is empirical, and has a
truth-value discoverable empirically, the
empirical investigation is not necessary in order to show that
(p and not-p) is not false. Moreover, by abstracting from the
precise features of the notation used, one
can see why the result does not depend on which language is used to
express the proposition.
Although this example raises more problems than there is time to discuss
here, it will suffice to illustrate my thesis that the way to explain a
case of necessity is to give a proof of that case. A proof shows why a
certain logical form guarantees non-falsity, or why no three flat sides
can enclose a solid (at least in central
cases), or why a certain position cannot be reached in chess played
according to the rules.
The explanation, that is, what the proof shows, is different in
different cases. What the proof does, and how, varies enormously.
For instance, a proof that a certain kind of thing is possible (e.g., a
solid bounded by four flat sides, or a set of ten
propositions no two of which can be false simultaneously) works
differently from a proof that something is impossible (e.g., a false
proposition of the form `not-(p and not-p)'). Proving something
about a specific type (`No three planes meeting at right angles can
enclose a space') may be different from proving something
more general ('No three planes can enclose a space'). Some
proofs use "construction lines" ('Suppose we alter the situation
thus....') while others do not. A proof using concepts implicitly
or explicitly defined in terms of others works differently from a
proof using only ostensively defined concepts. Some proofs use
only empirical concepts abstracted from actual instances, while
others use idealised concepts identified as limiting cases of some
sequence. (E.g., the concept of a perfectly straight line, or
perfectly flat surface, or the concept of a contingent proposition
whose
truth-value is definite in all possible circumstances, or the concept
of an indefinitely continuing sequence of tosses of a perfectly
balanced coin, or the concept of a perfect democracy. A proof
using such concepts must show, or assume, that the
extrapolation to the limiting case does identify a concept.
Belief in the absolute correctness of Euclidean geometry, or the
applicability of only classical logic in the theory of infinite sets may
arise from misplaced confidence in such assumptions.) Some proofs draw
attention to something obvious (but perhaps unnoticed), while
others reveal hidden connexions.
What is now needed is a detailed and systematic survey of this variety
of cases. It would surely undermine such common preconceptions as that
every proof proceeds by logical steps, that
every proof must be (in principle) checkable by some mechanical
procedure, that the purpose of a proof is to provide us with
absolute certainty, and that there are some simple, general and
basic necessary truths which underlie the rest. Such a survey
would put us in a better position to analyse the notion of a proof
`showing why such-and-such must be the case'. For the present, we need
merely note that since different proofs show different things it is
folly to expect one general theory to answer all questions of the form:
Why must so-and-so be the case?
Footnotes:
1An early version of this was
read to seminars at
Hull and Leeds universities
in in 1963. It was resurrected at short notice
and revised, largely under the stimulus of an unpublished paper by Cora
Diamond on Wittgenstein's views on necessity.
2The best way to identify propositions, for our
purposes, is not in terms of truth-conditions but in terms of a way of
identifying a set of truth-conditions and a set of falsity-conditions.
Thus `My lawn has three straight edges' and `My lawn has straight edges
meeting in three corners' both identify the same truth- and
falsity-conditions, but in different ways. They express different
propositions. Further clarification of this criterion cannot be
undertaken here. It is noteworthy that most of the facts about logical
necessity discussed herein do not presuppose any asymmetry between truth
and falsity. As long as recursive (semantic and syntactic) rules
associate with each sentence an empirical procedure for assigning that
sentence to one of two categories, all our problems can arise. Compare
my
"Functions and Rogators"
in J. N. Crossley and M. A. E. Dummett ed.
in Formal Systems and Recursive Functions, (North-Holland1965).
3E.g., see R. Carnap, Foundations of Logic and
Mathematics, p. 37.
4A. M. Quinton,
"The a priori and the analytic", Proc. Arist. Soc., Vol. 64
(1963-4), pp. 31-54 (reprinted in Philosophical Logic, ed. P. F.
Strawson, at p. 123).
5E.g.,
Philosophical Investigations, part I, $$ 186, 241-2, 292, 517,
and more explicitly in
Remarks on
the Foundations Mathematics, of e.g., p. 12-13, p. 23, p. 121ff,
though in
later passages,e.g.,
p. 193-6, he retreats to something
much more like the Tractatus doctrine
of "showing". He seems to have been strongly
pulled in several directions.
6E.g., "Two dogmas of empiricism" in
Philosophical Review, 1951, and "Necessary truth" in
The Ways of Paradox.
7I am indebted to I. Lakatos: "Proofs and
Refutations" (four parts) in British Journal of the Philosophy of
Science, 1963.
File translated from
TEX
by
TTH,
version 3.80.
On 26 Dec 2009, 10:51.