[mod.ai] non-monotonic reasoning

jbn@GLACIER.STANFORD.EDU (John B. Nagle) (11/02/86)

      Proper mathematical logic is very "brittle", in that two axioms
that contradict each other make it possible to prove TRUE=FALSE, from
which one can then prove anything.  Thus, AI systems that use
traditional logic should contain mechanisms to prevent the introduction
of new axioms that contradict ones already present; this is referred
to as "truth maintenance".  Systems that lack such mechanisms are prone
to serious errors, even when reasoning about things which are not
even vaguely related to the contradictory axioms; one contradiction
in the axioms generally destroys the system's ability to get useful
results.
       Non-monotonic reasoning is an attempt to make reasoning systems
less brittle, by containing the damage that can be caused by
contradiction in the axioms.  The rules of inference of non-monotonic
reasoning systems are weaker than those of traditional logic.  There
is not full agreement on what the rules of inference should be in
such systems.  There are those who regard non-monotonic reasoning as
hacking at the mathematical logic level.  Non-monotonic reasoning
lies in a grey area between the worlds of logic and heuristics.

				John Nagle

ether.allegra@btl.CSNET.UUCP (11/06/86)

John Nagle, in a recent posting, writes:
>         Non-monotonic reasoning is an attempt to make reasoning systems
>  less brittle, by containing the damage that can be caused by
>  contradiction in the axioms.  The rules of inference of non-monotonic
>  reasoning systems are weaker than those of traditional logic.  

Most nonmonotonic reasoning formalisms I know of (default logic,
autoepistemic logic, circumscription, NML I and II, ...) incorporate
a first-order logic as a subset.  Their rules of inference are thus
*stronger* than traditional logics'.  I think Nagle is thinking of
Relevance Logic (see Anderson & Belnap), which does make an effort
to contain the effects of contradiction by weakening the inference
rules (avoiding the paradoxes of implication).

As for truth-maintenance systems, contrary to Nagle and popular
mythology, these systems typically do *not* avoid contradictions
per se.  What they *do* do is prevent one from 'believing' all
of a set of facts explicitly marked as contradictory by the
system using the TMS.  These systems don't usually have any
deductive power at all, they are merely constraint satisfaction
devices.


David W. Etherington
AT&T Bell Laboratories
600 Mountain Avenue 
Murray Hill, NJ 07974-2070
ether%allegra@btl.csnet