Download Nonmonotonic Reasoning

Document related concepts

Neural modeling fields wikipedia , lookup

AI winter wikipedia , lookup

Existential risk from artificial general intelligence wikipedia , lookup

Fuzzy logic wikipedia , lookup

Philosophy of artificial intelligence wikipedia , lookup

Logic programming wikipedia , lookup

Knowledge representation and reasoning wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Transcript
Nonmonotonic Reasoning
Alexander Bochman
Computer Science Department,
Holon Institute of Technology (H.I.T.)
November 24, 2011
1
What is nonmonotonic reasoning
The field of nonmonotonic reasoning is now an essential part of the logical approach to Artificial Intelligence (AI). There exists a vast literature on the topic,
including a number of books [Antoniou, 1997; Besnard, 1989; Bochman, 2001;
2005; Brewka, 1991; Lukaszewicz, 1990; Makinson, 2005; Marek and Truszczyński,
1993; Schlechta, 1997; 2004]. Two collections are especially useful: [Ginsberg, 1987] is a primary source for the early history of the subject, while the
handbook [Gabbay et al., 1994] provides overviews of important topics and
approaches. [Minker, 2000] is a most recent collection of survey papers and
original contributions to logic-based AI. This chapter has also benefited from
a number of overviews of the field, especially [Reiter, 1987a; Minker, 1993;
Brewka et al., 1997; Thomason, 2003].
The relationship between nonmonotonic reasoning and logic is part of a
larger story of the relations between AI and logic (see [Thomason, 2003]). John
McCarthy, one of the founders of AI, has suggested in [McCarthy, 1959] and
consistently developed a research methodology that used logic to formalize the
reasoning problems in AI1 . McCarthy’s objective was to formalize common sense
reasoning used in dealing with everyday problems.
In a sense, nonmonotonic reasoning is an outgrowth of McCarthy’s program.
But though commonsense reasoning has always appeared to be an attractive
standard, the study of ‘artificial reasoning’ need not and actually has not been
committed to the latter. The basic formalisms of nonmonotonic reasoning could
hardly be called formalizations of commonsense reasoning. Still, in trying to
cope with principal commonsense reasoning tasks, the suggested formalisms
have succeeded in capturing important features of the latter and thereby have
broken new territory for logic.
Artificial Intelligence has practical purposes, which give rise to problems
and solutions of a new kind, apparently different from the questions relevant for
philosophers. The authors of first nonmonotonic theories have tried, of course, to
express their formalisms using available logical means, ranging from the classical
1 See
[Lifschitz, 1991a] for an overview of McCarthy’s research program.
1
first order language to modal logics. McCarthy himself has always believed
that anything that can be expressed, can be expressed in first order logic (he
considered this a kind of Turing thesis for logic). Still, from its very beginning,
logical AI has created formalisms and approaches that had no counterpart in
existing logical theories, and refined them to sophisticated logical systems. This
was achieved mostly by people that were not logicians as their primary speciality.
It is even advantageous to see nonmonotonic reasoning as a brand new approach
to logical reasoning; this would save us from hasty attempts to subsume such a
reasoning in existing logical formalisms at the price of losing the precious new
content. Already at this stage of its development, nonmonotonic reasoning is
not yet another application of logic, but a relatively independent field of logical
research that has a great potential in informing, in turn, future logical theory
as well as many areas of philosophical inquiry.
The origins of nonmonotonic reasoning within the broad area of logical AI
lied in dissatisfaction with the traditional logical methods in representing and
handling the problems posed by AI. Basically, the problem was that reasoning
necessary for an intelligent behavior and decision making in realistic situations
has turned out to be difficult, even impossible, to represent as deductive inferences in some logical system.
In commonsense reasoning, we usually have just partial information about
a given situation, and we make a lot of assumptions about how things normally
are in order to carry out further reasoning. For example, if we learn that Tweety
is a bird, we usually assume that it can fly. Without such presumptions, it would
be almost impossible to carry out the simplest commonsense reasoning tasks.
Speaking generally, human reasoning is not reducible to collecting facts and deriving their consequences; it embodies an active epistemic attitude that involves
making assumptions and wholesale theories about the world and acting in accordance with them. We do not only perceive the world, we also give it structure
in order to make it intelligible and controllable. Commonsense reasoning in this
sense is just a rudimentary form of a general scientific methodology.
The way of thinking in partially known circumstances suggested by nonmonotonic reasoning consists in using justified beliefs and reasonable assumptions that can guide us in our decisions. Accordingly, nonmonotonic reasoning
can be described as a theory of making and revising assumptions in a reasoned
or principled way ([Doyle, 1994]). Of course, the latter are only beliefs and
assumptions, so they should be abandoned when we learn new facts about the
circumstances that contradict them.
The sentence “Birds (normally) fly” is weaker than “All birds fly”; there is
a seemingly open-ended list of exceptions – ostriches, penguins, Peking ducks,
etc. etc. So, if we would try to use classical logic for representing “Birds fly”,
the first problem would be that it is practically impossible to enumerate all
exceptions to flight with an axiom of the form
(∀x).Bird(x)&¬P enguin(x)&¬Emu(x)&¬Dead(x)&... ⊃ F ly(x)
This fact indicates that our commonsense assumptions are often global in
character, saying something like “the world is as normal as possible, given the
2
known facts”. The second crucial problem is that, even if we could enumerate
all such exceptions, we still could not derive F ly(T weety) from Bird(T weety)
alone. This is so since we are not given that Tweety is not a penguin, or dead,
etc. The antecedent of the above implication cannot be derived, in which case
there is no way of deriving the consequent. Nevertheless, if told only about a
particular bird, say Tweety, without being told anything else about it, we would
be justified in assuming that Tweety can fly, without knowing that it is not one
of the exceptional birds. So the problem is how we can actually make such
assumptions in the absence of information to the contrary.
This suppositional character of commonsense reasoning conflicts with the
monotonic character of logical derivations. Monotonicity is just a characteristic
property of deductive inferences arising from the very notion of a proof being
a sequence of steps starting with accepted axioms and proceeding by inference
rules that remain valid in any context of its use. Consequently, if a set a of
formulas implies a consequence C, then a larger set a ∪ {A} will also imply C.
Commonsense reasoning is non-monotonic in this sense, because adding new
facts may invalidate some of the assumptions made earlier.
In his influential “frames paper” [Minsky, 1974], Marvin Minsky proposed
the notion of a frame, a complex data structure meant to represent a stereotyped and default information. While Minsky’s description of a frame was
informal, central to his notion were prototypes, default assumptions, and the
unsuitability of classical definitions for commonsense concepts. In the appendix
entitled “Criticism of the Logistic Approach”, Minsky explained why he thinks
that logical approaches will not work. To begin with, he directly questioned
the suitability of representing commonsense knowledge in a form of a deductive
system:
There have been serious attempts, from as far back as Aristotle, to
represent common sense reasoning by a ”logistic” system .... No
one has been able successfully to confront such a system with a
realistically large set of propositions. I think such attempts will
continue to fail, because of the character of logistic in general rather
than from defects of particular formalisms.
Minsky doubted the feasibility of representing ordinary knowledge effectively
in the form of many small, independently true propositions. On his opinion, such
a “logical” reasoning is not flexible enough to serve as a basis for thinking.
The strategy of complete separation of specific knowledge from general rules of inference is much too radical. We need more direct ways
for linking fragments of knowledge to advice about how they are to
be used.
As a result of this deficiency, traditional formal logic cannot discuss what
ought to be deduced under ordinary circumstances. Minsky was also one of the
first who mentioned monotonicity as a source of the problem:
MONOTONICITY: ... In any logistic system, all the axioms are
necessarily “permissive” - they all help to permit new inferences
3
to be drawn. Each added axiom means more theorems, none can
disappear. There simply is no direct way to add information to
tell such the system about kinds of conclusions that should not be
drawn! To put it simply: if we adopt enough axioms to deduce what
we need, we deduce far too many other things.
As yet another problematic feature, Minsky mentioned the requirement of
consistency demanded by Logic that makes the corresponding systems too weak:
I cannot state strongly enough my conviction that the preoccupation with Consistency, so valuable for Mathematical Logic, has been
incredibly destructive to those working on models of mind. At the
popular level it has produced a weird conception of the potential capabilities of machines in general. At the “logical” level it has blocked
efforts to represent ordinary knowledge, by presenting an unreachable image of a corpus of context-free “truths” that can stand separately by themselves.
First theories of nonmonotonic reasoning could be viewed as providing logical
answers to Minsky’s challenge.
2
Pre-history: problems and first solutions
Long before the emergence of first nonmonotonic systems, there have been a
number of problems and applications in AI that required and used some forms of
nonmonotonic reasoning. In fact, it is these problems and solutions, rather than
the strategic considerations of McCarthy or Minsky, that influenced the actual
shape of subsequent nonmonotonic formalisms. Initial solutions to commonsense
reasoning tasks worked (though in restricted applications), and this was an
incentive for trying to provide them with a more systematic logical basis.
On a most general level, nonmonotonic or default reasoning is intimately
connected to the notion of prototypes in psychology and natural kinds in philosophy. Just as the latter, default assumptions cannot be defined via necessary
and sufficient conditions, but involve a description of “typical” members. The
problem of representing and reasoning with such concepts has reappeared in
AI as a practical problem of building taxonomic hierarchies for large knowledge
bases. The basic reasoning principle in such hierarchies is that subclasses inherit
properties from their super-classes.
Much more complex logical issues have arisen when the organization of
a domain into hierarchies has been allowed to have exceptions. The theory
of reasoning in such taxonomies has been called nonmonotonic inheritance
(see [Horty, 1994] for an overview). The guiding principle in resolving potential conflicts in such hierarchies was a specificity principle ([Poole, 1985;
Touretzky, 1986]): more specific information should override more generic information in cases of conflict. Though obviously related to nonmonotonic reasoning, nonmonotonic inheritance relied more heavily on graph-based represen-
4
tations than on traditional logical tools (see, e.g., [Touretzky, 1986]). Nevertheless, it has managed to provide a plausible analysis of reasoning in this restricted
context. The relations of nonmonotonic inheritance to general nonmonotonic
formalisms, and especially the role of the specificity principle have been an active
area of research.
Default assumptions of a different kind have been ‘discovered’ in the framework of already existing systems, such as reasoning in databases and planning
domains, and even in formulation of ordinary problems, or puzzles. A common
assumption or, better, convention in such systems and problem formulations has
been that positive assertions that are not explicitly stated should be considered
false. Attempts to formalize this assumption have led to Reiter’s Closed World
Assumption principle and McCarthy’s circumscription (see below).
But first and foremost, the problem of default assumptions has shown itself
in attempts to represent reasoning about actions and change.
2.1
The frame problem
This, then, are the three problems in formalizing
action: [the qualification, frame, and ramification
problem]. Other than that, no worries. [Ginsberg, 1993]
It is difficult to overestimate the importance of the frame problem for AI. It
is central to virtually every interesting area of AI, such as planning, explanation,
and diagnosis. As such, solving the frame problem is necessary for the whole
logical approach to AI.
As was stated in [McCarthy and Hayes, 1969], a computer program capable
of acting intelligently in the world must have a general representation of the
world in terms of which its inputs are interpreted. It should decide what to do
by inferring in a formal language that a certain strategy will achieve its assigned
goal. This paper introduced also the main problem that prevented an adequate
formalization of this task - the Frame Problem. Basically, the problem was how
efficiently determine which things remain the same in a changing world.
As was rightly noted in [Thomason, 2003], the frame problem arises in the
context of predictive reasoning, a type of reasoning that has been neglected in the
traditional tense-logical literature, but is essential for planning and formalizing
intelligent behavior. Prediction involves the inference of later states from earlier
ones. Changes in this setting do not merely occur, but occur for a reason.
Furthermore, we usually assume that most things will be unchanged by the
performance of an action. It is this inertia assumption that connects reasoning
about action and change with nonmonotonic reasoning. In this reformulation,
the frame problem is to determine what stays the same about the world as time
passes and actions are performed – without having to explicitly state each time
all the things that stay the same ([Morgenstern, 1996]).
The frame problem has a number of formulations and levels of generalization.
The initial description of the problem in [McCarthy and Hayes, 1969] was made
in the specific context of a situation calculus, an instance of first-order logic
5
especially formulated for reasoning about action. In this model, changes are
produced by actions, so the basic relation is Result between an action, an initial
situation, and a situation resulting from the performance of the action. In order
to specify how propositions (fluents) do not change as actions occur (e.g., a red
block remains red after we have put it on top of another block), the authors
suggested to write down special axioms they called frame axioms. Of course, a
huge number of things stay the same after a particular action, so we would have
to add a very large number of frame axioms to the theory. This is precisely the
frame problem. More generally, it is the persistence problem [Shoham, 1988]:
the general problem of predicting the properties that remain the same as actions
are performed, within any reasonable formalism for reasoning about time.
On a more general level, the frame problem has been understood as a temporal projection problem encompassing the persistence, ramification and qualification problems, as well as the problem of backward temporal projection, or
retrodiction.
The ramification problem was formulated first in [Finger, 1987] and concerns
the necessity of taking into account numerous derived effects (ramifications) of
actions, effects created by logical and causal properties of a situation. Taking
an example from [Lin, 1995], suppose that a certain suitcase has two locks, and
is open if both locks are open. Then the action of opening one lock produces
an indirect effect of opening the suitcase if and only if the other lock is open.
Derived effects should be taken into account when combined with the above
mentioned inertia assumption, since they override the latter. Thus, if an action
of opening one lock is performed, its derived effect overrides the default inertia
assumption that the suitcase remains closed.
The ramification problem has raised general questions on the nature of causation and its role in temporal reasoning. It was an incentive for the causal
approach to the frame problem, described later.
The qualification problem is the problem of specifying what conditions must
be true in the world for a given action to have its intended effect. It was
introduced in [McCarthy, 1980] and described as follows:
It seemed that in order to fully represent the conditions for the
successful performance of an action, an impractical and implausible
number of qualifications would have to be included in the sentences
expressing them. Common sense reasoning is ordinarily ready to
jump to the conclusion that a tool can be used for its intended purpose unless something prevents its use. Considered purely extensionally, such a statement conveys no information; it seems merely
to assert that a tool can be used for its intended purpose unless it
can’t. Heuristically, the statement is not just a tautologous disjunction; it suggests forming a plan to use the tool.
If I turn the ignition key in my car, I expect the car to start. However, many
conditions have to be true in order for this statement to be true: The battery
must be alive, the starter must work, there must be gas in the tank, there is no
potato in the tailpipe, etc. - an open-ended list of qualifications. Still, without
6
knowing for certain about most of these facts, I normally assume that turning
the key will start the car.
The Qualification Problem has turned out to be one of the most stubborn
problems for the representation of action and change. The majority of subsequent nonmonotonic formalisms, such as default logic, have failed to deliver the
intended conclusions due to the fact that they do not provide natural ways of
formalizing the specificity principle, mentioned earlier in discussing nonmonotonic inheritance, according to which more specific defaults should override more
general defaults.
The idea behind first nonmonotonic solutions to the frame problem has been
to treat inertia as a default: changes are assumed to occur only if there is some
reason for them to occur. In action-based formalisms, the absence of change is
inferred when an action is performed unless a reason for the change can be found
in axioms for the action. Attempts to formalize such a reasoning obviously required a nonmonotonic reasoning system. In one of the earliest attempts to formalize such a reasoning, [Sandewall, 1972] used a modal operator U N LESS A
meaning “A cannot be proved” for dealing with the frame problem, namely
for expressing the inertia claim that every action leaves any fluent unaffected,
unless it is possible to deduce otherwise. One of the principal motivations of
McCarthy and Reiter for the study of nonmonotonic reasoning was the belief
that it would provide a solution to the frame problem.
The frame problem has ‘survived’, however, the first generation of nonmonotonic formalisms. [Hanks and McDermott, 1987] have suggested the Yale Shooting Anomaly and demonstrated that the apparently plausible nonmonotonic approaches to the frame problem fail. Actually, it was a major factor in conversion
of McDermott, one of the founders of nonmonotonic reasoning, to an antilogicist
(see [McDermott, 1987]).
The Yale Shooting Anomaly has shown, in effect, that a simple-minded
combination of default assumptions measured only by the number of defaults
that are violated can lead to spurious solutions. Still, the development of the
nonmonotonic approach to the frame problem has continued and led to new,
more adequate solutions (see, e.g., [Morgenstern and Stein, 1994; Sandewall,
1994; Shanahan, 1997]). In addition, monotonic solutions to the frame problem
has been suggested in [Schubert, 1990; Reiter, 1991] and successfully applied to
quite complex formalization problems. These solutions were based, however, on
writing explicit frame axioms stating what does not change when an action is
performed.
2.2
Procedural nonmonotonicity
Procedural solutions to the frame problem have been popular in AI since its earliest days. Perhaps the best known is the planning program STRIPS suggested
by Fikes and Nilsson in 1971. Given an initial state, a goal state, and a list of
actions, STRIPS finds a sequence of actions that achieves the goal state. In the
course of planning, however, it must reason about what changes and what stays
the same after an action is performed. STRIPS avoids (and thereby solves)
7
the frame problem by assuming that if an action is not known to change some
feature, it does not. In a setting devoid of deductive inferences, this principle
is easy to represent procedurally by associating with each action a list of preconditions that must be satisfied in order for the action to be performed, along
with an add list and a delete list. The add list is the set of statements that gets
added to the current state after an action is performed, and the delete list is
the set of statements that gets deleted from the current state after the action is
performed. By its very nature, STRIPS cannot handle conditional actions, and
worked only for limited ontologies devoid of causation and other constraints on
combination of propositional atoms.
Also in other areas of AI researchers have routinely been implementing procedural nonmonotonic reasoning systems, usually without reflecting on the underlying reasoning patterns on which their programs rely. Typically these patterns
were implemented using the so-called negation-as-failure, which occurs as an
explicit operator in programming languages like PROLOG.
In the database theory there is an explicit convention about the representation of negative information that provided a specific instance of nonmonotonic reasoning. For example, the database for an airline flight schedule does
not include the city pairs that are not connected, which clearly would be an
overwhelming amount of information. Instead of explicitly representing such
negative information, databases implicitly do so by appealing to the so-called
closed word assumption (CWA) [Reiter, 1978], which states that if a positive
fact is not explicitly present in the database, its negation is assumed to hold.
For simple databases consisting of atomic facts only, e.g. relational databases,
this approach to negative information is straightforward. In the case of deductive databases, however, it is no longer sufficient that a fact not be explicitly
present in order to conjecture its negation; the fact may be derivable. For this
case, Reiter defined the closure of a database as follows:
CW A(DB) = DB ∪ {¬P (t) | DB 2 P (t)}
where P(t) is a ground predicate instance. That is, if a ground atom cannot
be inferred from the database, its negation is added to the closure. Under the
CWA, queries are evaluated with respect to CWA(DB), rather than DB itself.
Logically speaking, the CWA singles out the least model of a database.
Consequently, it works only when the database possesses such a least model,
e.g., for Horn databases. Otherwise it becomes inconsistent. For example, for
a database containing just the disjunctive statement A ∨ B, neither A nor B
is deduced, so both ¬A and ¬B are in the closure, which is then inconsistent
with the original database. A suitable generalization of CWA for arbitrary
databases, the Generalized Closed World Assumption, has been suggested in
[Minker, 1982].
The Prolog programming language developed by Colmerauer and his students [Colmerauer et al., 1973] and the PLANNER language developed by Hewitt ([Hewitt, 1969]) were the first languages to have a nonmonotonic component. The not operator in Prolog, and the THNOT capability in PLANNER
8
provided default rules for answering questions about data where the facts did
not appear explicitly in the program.
In Prolog, the goal not G succeeds if the attempt to find a proof of G using
the Prolog program as axioms fails. Thus, Prologs negation is a nonmonotonic
operator: if G is nonprovable from some axioms, it needn’t remain nonprovable
from an enlarged axiom set. The way this procedural negation is actually used
in AI programs amounts to invoking the rule of inference “From failure of G,
infer ¬G.” This is really the closed world assumption. This procedural negation
can also be used to implement other forms of default reasoning; this has led to
developing a modern logic programming as a general representation formalism
for nonmonotonic reasoning.
A different formalization of the CWA was proposed in [Clark, 1978] in an
attempt to give a formal semantics for negation in Prolog. Clark’s idea was
that Prolog clauses provide sufficient but not necessary conditions on the predicates in their heads, while the CWA is the assumption that these sufficient
conditions are also necessary. Accordingly, for a propositional logic program Π
consisting of rules of the form p ← a, queries should be evaluated with respect
to its completion, which is a classical logical theory consisting of the following
equivalences:
_
p ↔ {∧ai | p ← ai ∈ Π},
for any ground atom p. The completion formulas embody two kinds of information. As implications from right to left, they contain the material implications
corresponding to the program rules. In addition, left-to-right implications state
that an atom belongs to the model only if one of its justifications is also in the
model. The latter justification requirement has become a central part of the
truth maintenance system, suggested by Jon Doyle.
2.3
Justification-based truth maintenance
Doyle’s truth maintenance system (TMS) can be viewed as one of the first
rigorous solutions to the problem of representing nonmonotonic reasoning (see
[Doyle, 1979]). In particular, it introduced the now familiar notion of nonmonotonic justification, subsequently used in default and modal nonmonotonic
logics.
The idea of the TMS was to keep track of the support of beliefs, and to use
the record of these support dependencies when it is necessary to revise beliefs.
In a TMS, part of the support for a belief can consist in the absence of other
beliefs. This introduced nonmonotonicity.
The TMS represented belief states by structures called nodes that the TMS
labeled as either in or out (of the current state). The TMS also recorded sets
of justifications or reasons for each node in the form of rules A//B c read
as “A without B gives c”, meaning that the node c should be in if each node
in the set A is in and each node in the set B is out. The TMS then seek to
construct labelings for the nodes from these justifications, labelings that satisfy
two principles:
9
• stability - a node is labeled in iff one of its reasons is valid in the labeling
(i.e., expresses hypotheses “A without B” that match the labeling);
• groundedness - labelings provide each node labeled in with a noncircular
argument in terms of valid reasons.
The TMS algorithm and its refinements had a significant impact on AI
applications, and called for a logical analysis. It provided a natural and highly
specific challenge for those seeking to develop a nonmonotonic logic. In fact,
both the modal nonmonotonic logic of [McDermott and Doyle, 1980] and the
default logic of [Reiter, 1980] can be seen as logical formalizations of the above
principles. Namely, each of these theories formalized nonmonotonic reasoning
by encoding groundedness and the presence and absence of knowledge in terms
of logical provability and unprovability (consistency).
It is interesting to mention, however, that Jon Doyle himself has always felt
a discrepancy between his original formulation and subsequent logical formalizations:
In the first place, the logical formalizations convert what in many
systems is a fast and computationally trivial check for presence and
absence of attitudes into a computationally difficult or impossible
check for provability, unprovability, consistency or inconsistency.
This inaptness seems especially galling in light of the initial problemsolving motivations for nonmonotonic assumptions, for which assumptions served to speed inference, not to slow it. ([Doyle, 1994])
One of the most important features of Doyle’s system was its emphasis on
justifications and the role of argumentation in constructing proper labelings.
This theme has been developed in subsequent theories.
3
Coming of age
Nonmonotonic reasoning obtained its impetus in 1980 with the publication of
a seminal issue of the Artificial Intelligence Journal, devoted to nonmonotonic
reasoning. The issue included papers representing three basic approaches to
nonmonotonic reasoning: circumscription [McCarthy, 1980], default logic [Reiter, 1980], and modal nonmonotonic logic [McDermott and Doyle, 1980]. These
theories suggested three different ways of meeting Minsky’s challenge by developing formalisms that do not have the monotonicity property.
On the face of it, the three approaches were indeed different, beginning with
the fact that they were based on three altogether different languages – the classical first order language in the case of circumscription, a set of inference rules
in default logic, and modal language in modal nonmonotonic logic. Still, behind these differences there was a common idea. The idea was that default
conditionals or inference rules used in commonsense derivations can be represented, respectively, as ordinary conditionals or inference rules by augmenting
their premises with additional assumptions, assumptions that could readily be
10
accepted in the absence of contrary information. In this respect, the differences
between the three theories amounted to different mechanisms of making such
default assumptions.
3.1
Circumscription
McCarthy’s circumscription was based on classical logic, and focused in a large
part on representation techniques. He stressed that circumscription is not a
“nonmonotonic logic”, but a form of nonmonotonic reasoning augmenting ordinary first order logic.
The first paper [McCarthy, 1980] connected the strategic ideas of [McCarthy
and Hayes, 1969] with the need for nonmonotonic reasoning, and described a
simplest kind of circumscription, namely domain circumscription. The second
paper [McCarthy, 1986] provided more thorough logical foundations, and introduced the more general and powerful predicate circumscription approach.
As a first description in [McCarthy, 1980], McCarthy characterized circumscription as a kind of conjectural reasoning by which humans and intelligent
computer programs jump to the conclusion that the objects they can determine
to have certain properties are the only objects that do.
The result of applying circumscription to a collection A of facts
is a sentence schema that asserts that the only tuples satisfying a
predicate are those whose doing so follows from the sentences of
A. Since adding more sentences to A might make P applicable to
more tuples, circumscription is not monotonic. Conclusions derived
from circumscription are conjectures that A includes all the relevant
facts and that the objects whose existence follows from A are all the
relevant objects.
Thus, circumscription is a tool of making conjectures. Conjectures may be
regarded as expressions of probabilistic notions such as “most birds can fly” or
they may be expressions of standard, or normal, cases. Such conjectures sometimes conflict, but there is nothing wrong with having incompatible conjectures
on hand. Besides the possibility of deciding that one is correct and the other
wrong, it is possible to use one for generating possible exceptions to the other.
Although circumscription was originally presented as a schema for adding
more formulas to a theory, just as Reiter’s CWA or Clark’s completion, it can
also be described semantically in terms of restricting the models of the theory
to those that have minimal extensions of (some of) the predicates and functions. Let P be a set of predicate symbols that we are interested in minimizing
and Z another set of predicate symbols that are allowed to vary across compared models. Predicates other than P and Z are called the fixed symbols. Let
A(P ; Z) be a first-order sentence containing the symbols P and Z. A (parallel
predicate) circumscription chooses models of A(P ; Z) that are minimal in the
extension of predicates P, assuming that these models have the same interpretation for all symbols not in P or Z. This characterization can be concisely written
11
as a second-order formula. More detailed description of circumscription in its
different forms can be found in [Lifschitz, 1994b].
The importance of predicates and functions that vary across compared interpretations has been recognized in [Etherington et al., 1985]. Without these,
it would be impossible to infer new positive instances of any predicates from
the preferred models. Accordingly, the general problem with using circumscription in applications amounted to specifying the circumscription policy: which
predicates should be varied, which should be minimized, and with what priority.
Widely varying results occur depending on these choices. In fact, as was noted
already in [McCarthy, 1980], the results of circumscription depend on the very
set of predicates used to express the facts, so the choice of representation has
epistemological consequences in making conjectures by circumscription.
The above models of circumscription can also be described as preferred models with respect to an appropriate preorder on all models. This view has later
led to the generalization of circumscription to a general preferential approach
to nonmonotonic reasoning.
3.1.1
Abnormality theories
It is somewhat misleading to reduce the essence of circumscription to minimization. Viewed as a formalism for nonmonotonic reasoning, the central concept of
McCarthy’s circumscriptive method is an abnormality theory - a set of conditionals containing the abnormality predicate ab that provides a representation
for default information. [McCarthy, 1980] justified the need in introducing such
auxiliary predicates as follows:
When we circumscribe the first order logic statement of the [missionaries and cannibals] problem together with the common sense facts
about boats etc., we will be able to conclude that there is no bridge
or helicopter. “Aha”, you say, “but there won’t be any oars either”.
No, we get out of that as follows: It is a part of common knowledge
that a boat can be used to cross a river unless there is something
wrong with it or something else prevents using it, and if our facts
don’t require that there be something that prevents crossing the
river, circumscription will generate the conjecture that there isn’t.
The price is introducing as entities in our language the “somethings”
that may prevent the use of the boat...
Using circumscription requires that common sense knowledge be expressed in a form that says a boat can be used to cross rivers unless there is something that prevents its use. In particular, it looks
like we must introduce into our ontology (the things that exist) a
category that includes something wrong with a boat or a category
that includes something that may prevent its use. Incidentally, once
we have decided to admit something wrong with the boat, we are
inclined to admit a lack of oars as such a something and to ask
questions like, “Is a lack of oars all that is wrong with the boat?”.
12
Some philosophers and scientists may be reluctant to introduce such
things; but since ordinary language allows “something wrong with
the boat” we shouldn’t be hasty in excluding it. Making a suitable
formalism is likely to be technically difficult as well as philosophically
problematical, but we must try. We challenge anyone who thinks he
can avoid such entities to express in his favorite formalism, “Besides
leakiness, there is something else wrong with the boat”.
In [McCarthy, 1986], McCarthy proposed a uniform principle for representing
default claims in circumscription.
It turns out that many common sense facts can be formalized in a
uniform way. A single predicate ab, standing for “abnormal” is circumscribed with certain other predicates and functions considered
as variables that can be constrained to achieve the circumscription
subject to the axioms. This also seems to cover the use of circumscription to represent default rules.
Many people have proposed representing facts about what is “normally” the case. One problem is that every object is abnormal in
some way, and we want to allow some aspects of the object to be
abnormal and still assume the normality of the rest. We do this
with a predicate ab standing for “abnormal”. We circumscribe ab z.
The argument of ab will be some aspect of the entities involved.
Some aspects can be abnormal without affecting others. The aspects themselves are abstract entities, and their unintuitiveness is
somewhat a blemish on the theory.
For example, to say that normally birds fly, we can use
∀x : Bird(x) ∧ ¬ab aspect1(x) ⊃ F ly(x).
Here the meaning of ab aspect1(x) is something like “x is abnormal with respect to flying birds”. There can be many different aspects of abnormality, and
they are indexed according to kind. The circumscription would then minimize
abnormalities, allowing relevant predicates to vary (e.g., F ly).
An important advantage of representation with abnormality predicates is
that, by asserting that a certain object is abnormal in some respect, we can
block, or defeat, an associated default rule without asserting that its consequent
is false. For example, by asserting ab aspect1(T weety), we block the application
of the default “Birds fly” to Tweety without asserting that it cannot fly. This
feature has turned out to be useful in many applications.
Abnormality theories have been widely used both in applications of circumscription, and in other theories. Some major examples are inheritance theories
[Etherington and Reiter, 1983], logic-based diagnosis [Reiter, 1987b], naming
defaults in the abductive approach of [Poole, 1988a], general representation of
defaults in [Konolige and Myers, 1989] and reasoning about time and action.
Abnormality theories have brought out, however, several problems in the
application of circumscription to commonsense reasoning. One of the most
13
pressing was the already mentioned specificity problem arising when there are
conflicting defaults. In combining two defaults, “Birds fly” and “Penguins can’t
fly”, the specificity principle naturally suggests that the second, more specific,
default should be preferred. A general approach to handle this problem in circumscription, suggested in [Lifschitz, 1985] and endorsed in [McCarthy, 1986],
was to impose priorities among minimized predicates and abnormalities. The
corresponding variant of circumscription has been called prioritized circumscription. [Grosof, 1991] has generalized prioritized circumscription to a partial order
of priorities.
3.1.2
Further developments and applications
[Lifschitz, 1985] described the concept of parallel circumscription and also treated
prioritized circumscription. He addressed also the problem of computing circumscription and showed that, in some cases, circumscription can be replaced by
an equivalent first-order formula. Further results in this direction have been
obtained in [Doherty et al., 1995].
Independently of McCarthy, [Bossu and Siegel, 1985] have provided semantic
account of nonmonotonic reasoning for a special class of minimal models of a
first-order theory. [Lifschitz, 1987a] has proposed a more expressive variation of
circumscription, called pointwise circumscription which, instead of minimizing
the extension of a predicate P (x) as a whole, minimizes the truth-value of P (a)
at each element a.
[Perlis, 1986] and [Poole, 1989b] showed the inadequacies of circumscription
to deal with counterexamples like the lottery paradox of Kyburg. In a lottery,
it is known that some person will win, yet, for any individual x, the default
should be that x does not win. If these facts are translated in a straightforward
way into circumscription, it is impossible to arrive at the default conclusion,
for any given individual x, that x will not win the lottery. To remedy this and
similar anomalies, [Etherington et al., 1991] proposed a scoped circumscription
in which the individuals over whom minimization proceeds are limited by a
scoping predicate.
We mentioned earlier that Reiter’s Closed World Assumption works only
for theories that have a unique least model. For this case, it has been shown
in [Lifschitz, 1985] that CWA is equivalent to circumscription (modulo unique
names and domain closure assumptions). Similarly, it has been shown in [Reiter, 1982] that an appropriate circumscription always implies Clark’s predicate
completion, but not vice versa. In other words, the minimal models determined
by circumscription form in general only a subset of models sanctioned by completion. Recently, it has been shown in [Lee and Lin, 2004] that circumscription
can be obtained from completion by augmenting the latter with so-called ‘loop’
formulas.
In the earliest circumscriptive solutions to the frame problem, the inertia
rule was stated using an abnormality predicate. This formalization has succumbed, however, to the Yale Shooting Problem, mentioned earlier. [Baker,
1989] presented another solution to the Yale Shooting problem in the situation
14
calculus, using a circumscriptive inertial axiom. A circumscriptive approach
to the qualification problem was presented in [Lifschitz, 1987b]; it used an explicit relation between an action and its preconditions, and circumscriptively
minimized preconditions, eliminating thereby unknown conditions that might
render an action inefficacious. Further details and references can be found in
[Shanahan, 1997].
3.2
Default Logic
A detailed description of default logic, its properties and variations can be found
in [Antoniou and Wang, 2006]. Accordingly, we will sketch below only general
features of the formalism, sufficient for determining its place in the nonmonotonic reasoning field.
In both circumscription and modal nonmonotonic logic (see below), default
statements are treated as logical formulas, while in default logic [Reiter, 1980]
they are represented as inference rules. In this respect, Reiter’s default logic
has been largely inspired by the need to provide logical foundations for the
procedural approach to nonmonotonicity found in deductive databases, logic
programming and Doyle’s truth maintenance. In particular, default logic begins
with interpreting “In the absence of any information to the contrary, assume
A” as “If A can be consistently assumed, then assume it.”
A default is a rule of the form A : b/C, intended to state something like: ‘if
A is believed, and each B ∈ b can be consistently believed, then C should be
believed’. A is called a prerequisite of a default rule, b a set of its justifications.
The flying birds default is represented by the rule Bird(x) : F ly(x)/F ly(x).
Reiter defined a default theory to be a pair (D, W ), where D is a set of default
rules and W is a set of closed first-order sentences. The set W represents what
is known to be true of the world. This knowledge is usually incomplete, and
default rules act as mappings from this incomplete theory to a more complete
extension of the theory. They partly fill in the gaps with plausible beliefs.
Extensions are defined by a fixed point construction. For any set S of firstorder sentences, Γ(S) is defined as the smallest set satisfying the following three
properties:
1. W ⊆ Γ(S);
2. Γ(S) is closed under first-order logical consequence;
3. If A : b/C belongs to D and A ∈ Γ(S) and ¬B ∈
/ S, for any B ∈ b, then
C ∈ Γ(S).
Then a set E is an extension of the default theory iff Γ(E) = E, that is, E
is a fixed point of the operator Γ.
The above definition embodies an idea that an extension must not contain
“ungrounded” beliefs, i.e., every formula in it must be derivable from W and
the consequents of applied defaults in a non-circular way. It is this property
that distinguishes default logic from circumscription. Though a standard way
15
of excluding unwanted elements employs a minimality requirement, minimality
alone is insufficient to exclude ungrounded beliefs. Reiter’s idea can also be
expressed as follows. At first stage we assume a conjectured extension and use
it to determine the set of applicable inference rules, namely default rules such
that their justifications are consistent with the assumption set. Then we take
the logical closure of these applicable inference rules, and if it coincides with the
candidate extension, the latter is vindicated. In this sense, an extension is a set
of beliefs which are in some sense “justified” or “reasonable” in light of what
is known about the world (cf. [Etherington, 1987]). This interpretation can
be seen as a paradigmatic form of general explanatory nonmonotonic reasoning
that we will discuss later.
Extensions are deductively closed sets that are closed also with respect to
the rules of the default theory. Moreover, they are minimal such sets, and hence
different extensions are incomparable with respect to inclusion. Still, not every
minimal theory closed with respect to default rules is an extension.
Multiple extensions of a default theory are possible. The perspective adopted
on these in [Reiter, 1980] was that any such extension is a reasonable belief set
for an agent. A typical example involves Richard Nixon who is quaker and
republican. Quakers (typically) are pacifists. Republicans (typically) are not
pacifists. One might conclude that Nixon is a pacifist using the first default, but
also that Nixon is not a pacifist because he is a republican. In this situation the
default logic generates two extensions, one containing the belief that Nixon is a
pacifist, the other one containing the belief that he is not. Speaking generally,
there are many applications where each of the extensions is of interest by itself.
For instance, diagnostic reasoning (see below) is usually modeled in such a way
that each diagnosis corresponds to a particular extension.
A default theory does not always have extensions, a simplest example being
the default theory (∅, {true : ¬A/A}). However, an important subclass consisting of normal default rules of the form A : B/B always has an extension.
On Reiter’s intended interpretation, normal defaults provided a representation
of commonsense claims “If A, then normally B”. Reiter developed a complete
proof theory for normal defaults and showed how it interfaces with a top-down
resolution theorem prover. He has considered this proof theory as one of the
advantages of default logic. Moreover, extensions of normal default theories are
semi-monotonic: if E is an extension of a normal default theory (W, D) then
the normal default theory (W, D ∪ D0 ) has an extension E0 such that E ⊆ E0 .
In other words, additional defaults may augment existing extensions or produce
new ones, but they never destroy the extensions obtained before.
Reiter has mentioned in [Reiter, 1980] that he knows of no naturally occurring default which cannot be represented in this form. However, the later paper
[Reiter and Criscuolo, 1981] showed that in order to deal with default interactions, we need at least semi-normal defaults of the form A : B ∧C/C. Moreover,
though some authors2 have questioned the usefulness of defaults which are not
semi-normal, Paul Morris has shown in [Morris, 1988] that such default rules
2 E.g.,
[Brewka et al., 1997].
16
can provide a solution for the Yale Shooting Anomaly.
There has been a number of attempts of providing a direct semantic interpretation of default logic. It was suggested already in [Reiter, 1980] that default
rules could be viewed as operators that restrict the models of the set of known
facts W . Developing this idea, [Etherington, 1987] has suggested a semantics
based on combining preference relations ≥δ on sets of models determined by
individual defaults δ. For normal default theories, extensions corresponded to
preferred sets of models. For full generality, however, an additional condition
of Stability was required which stated, roughly, that the corresponding model
set is ‘accessible’ from M OD(W ) via a sequence of ≥δ -links corresponding to
defaults δ with unrefuted justifications. This operational, or quasi-inductive
description of extensions has been developed later in a number of works – see
[Makinson, 2003; Antoniou and Wang, 2006].
Default logic is a more general and more expressive formalism than circumscription. In one direction, [Imielinski, 1987] has shown that even normal
default theories cannot be translated in a modular way to circumscription. In
the other direction, [Etherington, 1987] has shown that minimizing the predicate P , with all other predicates varying, corresponds to the use of the default
: ¬P (x)/¬P (x). Speaking generally, circumscription corresponds to default
theories involving only simplest normal defaults without prerequisites; skeptical reasoning in such theories is equivalent to minimization with respect to
the ordering on interpretations determined by the sets of violated defaults (cf.
Theorem 4.7 in [Poole, 1994a]).
[Etherington and Reiter, 1983] used default logic to formalize inheritance hierarchies, while [Reinfrank et al., 1989] have shown that justifications of Doyle’s
truth maintenance system can be directly translated into default rules, and then
extensions of the resulting default theory will exactly correspond to admissible
labelings of TMS.
A large part of the expressive power of default logic is due to the representation of defaults as inference rules. This representation avoids some problems arising with formula-based interpretations of defaults (such as contraposition) and provides a natural framework for logic programming and reasoning in
databases. Unfortunately, this representation also has its problems, problems
considered by many researchers as more or less serious drawbacks.
One kind of problems concerned an apparent discrepancy between commonsense claims “If A then normally B” and their representation in terms of normal
defaults A : B/B. For example, default logic does not allow for reasoning by
cases: From “Italians (normally) like wine” and “French (normally) like wine”
we cannot conclude that the person at hand that is Italian or French likes wine.
This is because, in default logic, a default can only be applied if its prerequisite
has already been derived. A more profound problem has been pointed out in
[Makinson, 1989]. A general class of preferential nonmonotonic inference relations (see below) satisfies a natural Cumulativity postulate stating that, if A
entails both B and C, then A ∧ B should entail C. Makinson has shown, however, that default logic does not satisfy Cumulativity (see [Antoniou and Wang,
2006]). This discrepancy was a first rigorous indication that default logic can17
not be directly subsumed by the preference-based approach to nonmonotonic
reasoning.
A problem of a different kind has been noticed in [Poole, 1989b] that concerned joint consistency of justifications of default rules. Namely, the definition
of extensions only enforces that each single justification of an applied default
is consistent with the generated extension, but nothing guarantees the joint
consistency of the justifications of all applied defaults (see again [Antoniou and
Wang, 2006] for examples and discussion).
These and other problems have led to a number of proposed modifications of
default logic. Most of them are described in [Antoniou and Wang, 2006]. Also,
driven mainly by the analogy with logic programming, a number of authors have
suggested to generalize the notion of extension to that of a partial extension
– see [Baral and Subrahmanian, 1991; Przymusinska and Przymusinski, 1994;
Brewka and Gottlob, 1997] (cf. also [Antonelli, 1999; Denecker et al., 2003]).
Unfortunately, these modifications still have not gained widespread acceptance.
A generalization of a different kind has been proposed in [Gelfond et al.,
1991], guided by the need to provide a logical basis for disjunctive logic programming, as well as more perspicuous ways of handling disjunctive information. A disjunctive default theory is a set of disjunctive defaults, rules of the
form a : b/c, where a, b, c are finite sets of formulas. The authors described a
semantics for such default theories that constituted a proper generalization of
Reiter’s default logic, as well as the semantics for disjunctive databases from
[Gelfond and Lifschitz, 1991]. They also have shown how the above mentioned
Poole’s problem of joint justifications can be avoided by using disjunctive default
rules.
[Marek and Truszczyński, 1989] introduced the notion of a weak extension
as a default counterpart of stable expansions in autoepistemic logic and models
of Clark’s completion in logic programming (see below). Weak extensions can
be defined3 as fixed points of a modified operator Γw , obtained by weakening
the third condition in the above definition of Γ to
(3’) If A : b/C ∈ D, A ∈ S and ¬B ∈
/ S, for any B ∈ b, then C ∈ Γ(S).
A number of studies have dealt with elaboration of default logic as a logical
formalism. Thus, [Marek and Truszczyński, 1989] suggested a more ‘logical’
description of default logic using the notion of a context-depended proof as a
way of formalizing Reiter’s operator Γ. [Marek et al., 1990] described a general
theory of nonmonotonic rule systems based on abstract rules a:b/A having an informal interpretation similar to Doyle’s justifications: ‘If all a’s are established,
and none of b’s is established now or ever, conclude A’. The authors studied
abstract counterparts of extensions and weak extensions in this framework. An
even more general default system in the framework of the domain theory has
been developed in [Zhang and Rounds, 1997]. A reconstruction of default derivability in terms of restricted Hilbert-type proofs has been described in [Amati
et al., 1994]. Finally, a sophisticated representation of default logic as a sequent
3 Cf.
[Halpern, 1997].
18
calculus based on both provability and unprovability sequents has been given in
[Bonatti and Olivetti, 2002].
Reiter has mentioned in [Reiter, 1987a] that, because the defaults are represented as inference rules rather than object language formulas, they cannot
be reasoned about within the logic. Thus, from “Normally canaries are yellow”
and “Yellow things are never green” we cannot conclude in default logic that
“Normally canaries are never green”. As far as we know, a first default system
with such meta-rules allowing to infer derived default rules from given ones has
been suggested in [Thiele, 1990] (see also [Brewka, 1992]). [Bochman, 1994]
introduced default consequence relations as a generalization of both default and
modal formalizations of nonmonotonic reasoning. Default consequence relation
is a logical (monotonic) inference system based on default rules of the form
a:b C similar to that in [Marek et al., 1990]. The basic system was required
to satisfy the following postulates:
Monotonicity
Cut
If a : b A and a ⊆ a0 , b ⊆ b0 , then a0 : b0 A.
If a : b A and a, A : b B, then a : b B.
Consistency
A:Af
together with postulates securing that default rules respect classical entailment
both in premises and conclusions. A binary derivability operator associated
with a default consequence relation was defined as Cn(u, v) = {A | u : v A}.
This operator simplified a characterization of extensions and weak extensions,
called expansions in [Bochman, 1994], due to their correspondence with stable
expansions of autoepistemic logic. Namely, a set u of propositions is an extension
of a default consequence relation if u = Cn(∅, u), and it is an expansion if
u = Cn(u, u) (where u denotes the complement of u).
Reiter’s default rules A:b/C were translated as the rules A:¬b C. Then
it was shown that extensions of the resulting default consequence relation correspond precisely to extensions of Reiter’s default logic. Moreover, it has been
shown that the following additional postulates also preserve extensions:
Reflexivity
A: A.
Negative Factoring If a, B : b f and a : b, B A, then a : b A.
The resulting class of default consequence relations can be viewed as an underlying monotonic logic of Reiter’s default logic. In contrast, the logic adequate
for reasoning with expansions can be obtained by adopting instead a powerful
alternative postulate:
Factoring
If a, B : b A and a : b, B A, then a : b A.
Such default consequence relations have been called autoepistemic. In autoepistemic consequence relations, extensions collapse to expansions (weak extensions). It was shown that this logic provides an exact non-modal counterpart
of Moore’s autoepistemic logic (see below).
19
3.3
Modal nonmonotonic logics
The third seminal paper in the 1980 issue of Artificial Intelligence, [McDermott
and Doyle, 1980], is the beginning of a modal approach to nonmonotonic reasoning. McDermott and Doyle provided first a broad picture of instances of
nonmonotonic reasoning in different parts of AI, and stressed the need for a formal logical analysis of these phenomena. The theory was also clearly influenced
by the need to provide a formal account of truth maintenance. As a general suggestion, the authors proposed to expand the notation in which logical inference
rules are stated by adding premises like ‘unless proven otherwise’.
The modal nonmonotonic logic was formulated in a modal language containing a modal operator M p with the intended meaning “p is consistent with
everything believed”. In accordance with this understanding, a direct way of
making assumptions in this setting might consist in accepting an inference rule
of the form “If 0 ¬A, then ` M A”. Such a rule, however, would be circular
relative to the underlying notion of derivability. And the way suggested by the
authors was a fixed-point construction much similar to that of Reiter’s default
logic, but formulated this time entirely in the modal object language.
For a set u of propositions, let us denote by M u the set {M A | A ∈ u},
and similarly for ¬u, etc. In addition, u will denote the complement of u with
respect to a given modal language. Using this notation, a set s is a fixed point
of a modal theory u, if it satisfies the equality
s = Th(u ∪ M ¬s)
The set of nonmonotonic conclusions of a modal theory was defined in [McDermott and Doyle, 1980] as an intersection of all its fixed points.
The initial formulation of modal nonmonotonic logic has turned out to be
unsatisfactory, mainly due to the fact that it secured no connection between
a modal formula M C and its objective counterpart C, so that even the nonmonotonic theory {M C, ¬C} was consistent. In response to this latter difficulty,
[McDermott, 1982] developed a stronger version of the formalism based on the
entailment relation of standard modal logics instead of first-order logic. In fact,
the decisive modification was an adoption of the Necessitation rule A ` LA.
The corresponding fixed points were defined now as sets satisfying the equality
s = CnS (u ∪ M ¬s),
where CnS is a provability operator of some modal logic S containing the necessitation rule.
In what follows, we will call such fixed points S-extensions (following [Konolige, 1988] and [Schwarz, 1990], but in contrast with the influential subsequent
terminology of Marek and Truszczyński who have called them S-expansions).
As has been shown by McDermott, the stronger is the underlying modal
logic, the smaller is the set of extensions, and hence the larger is the set of
nonmonotonic consequences of a modal theory. It has turned out, however, that
the modal nonmonotonic logic based on the strongest modal logic S5 collapses
20
to a monotonic system. So the resulting suggestion was somewhat indecisive,
namely a range of possible modal nonmonotonic logics without clear criteria
for evaluating the merits of the alternatives. However, this indecisiveness has
turned out to be advantageous in the subsequent development of the modal
approach.
Neither Doyle or McDermott pursued the modal approach much beyond
these initial stages. Actually, both have expressed later criticism about the
whole logical approach to AI.
3.3.1
Autoepistemic logic
As an alternative response to the problems with initial formulations of modal
nonmonotonic logics, Robert C. Moore proposed his autoepistemic logic (AEL)
in [Moore, 1985]. Autoepistemic logic has had a great impact on the development of nonmonotonic reasoning; for a certain period of time it has even been
considered as a correct replacement of the theory of McDermott and Doyle.
Moore has suggested that modal nonmonotonic logic should be interpreted
not as a theory of default (defeasible) reasoning, but as a model of (undefeasible)
autoepistemic reasoning of an ideally rational agent. Following a suggestion
made by Robert Stalnaker in 1980 (published much later as [Stalnaker, 1993]),
Moore reconstructed nonmonotonic logic as a model of an ideally rational agent’s
reasoning about its own beliefs. He argued that purely autoepistemic reasoning
is not defeasible; it is nonmonotonic because it is context-sensitive. He defined a
semantics for which he showed that autoepistemic logic is sound and complete.
Instead of the modal consistency operator M , he used the dual belief operator L. Moore stressed the epistemic interpretation of a default rule as one
that licenses a conclusion unless something that the agent knows blocks it.
Nonmonotonicity can be achieved by endowing an agent with the introspective
ability, namely an ability to reflect on its own beliefs in order to infer sentences
expressing what it doesn’t believe. Introspective here means that the agent is
completely aware about his beliefs: if a formula p belongs to the set of beliefs
B of the agent, then also Lp has to belong to B, and if p does not belong to B,
then ¬Lp must be in B. Sets of formulas satisfying these properties were first
discussed by Stalnaker, who called them stable sets.
Stable sets are essentially sets of formulas globally valid in S5-models (see
[Konolige, 1988]). As has been shown in [Moore, 1985] (and implicitly already
in [McDermott, 1982]) a stable set is uniquely determined by its objective (nonmodal) propositions. This fact has provided later a basic link between modal
and non-modal approaches to nonmonotonic reasoning.
A stable expansion of a set u of premises was defined as a stable set that is
grounded in u. Again, the formal definition was based on a fixed point equation:
s = Th(u ∪ Ls ∪ ¬Ls)
The groundedness condition ensured that every member of an expansion has
some reason tracing back to s. As in default logic, conflicting AEL rules can
lead to alternative stable sets of beliefs a reasoner may adopt.
21
Moore has argued that the original nonmonotonic logic of McDermott and
Doyle was simply too weak to capture the notions they wanted, while [McDermott, 1982] strengthened it in a wrong way. He observed that, in the modal
nonmonotonic logic, the component Ls is missing from the ‘base’ of the fixed
points, which means that McDermott and Doyle’s agents are omniscient as to
what they do not believe, but they may know nothing as to what they do believe.
A difference between the nonmonotonic modal and autoepistemic logics can
be seen on the theory {LP →P }, which has a single fixed point that does not
include P , though it also has a stable expansion containing P . As argued by
Moore, this makes the interpretation of L in nonmonotonic modal logic more like
“justified belief” than simple belief. On the other hand, already [Konolige, 1988]
argued that the second expansion is intuitively unacceptable. It corresponds to
an agent arbitrarily entering P, hence also LP, into her belief set.
Since nonmonotonic S5 collapses to monotonic S5, Moore (following Stalnaker) has suggested to retreat not to S4, as was suggested in [McDermott,
1982], but to K45 obtained from S5 by dropping reflexivity LP → P . Moore
has also shown that the axioms of K45 do not change stable expansions.
[Levesque, 1990] generalized Moore’s notion of a stable expansion to the full
first-order case. He provided a semantic account of stable expansions in terms
of a second modal operator O, where O(A) is read as “A is all that is believed.”
In this approach, nonmonotonicity was pushed entirely into the scope of the O
operator. Levesque’s ideas have been systematically presented and applied to
the theory of knowledge bases in [Levesque and Lakemeyer, 2000].
It has been shown in [Konolige, 1989] that autoepistemic logic is a more expressive formalism than circumscription (see also [Niemelä, 1992]). The relation
between autoepistemic and default logic, however, has turned out to be more
complex, and it has created an important incentive for further development of
a theory of nonmonotonic reasoning.
3.3.2
Unified modal theories
[Konolige, 1988] has attempted to translate default logic into autoepistemic
logic, and vice versa. To this end, he suggested to rephrase a default rule
A : B1 , . . . Bn /C as a modal formula (LA ∧ ¬L¬B1 ∧ · · · ∧ ¬L¬Bn ) ⊃ C.
Despite Konolige’s intentions, however, this translation has turned out to
be inappropriate for representing default rules. Furthermore, [Gottlob, 1994]
has shown that there can be no modular translation of default logic into autoepistemic logic, even if we restrict ourselves to normal defaults, or just to
inference rules without justifications. Still, [Marek and Truszczyński, 1989] has
shown that Konolige’s translation works for prerequisite-free defaults. As to the
general case, Marek and Truszczyński have shown that the translation provides
instead an exact correspondence between stable expansions and weak extensions.
A more organized picture started to emerge with the result of [Schwarz,
1990] according to which autoepistemic logic is just one of the nonmonotonic
logics in the general approach of [McDermott, 1982]. Namely, it is precisely the
nonmonotonic logic based on K45, or even KD45, if we ignore inconsistent ex22
pansions4 . This result, as well as difficulties encountered in interpreting default
logic, revived interest in the whole range of modal nonmonotonic logics based
on ‘negative introspection’ in the works of Marek, Truszczyński and Schwarz.
A minimal modal nonmonotonic logic in this range is based on a modal logic
N that does not contain modal axioms at all (see [Fitting et al., 1992]). Corresponding N -extensions have been introduced in [Marek and Truszczyński, 1990]
and called iterative expansions. A systematic study of modal nonmonotonic logics based on different underlying modal logics can be found in [Marek et al., 1993;
Marek and Truszczyński, 1993]. This study has shown the importance of many
otherwise esoteric modal logics for nonmonotonic reasoning, such as S4F, SW5,
and KD45. All these logics have, however, a common semantic feature: in the
terminology of [Segerberg, 1971], each of them is characterized by certain Kripke
models of depth two having a unique final cluster (see [Schwarz, 1992a]).
An adequate modal interpretation of defaults logic in modal nonmonotonic
logics of [McDermott, 1982] has been suggested in [Truszczyński, 1991]. The
interpretation was based on the following more complex translation:
A : B1 , . . . Bn /C ⇒ (LA ∧ LM B1 ∧ · · · ∧ LM Bn ) ⊃ LC
Under this interpretation, a whole range of modal logics between T − and
S4F can be used as a “host” logic. In addition, the translation can be naturally
extended to disjunctive default rules of [Gelfond et al., 1991]. In this sense,
modal nonmonotonic logics also subsumed disjunctive default logic. A different
translation of default logic into the modal logic KD4Z has been suggested in
[Amati et al., 1997].
[Bochman, 1994; 1995b] introduced the notion of a modal default consequence relation (see also [Bochman, 1998c]). These consequence relations were
defined in a language with a modal operator, but otherwise involved the same
rules as general default consequence relations, described earlier. A default consequence relation in a modal language has been called modal if it satisfied the
following two modal axioms:
A : LA and
: A ¬LA.
Basically, the transition from default to modal nonmonotonic logics amounts
to adding these default rules to a default theory. A similar translation has been
used in [Janhunen, 1996] and rediscovered a number of times in later publications. Modal default consequence relations have turned out to be a convenient
tool for studying modal nonmonotonic reasoning. Thus, both autoepistemic
reasoning and reasoning with negative introspection acquired a natural characterization in this framework. In particular, S-extensions of a modal theory u
have been shown to coincide with extensions of the least modal default consequence relation containing u and the modal axioms of the modal logic S.
Under certain reasonable conditions, modal consequence relations have turned
out to be reducible to their nonmodal, default sub-relations in a way that
4 KD45 is obtained from S5 by replacing the T axiom LA⊃A with a weaker D axiom
LA⊃¬L¬A.
23
preserved the associated nonmonotonic semantics. These results were used in
[Bochman, 1998c] for establishing a two-way correspondence between modal and
default formalizations. In particular, it has been shown that modal autoepistemic logic is equivalent to a nonmodal autoepistemic consequence relation,
described earlier.
Lin and Shoham have suggested in [Lin and Shoham, 1992] a bimodal system
purported to combine preferential and fixed-point approaches to nonmonotonic
reasoning. The bridge was provided by a preference relation on models that
minimized knowledge for fixed sets of assumptions. Semantic models of a bimodal system are triples (Mk , Mb , V ) such that Mk and Mb are sets of possible
worlds, Mb ⊆ Mk , and V is a valuation function assigning each world a propositional interpretation. Mk and Mb serve as ranges of two modal operators K
and B representing, respectively, the knowledge and assumptions (beliefs) of an
introspective agent. A model M1 is preferred over a model M2 , if they have
the same B-worlds, but M1 has a larger set of K-worlds. Finally, a preferred
model is a model M = (Mk , Mb , V ) such that Mk = Mb , and there is no model
M1 that is preferred over M.
Lin and Shoham were able to show that, by using suitable translations, default and autoepistemic logic, as well as the minimal belief logic of [Halpern
and Moses, 1985] can be embedded into their system. [Lifschitz, 1991b; 1994a]
described a simplified version of such a bimodal logic, called MBNF (Minimal
Belief with Negation-as-Failure), extended it to the language with quantifiers,
and considered a representation of default logic, circumscription and logic programming in this framework. Yet another bimodal theory, Autoepistemic Logic
of Minimal Beliefs (AELB) has been suggested in [Przymusinski, 1994] as an extension of Moore’s autoepistemic logic and has been shown to subsume circumscription, CWA and its generalizations, epistemic specifications and a number
of semantics for logic programs.
It has been shown in [Lifschitz and Schwarz, 1993] that a significant part of
MBNF, namely theories with protected literals, can be embedded into autoepistemic logic (see also [Chen, 1994]). Furthermore, it has been shown in [Schwarz
and Truszczyński, 1994] that in most cases bimodal nonmonotonic logics can be
systematically reduced to ordinary unimodal nonmonotonic logics by translating the belief (assumption) operator B as ¬K¬K (see also [Bochman, 1995c]
for a similar reduction of MBNF).
Finally, a partial (four-valued) generalization of modal and default nonmonotonic logics has been studied in [Denecker et al., 2003].
3.4
Logic Programming
Logic programming has been based on the idea that program rules should have
both a procedural and declarative (namely, logical) meaning. This double interpretation was intended to elevate the programming process by basing it on
a transparent and systematic logical representation of real world information.
Fortunately, it has been discovered quite early that logic programming with
negation as failure allows us to express significant forms of nonmonotonic rea24
soning. Moreover, general nonmonotonic formalisms, described earlier, inspired
the development of new semantics for logic programs. These developments have
made logic programming an integral part of nonmonotonic reasoning research
functioning as a general computational mechanism for knowledge representation
and nonmonotonic reasoning (see [Baral, 2003]).
A normal logic program is a set of program rules of the form A ← a, not b,
where A is a propositional atom (the head of the rule), a and b are finite sets
of atoms (forming the body of the rule), and not is negation as failure.
The semantics for logic programs have gradually evolved from definite rules
without not to more general program rules involving negation as failure. For
definite logic programs, such a semantics has been identified with the unique
minimal model of the corresponding logical theory [van Emden and Kowalski,
1976]. Both the Closed World Assumption (CWA) [Reiter, 1978] and completion
theory [Clark, 1978] have provided its syntactic characterization. This semantics has been conservatively extended in [Apt et al., 1988] to stratified normal
programs. It has been called the perfect semantics in [Przymusinski, 1988]. See
[Shepherdson, 1988] for a detailed description of these initial semantics.
Unlike the CWA, Clark’s completion is naturally definable for any normal
program. It has turned out, however, that the corresponding supported semantics for logic programs produces inadequate results when applied to arbitrary
(i.e., non-stratified) programs. In response to this problem, two competing semantics have been suggested for arbitrary normal programs: the stable model
semantics [Gelfond and Lifschitz, 1988], and the well-founded semantics [van
Gelder et al., 1991]. A large literature has been devoted to studying the relationships between these three semantics, as well as between the latter and the
nonmonotonic formalisms.
Under the stable model semantics, a normal program rule A ← a, not b
corresponds to a rule a : ¬b/A of default logic (see [Marek and Truszczyński,
1989]). This translation establishes a one-to-one correspondence between stable
models of a normal logic program and extensions of the associated default theory.
In this sense, logic programs under the stable semantics capture the main idea, as
well as primary representation capabilities, of default logic. Still, this embedding
of logic programs into default logic is unidirectional, since not every default
theory corresponds in this sense to a logic program.
The stable model semantics is sufficient, in particular, for an adequate representation of Doyle’s truth-maintenance system [Doyle, 1979]. If we translate
each nonmonotonic justification as a program rule then there is a 1-1 correspondence between stable models of the resulting logic program and admissible
labelings of the original justification network [Elkan, 1990].
Guided partly by the correspondence with default logic, [Gelfond and Lifschitz, 1991] have suggested to include classical negation ¬ into logic programs,
in addition to negation-as-failure not. They argued that some facts of commonsense reasoning can be represented more easily when classical negation is
available. [Alferes and Pereira, 1992] defined a parameterizable schema to encompass and characterize a range of proposed semantics for such extended logic
programs. By adjusting the parameters they have been able to specify several
25
semantics using two kinds of negation. [Gelfond, 1994] expanded the syntax
and semantics of logic programs even further to an epistemic formalism with
modal operators allowing for the representation of incomplete information in
the presence of multiple extensions.
[Gelfond and Lifschitz, 1988] showed that the stable model semantics is also
equivalent to some translation of logic programs into autoepistemic theories.
[Pearce, 1997] has suggested that a certain three-valued logic called here-andthere (HT) can serve as an underlying logic of the stable model semantics.
Continuing this line of research, [Lifschitz et al., 2001] have shown that HT
provides a characterization of strong equivalence for logic programs under the
stable semantics5 .
A stable model of a logic program is also a model of its completion, while
the converse does not hold. [Marek and Subrahmanian, 1992] showed the relationship between supported models of normal programs and expansions of autoepistemic theories. [Fages, 1994] has shown, however, that a natural syntactic
acyclicity condition on the rules of a program, subsequently called “tightness”,
is sufficient for coincidence of the stable and supported semantics. As a practical consequence, this has reduced the task of computing the stable models for
tight programs to classical satisfiability [Babovich et al., 2000]. Finally, as a
most important recent development, [Lin and Zhao, 2002] showed that a classical logical description of the stable semantics for an arbitrary normal program
can be obtained by augmenting its Clark’s completion with what they called
“loop formulas”. This formulation has opened the way of computing the stable
semantics of logic programs using SAT solvers.
Unlike the stable semantics, the well-founded semantics suggested in [van
Gelder et al., 1991] does not correspond to any of the nonmonotonic formalisms,
described earlier. It determines, however, a unique model for every normal
logic program. This semantics has been elaborated upon by Przymusinski to a
theory of partial stable models [Przymusinski, 1991b; 1990]; the well-founded
model constituted the least model in this hierarchy. Przymusinski has shown,
in effect, that the well-founded semantics can be seen as the three-valued stable
semantics (see also [Dung, 1992]). [Fitting, 1991] has extended the well-founded
semantics to a family of lattice-based logic programming languages. Denecker
has characterized the WFS as based on the principle of inductive definitions
- see [Denecker et al., 2001]. Finally, [Cabalar, 2001] has shown that a twodimensional variant of the logic HT, called HT 2 , can serve as the underlying
logic of WFS.
An argumentation-theoretic analysis of the different semantics for normal
logic programs has been suggested in [Dung, 1995a]. He has shown, in particular, that the stable, partial stable and well-founded semantics are based on
different argumentation principles of rejecting negation-as-failure assumptions
(see below). An alternative representation of normal logic programs and their
semantics has been given in [Bochman, 1995a; 1996b] in the framework of de5 Two programs are strongly equivalent if they have the same stable model semantics for
any extension of these programs with additional rules.
26
fault consequence relations from [Bochman, 1994]. The framework has provided
a relatively simple representation of various semantics for such programs, as well
as allowed to single out different kinds of logical reasoning that are appropriate
with respect to these semantics. It has been shown, in particular, that stable and supported models are precise structural counterparts of, respectively,
extensions and expansions of general default consequence relations.
An important line of research in logic programming has concentrated on
extending the expressive capabilities of program rules by allowing disjunction
in their heads, namely to rules of the form c ← a, not b, where c is a set of
atoms (treated disjunctively). For such disjunctive logic programs, Reiter’s
Closed World Assumption has been generalized to the Generalized Closed World
Assumption (GCWA) [Minker, 1982] and its derivatives. The model theoretic
definition of the GCWA states that one can conclude the negation of a ground
atom if it is false in all minimal Herbrand models. This makes disjunctive
programming quite close in spirit to McCarthy’s circumscription - see [Oikarinen
and Janhunen, 2005] for a detailed comparison. Initial semantics for disjunctive
programs containing negation in premises of the rules have been summarized in
[Lobo et al., 1992].
[Przymusinski, 1991a] introduced the stable model semantics for disjunctive logic programs that generalized corresponding semantics for normal logic
programs. Depending upon whether only total (2-valued) or partial (3-valued)
models are used, one obtains the disjunctive stable semantics or the partial
disjunctive stable semantics, respectively. He showed that for locally stratified disjunctive programs, both disjunctive semantics coincide with the perfect
model semantics.
Unfortunately, absence of clear grounds for adjudicating semantics for logic
programs has resulted in a rapid proliferation of suggested semantics, especially
for disjunctive programs. This proliferation has obviously created a severe problem. [Dix, 1991; 1992] has provided a systematic analysis of various semantics
for normal and disjunctive programs based on such properties as associated nonmonotonic inference relations, modularity, relevance and the principle of partial
evaluation.
Even the two-valued stable semantics for disjunctive logic programs is already not subsumed by default logic, but requires its extension to disjunctive
default logic [Gelfond et al., 1991]. It can be naturally embedded, however, into
more general modal nonmonotonic formalisms such as Lifschitz’ MBNF [Lifschitz, 1994a] that we mentioned at the end of the preceding section. These formalisms have suggested, in turn, a generalization of disjunctive program rules to
rules involving negation as failure in heads, namely rules of the form not d, c ←
a, not b, where c and d are sets of atoms - see [Lifschitz and Woo, 1992; Inoue
and Sakama, 1994]. The stable model semantics is naturally extendable to such
generalized programs in a way that preserves the correspondence with these nonmonotonic formalisms. Moreover, the alternative well-founded and partial stable
semantics are representable in a many-valued extension of such a modal nonmonotonic system, suggested in [Denecker et al., 2003]. Yet more powerful generalization has been introduced in [Lifschitz et al., 1999]) that made use of pro27
gram rules with arbitrary logical formulas in their heads and bodies. Such results
as the correspondence between the stable semantics and completion with loop
formulas have been extended to such programs (see [Erdem and Lifschitz, 2003;
Lee and Lifschitz, 2003]).
[Bochman, 1996a] has suggested a general scheme for constructing semantics
of logic programs of a most general kind in the logical framework of biconsequence relations that we will describe later in this study. A detailed description
was given in [Bochman, 1998a; 1998b]. Briefly, the formalism involved rules
(bisequents) of the form a : b c : d that directly represented general program
rules not d, c ← a, not b. The notion of circumscription of a biconsequence
relation was defined, and then a nonmonotonic completion was constructed as
a closure of the circumscription with respect to certain coherence rules. The
strength of these rules depended on the language L in which they were formulated. Finally the nonmonotonic semantics of a biconsequence relation was
defined as a pair of sets of propositions in the language L that are, respectively,
provable and refutable in the nonmonotonic completion. A uniform representation of major existing semantics for logic programs was provided simply by
varying the language L of the representation. A related, though more ‘logical’, representation of general logic programs and their semantics has been
suggested in [Bochman, 2004b] in the framework of causal inference relations
from [Bochman, 2004a] that will be described later.
3.5
Argumentation theory
Since Doyle’s seminal work on truth maintenance [Doyle, 1979], the importance
of argumentation in choosing default assumptions has been a recurrent theme
in the literature on nonmonotonic reasoning (see, e.g., [Lin and Shoham, 1989;
Geffner, 1992]). In a parallel development, a number of researchers in traditional
argumentation theory have shown that, despite influential convictions and prejudices, the ordinary, human argumentation is within the reach of formal logical
methods (see [Chesñevar et al., 2000] for a survey of this development).
A powerful version of abstract argumentation theory, especially suitable for
nonmonotonic reasoning, has been suggested in [Dung, 1995b]. We will give
below a brief description of this theory.
Definition. An abstract argumentation theory is a pair hA, ,→i, where A is a
set of arguments, while ,→ a binary relation of an attack on A.
If α ,→ β holds, then the argument α attacks, or undermines, the argument β.
A general task of argumentation theory consists in determining sets of arguments
that are safe (justified) in some sense with respect to the attack relation. To
this end, we should extend first the attack relation to sets of arguments: if Γ, ∆
are sets of arguments, then Γ ,→ ∆ is defined to hold if α ,→ β, for some α ∈ Γ
and β ∈ ∆.
Let us say that an argument α is allowable for the set of arguments Γ, if Γ
does not attack α. For any set of arguments Γ, we will denote by [Γ] the set of
28
all arguments allowable by Γ, that is
[Γ] = {α | Γ 6,→ α}
An argument α will be said to be acceptable for the set of arguments Γ, if Γ
attacks any argument against α. As can be easily checked, the set of arguments
that are acceptable for Γ coincides with [[Γ]].
Using the above notions, we can give a quite simple characterization of the
basic objects of an abstract argumentation theory.
Definition. A set of arguments Γ will be called
• conflict-free if Γ ⊆ [Γ];
• admissible if it is conflict-free and Γ ⊆ [[Γ]];
• a complete extension if it is conflict-free and Γ = [[Γ]];
• a preferred extension if it is a maximal complete extension;
• a stable extension if Γ = [Γ].
A set of arguments Γ is conflict-free if it does not attack itself. A conflictfree set Γ is admissible if and only if any argument from Γ is also acceptable for
Γ, and it is a complete extension if it coincides with the set of arguments that
are acceptable with respect to it. Finally, a stable extension is a conflict-free
set of arguments that attacks any argument outside it. Any stable extension
is also a preferred extension, any preferred extension is a complete extension,
and any complete extension is an admissible set. Moreover, as has been shown
in [Dung, 1995b], any admissible set is included in some complete extension.
Consequently, preferred extensions coincide with maximal admissible sets. In
addition, the set of complete extensions forms a complete lower semi-lattice: for
any set of complete extensions, there exists a unique greatest complete extension
that is included in all of them. In particular, there always exists a least complete
extension of an argumentation theory.
As has been shown in [Dung, 1995a] the above objects exactly correspond
to the semantics suggested for normal logic programs. Thus, stable extensions
correspond to stable models, complete extensions correspond to partial stable
models, preferred extensions correspond to regular models, while the least complete extension corresponds in this sense to the well-founded semantics. These
results have shown, in effect, that the abstract argumentation theory successfully captures the essence of logical reasoning behind normal logic programs (see
also [Kakas and Toni, 1999]).
The notion of an argument is often taken as primitive in argumentation theory, which even allows for a possibility of considering arguments that are not
propositional in character (e.g., arguments as inference rules, or derivations). As
has been shown in [Bondarenko et al., 1997], however, a powerful instantiation
of argumentation theory can be obtained by identifying arguments with propositions of a special kind called assumptions (see also [Kowalski and Toni, 1996]).
29
Thus, for logic programs assumptions can be identified with negation-as-failure
literals of the form not A, while consistency-based justifications of default rules
(see above) can serve as assumptions in default logic.
Slightly simplifying definitions from [Bondarenko et al., 1997], an assumptionbased argumentation framework can be defined as a triple consisting of an underlying deductive system, a distinguished subset of propositions Ab called assumptions, and a mapping from Ab to the set of all propositions of the language
that determines the contrary α of any assumption α. For instance, the contrary of the negation-as-failure literal not A in logic programming is the atom
A itself, while ¬A is the contrary of the justification A in default logic. A set
of assumptions Γ attacks an assumption α, if it implies its contrary, α. It has
been shown that the main nonmonotonic formalisms such as default and modal
nonmonotonic logics, as well as semantics of normal logic programs are representable in this framework. A dialectical proof procedure for finding admissible
sets of assumptions has recently been described in [Dung et al., 2006].
In [Bochman, 2003b] an extension of an abstract argumentation framework
was introduced in which the attack relation is defined directly among sets of
arguments. The extension, called collective argumentation, has turned out to
be suitable for representing semantics of disjunctive logic programs. In collective argumentation ,→ is an attack relation on sets of arguments satisfying the
following monotonicity condition:
(Monotonicity)
If Γ ,→ ∆, then Γ ∪ Γ0 ,→ ∆ ∪ ∆0 .
Dung’s argumentation theory can be identified with a normal collective argumentation theory in which no set of arguments attacks the empty set ∅ and
the following condition is satisfied:
(Locality)
If Γ ,→ ∆, ∆0 , then either Γ ,→ ∆ or Γ ,→ ∆0 .
In a normal collective argumentation theory the attack relation is reducible
to the relation Γ ,→ α between sets of arguments and single arguments, and the
resulting theory coincides with that given in [Dung, 1995a].
The attack relation of an argumentation theory can be given a natural fourvalued semantics based on independent evaluations of acceptance and rejection
of arguments. On this interpretation the attack Γ ,→ ∆ means that at least one
of the arguments in ∆ should be rejected whenever all the arguments from Γ
are accepted. The expressive capabilities of the argumentation theory depend,
however, on the absence of usual ‘classical’ constraints on the acceptance and
rejection of arguments, so it permits situations in which an argument is both
accepted and rejected, or, alternatively, neither accepted, nor rejected. Such
an understanding can be captured formally by assigning to any argument an
arbitrary subset of the set {t, f }, where t denotes acceptance (truth), while f
denotes rejection (falsity) (cf. [Jakobovits and Vermeir, 1999]). This interpretation is nothing other than the well-known Belnap’s interpretation of four-valued
logic (see [Belnap, 1977]).
30
The language of arguments can be extended with a global negation connective
∼ having the following semantic interpretation:
∼A is accepted iff A is rejected
∼A is rejected iff A is accepted.
An axiomatization of this negation in argumentation theory can be obtained
by imposing the following rules on the attack relation (see [Bochman, 2003b]):
A ,→ ∼A
∼A ,→ A
If a ,→ A, b and a, ∼A ,→ b, then a ,→ b
(AN)
If a, A ,→ b and a ,→ b, ∼A, then a ,→ b
It turns out that the resulting N-attack relations are interdefinable with
certain consequence relations.
A Belnap consequence relation in a propositional language with a global
negation ∼ is a Scott (multiple-conclusion) consequence relation satisfying
the postulates
(Reflexivity)
A A;
(Monotonicity) If a b and a ⊆ a0 , b ⊆ b0 , then a0 b0 ;
(Cut) If a b, A and a, A b, then a b,
as well as the following two Double Negation rules for ∼:
A ∼∼A
∼∼A A.
For a set u of propositions, ∼u will denote the set {∼A | A ∈ u}. Now, for
a given N-attack relation, we can define the following consequence relation:
a b ≡ a ,→ ∼b
(CA)
Similarly, for any Belnap consequence relation we can define the corresponding attack relation as follows:
a ,→ b ≡ a ∼b
(AC)
As has been shown in [Bochman, 2003b], the above definitions establish an
exact equivalence between N-attack relations and Belnap consequence relations.
In this setting, the global negation ∼ serves as a faithful logical formalization of
the operation of taking the contrary from [Bondarenko et al., 1997]. Moreover,
given an arbitrary language L that does not contain ∼, we can define assumptions as propositions of the form ∼A, where A ∈ L. Then, since ∼ satisfies
double negation, a negation of an assumption will be a proposition from L.
Such a ‘negative’ representation of assumptions will agree with the applications
of the argumentation theory to other nonmonotonic formalisms described in
[Bondarenko et al., 1997].
31
3.6
Abduction and causal reasoning
It does not seem necessary to argue that abduction and causation are essential
for the human, commonsense reasoning about the world. It has been gradually
realized, however, that these kinds of reasoning are essential also for efficient
‘artificial’ reasoning. Moreover, a new outlook on these kinds of reasoning can
be achieved by viewing them as special, though important, instances of nonmonotonic reasoning.
3.6.1
Abductive reasoning
Abduction is the process of finding explanations for observations. The importance of abduction for AI can be seen already in the Minsky’s frame paper
[Minsky, 1974]. A frame that contains default and stereotypical information
should be imposed on a particular situation:
Once a frame is proposed to represent a situation, a matching process
tries to assign values to each frame’s terminals, consistent with the
markers at each place. The matching process is partly controlled by
information associated with the frame (which includes information
about how to deal with surprises) and partly by knowledge about
the system’s current goals.
[Israel, 1980] criticized first nonmonotonic formalisms by objecting to the
centrality of deductive logic in these formalisms as a mechanism for justification.
He argued that what we need is not a new logic, but a good scientific methodology. Abductive reasoning to a best explanation requires rational epistemic
policies that lie, on Israel’s view, outside nonmonotonic logics. [McDermott,
1987] levied a similar criticism about the absence of a firm theoretical basis
behind diagnostic and other programs dealing with abduction:
This state of affairs does not stop us from writing medical diagnosis
programs. But it does keep us from understanding them. There is
no independent theory to appeal to that can justify the inferences a
program makes .... these programs embody tacit theories of abduction; these theories would be the first nontrivial formal theories of
abduction, if only one could make them explicit.
Despite this criticism, however, formal properties of abductive inference
methods and relations to other formalisms have been actively explored. [de
Kleer, 1986] has developed the Assumption-based Truth Maintenance Systems
(ATMS), a method for finding explanations in the context of a propositional
Horn-clause theories, when the hypotheses and observations are positive atoms.
Similarly to Doyle’s JTMS [Doyle, 1979], described earlier, the ATMS was developed by de Kleer as a subcomponent of a more general problem-solving system.
The main inference mechanism in the ATMS is the computation of a label at
each node. A label for an atom c is a set of environments, that is, sets of hypotheses that explain c. The ATMS differed from Doyle’s TMS, however, in
32
keeping track of multiple explanations or contexts, and especially in using only
monotonic propositional inference.
More general systems of logical abduction used more general theories. The
most prominent of these is the Clause Maintenance System (CMS) of [Reiter and
de Kleer, 1987]. To generalize the ATMS definition of explanation, the authors
considered a propositional domain theory consisting of general clauses and drew
on the concept of prime implicates to extend the inference technique of the
ATMS. Prime implicates can be used to find all the parsimonious explanations
of unit clauses. It was shown in [Reiter and de Kleer, 1987] that the label in
ATMS is exactly the set of such explanations, so the ATMS can be used to
compute parsimonious explanations for propositional Horn-clause theories.
Under diagnosis from first principles [Reiter, 1987b], or diagnosis from structure and behavior, the only information at hand is a description of some system,
say a physical device, together with an observation of that system’s behavior.
If this observation conflicts with intended system behavior, then the diagnostic
problem is to determine which components could by malfunctioning account
for the discrepancy. Since components can fail in various and often unpredictable ways, their normal or default behaviors should be described. These
descriptions fit the pattern of nonmonotonic reasoning. For example, an AND
gate in a digital circuit would have the description: Normally, an AND-gate’s
output is the Boolean AND function of its inputs. In diagnosis, such component descriptions are used in the following way: We first assume that all of
the system components are behaving normally. Suppose, however, the system
behavior predicted by this assumption conflicts with (i.e. is inconsistent with)
the observed system behavior. Thus some of the components we assume to be
behaving normally must really be malfunctioning. By retracting enough of the
original assumptions about correctly behaving components, we can remove the
inconsistency between the predicted and observed behavior. The retracted components yield a diagnosis. This approach to diagnosis from first principles was
called a consistency-based diagnosis, and it forms the basis for several diagnostic
reasoning systems (see also [de Kleer et al., 1992]).
From a logical point of view, the consistency-based diagnosis reduces to
classical consistency reasoning with respect to a certain abnormality theory, in
which abnormalities are minimized as in circumscription. This kind of diagnosis
is still not fully abductive, since it determines only what is a minimal set of
abnormalities that is consistent with the observed behavior. In other words, it
does not explain observations, but only excuses them.
In medical diagnosis, the type of reasoning involved is abductive in nature
and consists in explaining observations or the symptoms of a patient. Moreover,
as in general nonmonotonic reasoning, adding more information can cause a
previously held conclusion to become invalid. If it is known only that a patient
has a fever, the most reasonable explanation is that he has the flu. But if we
learn that he also has jaundice, then it becomes more likely that he has a disease
of the liver. The first, ‘procedural’ formalization of this reasoning was the setcovering model, proposed in [Reggia et al., 1985]. This model had a set of causes
and a set of symptoms, along with a relation that maps a cause to the set of
33
symptoms that it induces. Given an observation of symptoms for a particular
case, a diagnosis is a set of causes that covers all of the observed symptoms and
contains no irrelevant causes.
On the face of it, consistency-based and abductive diagnosis appear very
different. To begin with, rather than abducing causes that imply the observations, the consistency approach tries to minimize the extent of the causation
set by denying as many of its elements as possible. Moreover, in the abductive
framework, the causes have implications for the effects, while in the consistency
based systems, the most important information seems to be the implication of
the observations for possible causes.
Despite these differences, it is now known that, under certain conditions,
consistency-based explanations in the Clark completion of the domain theory
coincide with abductive explanations in the source theory.
Theorem. ([Poole, 1988b; Console et al., 1991]) Let Σ be a set of nonatomic
definite clauses whose directed graph of dependencies is acyclic, and let Π be
the Clark completion of Σ. Then the consistency-based explanations of an
observation O in Π are exactly the abductive explanations of O in Σ.
For more complicated domain theories, Clark completion does not give the
required closure to abductive explanations. For such more general cases the
correct generalization of Clark completion is explanatory closure (see [Konolige,
1992]).
Classical abduction has also been used as a practical proof method for circumscription (see, e.g., [Ginsberg, 1989]). [Levesque, 1989] suggested a knowledge level analysis of abduction in which the domain theory is represented as
the beliefs of an agent. Motivated primarily by the intractability of logic-based
abduction, this representation allowed for incomplete deduction in connecting
hypotheses to observations for which tractable abduction mechanisms can be
developed.
[Poole, 1988a] has developed the Theorist system in which abduction is used
as an inference method in default theories. In the spirit of David Israel, Poole
argued that there is nothing wrong with classical logic; instead, nonmonotonicity is a problem of how the logic is used. Theorist assumed the typical abductive
machinery: a set of hypotheses, a first-order background theory, and the concept
of explanation. A scenario is any subset of the hypotheses consistent with the
background theory; in the language of Theorist, an explanation for an observation O is a scenario that implies O. Theorist defined the notion of extension as a
set of propositions generated by a maximal consistent scenario. In other words,
by taking defaults as possible hypotheses, default reasoning has been reduced
to a process of theory formation.
Defaults of Poole’s abductive system corresponded to a simplest kind of Reiter’s default rules, namely normal defaults of the form : A/A. In addition, Poole
employed the mechanism of naming defaults (closely related to McCarthy’s abnormality predicates) that has allowed him to say, in particular, when a default
is inapplicable. The resulting system has been shown to capture many of the
representative capabilities of Reiter’s default logic in an almost classical logical
framework.
34
Systems such as the ATMS and Theorist are popular in many AI applications, because they are easy to understand and relatively easy to implement.
The use of abductive methods is growing within AI, and they are now a standard part of most AI representation and reasoning systems. In fact, the study
of abduction is one of the success stories of nonmonotonic reasoning, and it has
a major impact on the development of an application area.
Abductive Logic Programming. One of the important applications of abduction has been developed in the framework of logic programming. A comprehensive survey of the extension of logic programming to perform abductive
reasoning (referred to as abductive logic programming) can be found in [Kakas
et al., 1992] together with an extensive bibliography on abductive reasoning. As
their primary goal, the authors introduced an argumentation theoretic approach
to the use of abduction as an interpretation of negation-as-failure. Abduction
was shown to generalize negation-as-failure to include not only negative but also
positive hypotheses, and to include general integrity constraints. They showed
that abductive logic programming is related to the justification-based truth
maintenance system of Doyle and the assumption-based truth maintenance system of de Kleer.
Abductive logic programs are defined as pairs (Π, A), where Π is a logic program, and A a set abducible atoms. A formalization of abductive reasoning in
this setting is provided by the generalized stable semantics [Kakas and Mancarella, 1990], in which an abductive explanation of a query q is a subset S
of abducibles such that there exists a stable model of the program Π ∪ S that
satisfies q.
[Denecker and Schreye, 1992] developed a family of extensions of SLDNF
resolution for normal abductive programs. Further details and directions in
abductive logic programming can be found, e.g., in [Brewka and Konolige, 1993;
You et al., 2000; Lin and You, 2002].
It has been shown in [Inoue and Sakama, 1998] that abductive logic programs
under the generalized stable semantics are reducible to general disjunctive logic
programs under the stable semantics. The relevant transformation of abductive
programs can be obtained simply by adding to Π the program rules p, not p ←,
for any abducible atom p from A. This reduction has confirmed, in effect,
that general logic programs are inherently abductive and hence have the same
representation capabilities as abductive logic programs.
An abstract abductive framework. To end this section on abduction, we
will provide below a brief description of an abstract abductive system that is
sufficient for the main applications of abduction in AI. Further details on this
system and its expressive capabilities can be found in [Bochman, 2005].
An abductive system is a pair A = (Cn, A), where Cn is a supraclassical consequence relation, while A a distinguished set of propositions called abducibles.
A set of abducibles a ⊆ A is an explanation of a proposition A, if A ∈ Cn(a).
In applications, the consequence relation Cn is usually given indirectly by
35
a generating conditional theory ∆, in which case the corresponding abductive
system can be defined as (Cn∆ , A). Many abductive frameworks also impose
syntactic restrictions on the set of abducibles A (Poole’s Theorist being a notable
exception). Thus, A is often restricted to a set of special atoms (e.g., those
built from abnormality predicates ab), or to the corresponding set of literals.
The restriction of this kind is not essential, however. Indeed, for any abducible
proposition A we can introduce a new abducible propositional atom pA , and add
the equivalence A ↔ pA to the underlying theory. The new abductive system
will have much the same properties.
An abductive system (Cn, A) will be called classical if Cn is a classical
consequence relation. A classical abductive system can be safely equated with
a pair (Σ, A), where Σ is a set of classical propositions (the domain theory). An
example of such a system in diagnosis is [de Kleer et al., 1992], a descendant of
the consistency-based approach of [Reiter, 1987b].
In abductive systems, acceptance of propositions depends on existence of explanations, and consequently such systems sanction not only forward inferences
determined by the consequence relation, but also backward inferences from facts
to their explanations, and combinations of both. All these kinds of inference
can be captured formally by considering only theories of Cn that are generated
by the abducibles. This suggests the following notion:
Definition. The abductive semantics SA of an abductive system A is the set of
theories {Cn(a) | a ⊆ A}.
By restricting the set of theories to theories generated by abducibles, we
obtain a semantic framework containing more information. Generally speaking,
all the information that can be discerned from the abductive semantics of an
abductive system can be seen as abductively implied by the latter.
The information embodied in the abductive semantics can be made explicit
by using the associated Scott (multiple-conclusion) consequence relation, defined
as follows6 : for any sets b, c of propositions,
b `A c ≡ (∀a ⊆ A)(b ⊆ Cn(a) → c ∩ Cn(a) 6= ∅)
This consequence relation describes not only forward explanatory relations,
but also abductive inferences from propositions to their explanations. Speaking
generally, it describes the explanatory closure, or completion, of an abductive
system, and thereby captures abduction by deduction (cf. [Console et al., 1991;
Konolige, 1992]).
Example. The following abductive system describes a variant of the well-known
Pearl’s example. Assume that an abductive system A is determined by the set
∆ of rules
Rained ` Grasswet
Sprinkler ` Grasswet
Rained ` Streetwet,
6 A Tarski consequence relation of this kind has been used for the same purposes in [Lobo
and Uzcátegui, 1997].
36
and the set of abducibles Rained, ¬Rained, Sprinkler, ¬Sprinkler, ¬Grassswet.
Since Rained and ¬Rained are abducibles, Rained is an independent (exogenous) parameter, and similarly for Sprinkler. However, since only ¬ Grassswet
is an abducible, non-wet grass does not require explanation, but wet grass does.
Thus, any theory of SA that contains Grasswet should contain either Rained,
or Sprinkler, and consequently we have
Grasswet `A Rained, Sprinkler.
Similarly, Streetwet implies in this sense both its only explanation Rained
and a collateral effect Grasswet.
3.6.2
Causal reasoning
The last example of the preceding section also illustrates that one of the central
applications of abductive inference consists in generating causal explanations:
reasons connecting causes and their effects. In this case, the hypotheses, or
abducibles, represent primitive causes, observations are about their effects, and
the background theory encodes the relation between them. This is just one of the
many ways in which the notion of causation that once has been expelled from
exact sciences (see [Russell, 1957]) reappears as an important representation
tool in Artificial Intelligence. Moreover, the theories of causality emerging in AI
are beginning to illuminate in turn the actual role of causality in our reasoning.
As a mater of fact, causal considerations play an essential role in abduction
in general. They determine, in particular, the very choice of abducibles, as well
as the right form of descriptions and constraints (even in classical first-order
representations). As has been shown already in [Darwiche and Pearl, 1994],
system descriptions that do not respect the natural causal order of things can
produce inadequate predictions and explanations.
The intimate connection between causation and abduction has become especially vivid in the abductive approach to diagnosis (see especially [Cox and
Pietrzykowski, 1987; Poole, 1994b; Konolige, 1994]). As has been acknowledged
in these studies, reasoning about causes and effects should constitute a logical
basis for diagnostic reasoning. Unfortunately, the absence of an adequate logical formalization for causal reasoning has relegated the latter to the role of an
informal heuristic background, with classical logic serving as the representation
language.
Abduction and diagnosis are not the only areas of AI in which causality has
emerged. Thus, causality is an essential part of general qualitative reasoning;
see, e.g., [Iwasaki and Simon, 1986; Nayak, 1994], which goes back to Herbert
Simon’s important work [Simon, 1952]. Judea Pearl and his students and associates have developed a profound research program in the study of causation
derived from the use of causal diagrams in reasoning about probabilities known
as Bayesian Belief Networks. A detailed description of the emerged theory of
causality and its applications can be found in [Pearl, 2000]. See also [Halpern,
1998; Halpern and Pearl, 2001a; 2001b; Eiter and Lukasiewicz, 2004].
37
But perhaps the most significant development concerning the role of causation in our reasoning has occurred in of one of the central fields of logical AI,
reasoning about action and change. The starting point in this development was
the discovery of the Yale Shooting Problem [Hanks and McDermott, 1987].
We are told that a gun is loaded at time 1, and that the gun is fired at Fred
at time 5. Loading a gun causes the gun to be loaded, and firing a loaded gun
at an individual causes the person to be dead. In addition, the fluents alive and
loaded persist as long as possible; that is, these fluents stay true unless an action
occurs that is abnormal with respect to these fluents. Thus, a person who is
alive tends to remain alive, and a gun that is loaded tends to remain loaded.
What can we conclude about Fred’s status at time 6? Although common sense
argues that Fred is dead at time 6, existed nonmonotonic representations of this
story supported two models. In one model (the expected model), the fluent
loaded persists as long as possible. Therefore, the gun remains loaded until it is
fired at Fred, and Fred dies. In this model, at time 5, shooting is abnormal with
respect to Fred’s being alive. In the other, unexpected model, the fluent alive
persists as long as possible (i.e., Fred is alive after the shooting). Therefore,
the fluent loaded did not persist; somehow the gun has become unloaded. That
is, in some situation between 2 and 5, the empty action Wait was abnormal
with respect to the gun being loaded. This existence of multiple extensions has
created a genuine problem. Hanks and McDermott in fact argued that the Yale
Shooting Problem underscored the inadequacy of the whole logicist temporal
reasoning.
Early solutions to the Yale Shooting problem has been based on chronological
minimization, but they quickly lost favor mainly due to their inability to handle
backward temporal reasoning. As was rightly pointed out in [McDermott, 1987],
the main source of the problem was that the previous nonmonotonic logics drew
conclusions that minimized disturbances instead of avoiding disturbances with
unknown causes. And the emerged alternative approaches has begun to employ
various forms of causal reasoning.
[Lifschitz, 1987b] was the beginning of a sustained line of research in the
causal approach. The causal-based approach argued that we expect Fred to die
because there is an action that causes Fred’s death, but there is no action that
causes the gun to become unloaded. All the causal representations formalize this
principle in some way. Lifschitz has provided also a solution to the qualification
problem in this framework. Several modifications of Lifschitz’s solution have
been suggested (see, e.g., [Baker, 1989]).
A broadly explanatory approach to nonmonotonic reasoning was pursued
by Hector Geffner in [Geffner, 1992]. In the last chapters of his book, Geffner
argued for the necessity of incorporating causation as part of the meaning of
default conditionals. He used a unary modal operator Cp meaning ‘p is caused’
and expressed causal claims as conditionals of the form p → Cq. Using this
language, he formalized a simple causal solution to the Yale Shooting Problem.
In [Geffner, 1992], however, the causal theory was only sketched; in particular
the Ramification Problem was left untouched.
One of the first causal approaches, Motivated Action Theory [Morgenstern
38
and Stein, 1994], was based on the idea of preferring models in which unexpected
actions do not happen. A description of a problem scenario in MAT consisted of
a theory and a partial chronicle description. The theory contained causal rules
and persistence rules. Causal rules described how actions change the world;
persistence rules described how fluents remain the same over time. Central to
MAT was the concept of motivation. Intuitively, an action is motivated if there
is a “reason” for it to happen. The effects of causal chains are motivated in
this sense. A model is preferred if it has as few unmotivated actions as possible.
This kind of reasoning was clearly nonmonotonic, and it allowed for a solution
to the Yale Shooting Problem. In the expected model, where the gun remains
loaded and Fred dies, there are no unmotivated actions. In the unexpected
model, there is an unmotivated action – the Unload action. Thus we prefer the
expected models over the unexpected models. However, MAT falled short as
a solution to the frame problem due to its persistence rules – which are just
another form of frame axioms.
A first systematic causal representation of reasoning about action and change
has been developed by Fangzhen Lin in [Lin, 1995; 1996]. The representation
has allowed, in particular, to handle natural causes in addition to actions, and
provided a natural solution to the ramification problem. Lin’s representation
was formulated in a purely classical first-order language, but employed a (reified)
distinction between facts that hold in a situation versus facts that are caused in
it. The adequate models have been obtained by minimizing (i.e., circumscribing)
the caused facts.
A formalization of this causal reasoning in action theories has been suggested
in [McCain and Turner, 1997] in the framework of what they called causal theories. A causal theory is a set of causal rules that express causal relations among
propositions. The inertia (or persistence) principle is also expressed as a kind
of a causal rule. Then a true proposition can be caused either because it is
the direct or indirect effect of an action, or because it involves the persistence
of a caused proposition. Initial conditions are also considered to be caused,
by stipulation. Finally, the nonmonotonic semantics of a causal theory is determined by causally explained models, namely the models that both satisfy
the causal rules and such that every fact holding in them is caused by some
causal rule. In other words, in causally explained models the caused propositions coincide with the propositions that are true, and this must be the only
possibility consistent with the extensional part of the model. The resulting
nonmonotonic formalism has been shown to provide a plausible and efficient
solution for both the frame and ramification problem. See [Lifschitz, 1997;
Turner, 1999; Giunchiglia et al., 2004] for a detailed exposition of this theory
and applications in representing action domains. Related causal approaches
to representing actions and change have been suggested in [Thielscher, 1997;
Schwind, 1999; Zhang and Foo, 2001], to mention only a few.
The logical foundations of causal reasoning of this kind have been formulated
in [Bochman, 2003c; 2004a] in the framework of an inference system for causal
rules originated in input/output logics of [Makinson and van der Torre, 2000].
Formally, a causal inference relation is a binary relation ⇒ on the set of classical
39
propositions satisfying the following postulates:
(Strengthening)
(Weakening)
If A ⇒ B and B C, then A ⇒ C;
If A ⇒ B and A ⇒ C, then A ⇒ B ∧ C;
(And)
(Or)
If A B and B ⇒ C, then A ⇒ C;
If A ⇒ C and B ⇒ C, then A ∨ B ⇒ C.
(Cut)
(Truth)
(Falsity)
If A ⇒ B and A ∧ B ⇒ C, then A ⇒ C;
t ⇒ t;
f ⇒ f.
From a logical point of view, the most significant ‘omission’ of the above set
is the absence of the reflexivity postulate A ⇒ A. It is precisely this feature of
causal inference that creates a possibility of nonmonotonic reasoning. Causal
inference relations can be given a standard possible worlds semantics. Namely,
given a relational possible worlds model (W, R, V ), where W is a set of possible
worlds, R a binary accessibility relation on W , and V a valuation function, the
validity of causal rules can be defined as follows:
Definition. A rule A ⇒ B is valid in a possible worlds model (W, R, V ) if, for
any α, β ∈ W such that αRβ, if A holds in α, then B holds in β.
Causal inference relations are determined by possible worlds models in which
the relation R is quasi-reflexive, that is, αRβ holds only if αRα.
Causal rules are extended to rules with sets of propositions
V in premises by
stipulating that, for a set u of propositions, u ⇒ A holds if a ⇒ A for some
finite a ⊆ u. C(u) denotes the set of propositions explained by u:
C(u) = {A | u ⇒ A}
The production operator C plays the same role as the usual derivability
operator for consequence relations. In particular, it is a monotonic operator,
that is, u ⊆ v implies C(u) ⊆ C(v). Still, it does not satisfy inclusion, that is,
u ⊆ C(u) does not in general hold.
InVcausalWinference relations, any causal rule is reducible to a set of clausal
rules li ⇒ lj , where li , lj are classical literals. In addition, any rule A ⇒ B
is equivalent to a pair of rules A ∧ ¬B ⇒ f and A ∧ B ⇒ B.
The rules A ∧ B ⇒ B are explanatory rules. Though logically trivial, they
play an important explanatory role in causal reasoning by saying that, if A
holds, B is self-explanatory (and hence does not require explanation). Such
rules has been used in [McCain and Turner, 1997] for representing inertia or
persistence claims (see an example at the end of this section). On the other
hand, the rule A ∧ ¬B ⇒ f is a constraint that does not have an explanatory
content, but imposes a factual restriction A→B on the set of interpretations.
Causal inference relations determine also a natural nonmonotonic semantics, and provide thereby a logical basis for a particular form of nonmonotonic
reasoning.
40
Definition. A nonmonotonic semantics of a causal inference relation is the set
of all its exact worlds — maximal deductively closed sets u of propositions such
that u = C(u).
Exact worlds are worlds that are fixed points of the production operator C.
An exact world describes a model that is closed with respect to the causal rules
and such that every proposition in it is caused by other propositions accepted
in the model. Accordingly, they embody an explanatory closure assumption,
according to which any accepted proposition should also have explanation for
its acceptance. Such an assumption is nothing other than the venerable principle
of Universal Causation (cf. [Turner, 1999]).
The nonmonotonic semantics for causal theories is indeed nonmonotonic in
the sense that adding new rules to the causal relation may lead to a nonmonotonic change of the associated semantics, and thereby of derived information.
This happens even though causal rules themselves are monotonic, since they
satisfy the postulate of Strengthening (the Antecedent).
The nonmonotonic semantics of causal theories coincides with the semantics suggested in [McCain and Turner, 1997]. Moreover, it has been shown
in [Bochman, 2003c] that causal inference relations constitute a maximal logic
adequate for this kind of nonmonotonic semantics.
Causal inference relations and their variations have turned out to provide a
new general-purpose formalism for nonmonotonic reasoning with representation
capabilities stretching far beyond reasoning about action. In particular, it has
been shown in [Bochman, 2004a] that they provide a natural logical representation of abduction, based on treating abducibles as self-explanatory propositions satisfying reflexivity A ⇒ A. In addition, it has been shown in [Bochman,
2004b] that causal reasoning provides a precise interpretation for general logic
programs. Namely, any program rule c, not d ← a, not b can be translated as a
causal rule d, ¬b ⇒ ∧a →∨c. Then the reasoning underlying the stable model semantics for logic programs (see above) is captured by augmenting the resulting
causal theory with the causal version of the Closed World Assumption stating
that all negated atoms are self-explanatory:
Default Negation ¬p ⇒ ¬p, for any propositional atom p.
The causal nonmonotonic semantics of the resulting causal theory will correspond precisely to the stable semantics of the source logic program. Moreover,
unlike known embedding of logic programs into other nonmonotonic formalisms,
namely default and autoepistemic logics, the causal interpretation of logic programs turns out to be bi-directional in the sense that any causal theory is
reducible to a general logic program.
As can be shown, a world α is an exact world of a causal inference relation
if and only if, for any propositional atom p,
p ∈ α if and only if α ⇒ p
and
¬p ∈ α if and only if α ⇒ ¬p.
41
By the above description, the exact worlds of a causal relation are determined
ultimately by rules of the form A ⇒ l, where l is a literal. Such rules are called
determinate. McCain and Turner have established an important connection
between the nonmonotonic semantics of a determinate causal theory and Clark’s
completion of the latter. A finite causal theory ∆ is definite, if it consists of
determinate rules, or rules A ⇒ f , where f is a falsity constant. A completion
of such a theory is the set of all classical formulas
_
_
p ↔ {A | A ⇒ p ∈ ∆}
¬p ↔ {A | A ⇒ ¬p ∈ ∆}
for any propositional atom p, plus the set {¬A | A ⇒ f ∈ ∆}. Then the classical models of the completion precisely correspond to exact worlds of ∆ (see
[Giunchiglia et al., 2004]).
The completion formulas embody two kinds of information. As (forward)
implications from right to left, they contain the material implications corresponding to the causal rules from ∆. In addition, left-to-right implications
state that a literal belongs to the model only if one of its causes is also in
the model. These implications reflect the impact of causal descriptions using
classical logical means. In other words, the completion of a causal theory is a
classical logical theory that embodies the required causal content, so it obliterates in a sense the need in causal representation. In fact, a classical theory
of actions and change based directly on such a completion has been suggested
by Ray Reiter as a simple solution to the frame problem — see [Reiter, 1991;
2001]. This solution can now be reformulated as follows.
Reiter’s simple solution. The following toy representation contains the
main ingredients of causal reasoning in temporal domains. The temporal behavior of a propositional fluent F is described using two propositional atoms F0
and F1 saying, respectively, that F holds now and F holds in the next moment.
C − ⇒ ¬F1
C + ⇒ F1
F0 ∧ F1 ⇒ F1
F0 ⇒ F0
¬F0 ∧ ¬F1 ⇒ ¬F1
¬F0 ⇒ ¬F0 .
The first pair of causal rules describes the actions or natural factors that
can cause F and, respectively, ¬F (C + and C − normally describe the present
situation). Second, we have a pair of inertia axioms, purely explanatory rules
stating that if F holds (does not hold) now, then it is self-explanatory that it
will hold (resp., not hold) in the next moment. The last pair of initial axioms
states that F0 is an exogenous parameter.
The above causal theory is determinate, and its completion is as follows:
F1 ↔ C + ∨ (F0 ∧ F1 )
¬F1 ↔ C − ∨ (¬F0 ∧ ¬F1 ).
These formulas are equivalent to the conjunction of ¬(C + ∧ C − ) and
F1 ↔ C + ∨ (F0 ∧ ¬C − ).
42
The above formulas provide an abstract description of Reiter’s simple solution: the first formula corresponds to his consistency condition, while the last
one - to the successor state axiom for F .
4
Preferential Nonmonotonic Reasoning
As a theory of a rational use of assumptions, the main problem nonmonotonic
reasoning deals with is that assumptions are often incompatible with one another, or with known facts. In such cases of conflict we must have a reasoned
choice. The preferential approach follows here the slogan “Choice presupposes
preference”. According to this approach, the choice of assumptions should be
made by forming the space of options for choice and establishing preference relations among them. This makes preferential approach a special case of a general
methodology that is at least as old as the decision theory and theory of social
choice.
McCarthy’s circumscription can be seen as the ultimate origin of the preferential approach. A generalization of this approach was initiated by Gabbay
in [Gabbay, 1985] on the logical side, and by Shoham [Shoham, 1988] on the AI
side. A first overview of the preferential approach has been given in [Makinson,
1994]. A detailed description of the approach as we see it now can be found in
[Bochman, 2001; Schlechta, 2004; Makinson, 2005]; see also the paper of Karl
Schlechta in this volume.
Both the Closed World Assumption and circumscription can be seen as working on the principle of preferring interpretations in which positive facts are minimized; this idea was pursued already in [Lifschitz, 1985]. Generalizing this idea,
[Shoham, 1988] argued that any form of nonmonotonicity necessarily involves
minimality of one kind or another. He argued also for a shift in emphasis from
syntactic characterizations in favor of semantic ones. Namely, the relevant nonmonotonic entailment should be defined in terms of truth in all those models
of a given axiomatization minimal with respect to some application dependent
criterion. The ability to characterize such minimality criteria axiomatically is
not essential. In effect, on Shoham’s view, an axiomatization of an application
domain coupled with a characterization of its preferred minimal models is a
sufficient specification of the required entailments.
Shoham defined a model preference logic by using an arbitrary preference
ordering of the interpretations of a language.
Definition. An interpretation i is a preferred model of A if it satisfies A and
there is no better interpretation j > i satisfying A. A preferentially entails B
(written A |∼ B) iff all preferred models of A satisfy B.
In support of his conclusions, Shoham offered his own theory of temporal
minimization, as well as a minimal model semantics for a certain simplification
of Reiter’s default logic.
Shoham’s approach was very appealing, and suggested a unifying perspective
on nonmonotonic reasoning. This treatment of nonmonotonicity was also similar
43
to the earlier modal semantic theories of conditionals and counterfactuals, which
have been studied in the philosophical literature — see, e.g., [Stalnaker, 1968;
Lewis, 1973]. Just as the nonmonotonic entailment, counterfactual conditionals
do not satisfy monotonicity, that is, A > C does not imply A ∧ B > C. The
interrelations between these two theories have become an important theme.
Thus, [Delgrande, 1987; 1988] employed conditional logics for reasoning about
typicality.
The paper [Kraus et al., 1990] constituted a turning point in the development of the preferential approach. Based on Gabbay’s earlier description of
generalized inference relations, given in [Gabbay, 1985], and on semantic ideas
of [Shoham, 1988], the authors described both semantics and axiomatization for
a range of nonmonotonic inference relations. They also strongly argued that
preferential conditionals provide a more adequate and versatile formalization of
the notion of normality than, say, default logic. [Kraus et al., 1990] has established the logical foundations for a research program that has attracted many
researchers, both in AI and in logic. It has been found, in particular, that the
new approach to nonmonotonic reasoning is intimately connected also with the
influential theory of belief change suggested in [Alchourrón et al., 1985]. As a
consequence, an alternative but equivalent formalization of nonmonotonic inference relations was developed in [Gärdenfors and Makinson, 1994] based on
expectation ordering of classical formulas.
It is worth mentioning here, however, that despite the general enthusiasm
with this new approach, there have also been voices of doubt. Already [Reiter,
1987a] has warned against overly hasty generalizations when it comes to nonmonotonic reasoning. Moreover, Reiter has rightly argued that, for the purposes
of representing nonmonotonic reasoning, these preferential logics have two fatal
flaws; they are (globally) monotonic and extremely weak. On Reiter’s view,
nonmonotonicity is achieved in these logics by pragmatic considerations affecting how the logic is used, which destroys the principled semantics on which these
logics were originally based.
4.1
Epistemic States
Semantic interpretation constitutes one of the main components of a viable reasoning system, monotonic or not. A formal inference engine, though important,
can be effectively used for representing and solving reasoning tasks only if its
basic notions have clear meaning allowing to discern them from a description of
a situation at hand.
The standard semantics of preferential inference relations is based on abstract possible worlds models in which worlds are ordered by a preference relation. A more specific semantic interpretation for such inference relations, suitable for nonmonotonic reasoning, can be obtained, however, based on preference
ordering on sets of default assumptions. This strategy was pursued by Hector
Geffner in [Geffner, 1992] in the context of an ambitious general project in
nonmonotonic reasoning, which showed also how to apply the preferred model
approach to particular reasoning problems. This more specific interpretation
44
also subsumes the approach to nonmonotonic reasoning employed in Poole’s
Theorist system [Poole, 1988a] (see above).
It has been suggested in [Bochman, 2001] that a general representation
framework for preferential nonmonotonic reasoning can be given in terms of
epistemic states, defined below.
Definition. An epistemic state is a triple (S, l, ≺), where S is a set of admissible
belief states, ≺ a preference relation on S, while l is a labeling function assigning
a deductively closed belief set to every state from S.
On the intended interpretation, admissible belief states are generated as logical closures of allowable combinations of default assumptions. Such states are
taken to be the options for choice. The preference relation on admissible belief
states reflects the fact that not all admissible combinations of defaults constitute equally preferred options for choice. For example, defaults are presumed
to hold, so an admissible belief state generated by a larger set of defaults is
normally preferred to an admissible state generated by a smaller set of defaults.
In addition, not all defaults are born equal, so they may have some priority
structure that imposes, in turn, additional preferences among belief states (see
below).
Epistemic states guide our decisions what to believe in particular situations.
They are epistemic, however, precisely because they say nothing directly about
what is actually true, but only what is believed (or assumed) to hold. This makes
epistemic states relatively stable entities; change in facts and situations will not
necessary lead to change in epistemic states. The actual assumptions made in
particular situations are obtained by choosing preferred admissible belief states
that are consistent with the facts.
4.1.1
Prioritization
An explicit construction of epistemic states generated by default bases provides
us with characteristic properties of epistemic states arising in particular reasoning contexts. An epistemic state is base-generated by a set ∆ of propositions
with respect to a classical Tarski consequence relation Th if
• the set of its admissible states is the set P(∆) of subsets of ∆;
• l is a function assigning each Γ ⊆ ∆ a theory Th(Γ);
• the preference order is monotonic on P(∆): if Γ ⊂ Φ, then Γ ≺ Φ.
The preference order on admissible belief states is usually derived in some
way from priorities among individual defaults. This task turns out to be a
special case of a general problem of combining a set of preference relations into
a single ‘consensus’ preference order. Let us suppose that the set of defaults
∆ is ordered by some priority relation C which will be assumed to be a strict
partial order: α C β will mean that α is prior to β.
Recall that defaults are beliefs we are willing to hold insofar as it is consistent
to do so. Hence any default δ determines a primary preference relation 4δ on
45
P(∆) by which admissible belief sets containing the default are preferred to
belief sets that do not contain it:
Γ 4δ Φ ≡ if δ ∈ Γ then δ ∈ Φ
Each 4δ is a weak order having just two equivalence classes, namely sets
of defaults that contain δ, and sets that don’t. In this setting, the problem
of finding a global preference order amounts to constructing an operator that
maps a set of preference relations {4δ | δ ∈ ∆} to a single preference relation 4
on P(∆). As has been shown in [Andreka et al., 2002], any finitary operator of
this kind satisfying the so-called Arrow’s conditions is definable using a priority
graph (N, C, v), where C is a priority order on a set of nodes N , and v is a
labeling function assigning each node a preference relation. The priority graph
determines a single resulting preference relation via the lexicographic rule, by
which t is weakly preferred to s overall if it is weakly preferred for each argument preference, except possibly those for which there is a prior preference that
strictly prefers t to s:
s 4 t ≡ ∀i ∈ N (s 4v(i) t ∨ ∃j ∈ N (j C i ∧ s ≺v(j) t))
In our case, the prioritized base (∆, C) can be viewed as a priority graph in
which every node δ is assigned a preference relation 4δ . Consequently, we can
apply the lexicographic rule and arrive at the following definition:
Γ 4 Φ ≡ (∀α ∈ Γ\Φ)(∃β ∈ Φ\Γ)(β C α)
Γ 4 Φ holds when, for each default in Γ \ Φ, there is a prior default in Φ \ Γ.
The corresponding strict preference Γ ≺ Φ is defined as Γ 4 Φ ∧ Γ 6= Φ.
[Lifschitz, 1985] was apparently the first to use this construction in prioritized circumscription, while [Geffner, 1992] employed it for defining preference
relations among sets of defaults (see also [Grosof, 1991]).
4.2
Nonmonotonic inference and its kinds
In particular situations, we restrict our attention to admissible belief sets that
are consistent with the facts, and choose preferred among them. The latter are
used to support the assumptions and conclusions we make about the situation
at hand. Accordingly, all kinds of nonmonotonic inference relations, described
below, presuppose a two-step selection procedure: for a current evidence A, we
consider admissible belief states that are consistent with A and choose preferred
elements in this set.
An admissible belief state s ∈ S will be said to be compatible with a proposition A, if ¬A does not belong to its belief set, that is, ¬A ∈
/ l(s). The set of
all admissible states that are compatible with A will be denoted by hAi.
A skeptical inference (or prediction) with respect to an epistemic state is
obtained when we infer only what is supported by each of the preferred states.
In other words, B will be a skeptical conclusion from the evidence A in an
epistemic state E if each preferred admissible belief set in E that is consistent
with A, taken together with A itself, implies B.
46
Definition. B is a skeptical consequence of A (notation A |∼ B) in an epistemic
state if A → B is supported by all preferred belief states in hAi.
A set of conditionals A |∼ B that are valid in an epistemic state E will be
called a skeptical inference relation determined by E.
The above definition generalizes the notion of prediction from [Poole, 1988a],
as well as an expectation-based inference of [Gärdenfors and Makinson, 1994].
But in fact, it is much older. While the semantics of nonmonotonic inference
from [Kraus et al., 1990] derives from the possible worlds semantics of Stalnaker
and Lewis, the above definition can be traced back to the era before the discovery
of possible worlds, namely to Frank Ramsey and John S. Mill:
“In general we can say with Mill that ‘If p then q’ means that q
is inferrable from p, that is, of course, from p together with certain
facts and laws not stated but in some way indicated by the context.”
[Ramsey, 1978, page 144]
This definition has also been used in the ‘premise-based’ semantics for counterfactuals [Veltman, 1976; Kratzer, 1981] (see also [Lewis, 1981]).
A credulous inference (or explanation) with respect to an epistemic state is
obtained by assuming that we can reasonably infer (or explain) conclusions that
are supported by at least one preferred belief state consistent with the facts. In
other words, B will be a credulous conclusion from A if at least one preferred
admissible belief set in E that is consistent with A, taken together with A itself,
implies B. In still other words,
Definition. B is a credulous consequence of A in an epistemic state if A→B is
supported by at least one preferred belief state in hAi.
The set of conditionals that are credulously valid in an epistemic state E
forms a credulous inference relation determined by E. The above definition
constitutes a generalization of the corresponding definition of explanation in
Poole’s abductive system [Poole, 1988a]. Credulous inference is only one, though
important, instance of a broad range of non-skeptical inference relations (see
[Bochman, 2003a]).
4.3
Syntactic characterizations
As we mentioned, [Gabbay, 1985] was a starting point of the approach to nonmonotonic reasoning based on describing associated inference relations. This
approach was primarily designed to capture the skeptical view of nonmonotonic
reasoning. We have seen, however, that many reasoning tasks in AI, such as
abduction and diagnosis, are based on a credulous understanding of nonmonotonic reasoning. A common ground for both skeptical and credulous inference
can be found in the logic of conditionals suggested in [van Benthem, 1984].
The main idea behind van Benthem’s approach was that a conditional can
be seen as a generalized quantifier representing a relation between the respective
sets of instances or situations supporting and refuting it. A situation confirms
47
a conditional A |∼ B if it supports the classical implication A→B, and refutes
A |∼ B if it supports A→¬B. Then the validity of a conditional in a set of
situations is determined by appropriate, ‘favorable’ combinations of confirming
and refuting instances. Whatever they are, we can assume that adding new
confirming instances to a valid conditional, or removing refuting ones, cannot
change its validity. Accordingly, we can accept the following principle:
If all situations confirming A |∼ B confirm also C |∼ D, and all situations refuting C |∼ D refute also A |∼ B, then validity of A |∼ B
implies validity of C |∼ D.
The above principle is already sufficient for justifying the rules of the basic inference relation, given below. It supports also a representation of conditionals in terms of expectation relations. More exactly, we can say that a
conditional is valid if the set of its confirming situations is sufficiently good
compared with the set of its refuting situations. Accordingly, A |∼ B can be
defined as A→¬B < A→B for an appropriate expectation relation <, while the
above principle secures that this is a general expectation relation as defined in
[Bochman, 2001].
A basic inference relation B satisfies the following postulates:
Reflexivity
A |∼ A
Left Equivalence
Right Weakening
Antecedence
Deduction
If A ↔ B and A |∼ C, then B |∼ C
If A |∼ B and B C, then A |∼ C
If A |∼ B, then A |∼ A ∧ B
If A ∧ B |∼ C, then A |∼ B → C
Very Cautious Monotony
If A |∼ B ∧ C, then A ∧ B |∼ C
A set Γ of conditionals implies a conditional α with respect to basic inference
if α belongs to the least basic inference relation containing Γ. Note, however,
that all the above postulates involve at most one conditional premise. As a
result, the basic entailment boils down to a derivability relation among single
conditionals. The following theorem describes this derivability relation in terms
of the classical entailment.
Theorem. A |∼ B B C |∼ D if and only if either C D, or A → B C → D
and C→¬D A→¬B.
Theorem 4.3 justifies the principle stated earlier: A |∼ B implies C |∼ D if
and only if either all situations confirm C |∼ D, or else all confirming instances
of the former are confirming instances of the latter and all refuting instances of
the latter are refuting instances of the former.
Basic inference relation is a weak inference system, first of all because it does
not allow to combine different conditionals. Nevertheless, it is in a sense complete so far as we are interested in derivability among individual conditionals.
48
More exactly, basic derivability captures exactly the one-premise derivability of
both skeptical and credulous inference relations.
As was noted already in [Gabbay, 1985], a characteristic feature of sceptical
inference is the validity of the following postulate:
(And) If A |∼ B and A |∼ C, then A |∼ B ∧ C.
Indeed, in the framework of basic inference, And is all we need for capturing
precisely the preferential inference relations from [Kraus et al., 1990]. Such
inference relations provide a complete axiomatization of skeptical inference with
respect to epistemic states. An important special case of preferential inference,
rational inference relations, are determined by linearly ordered epistemic states;
they are obtained by adding further
(Rational Monotony) If A |∼ B and A | ¬C, then A ∧ C |∼ B.
In contrast, credulous inference relations do not satisfy And. Still, they are
axiomatized as basic inference relations satisfying Rational Monotony.
Rational Monotony is not a ‘Horn’ rule, so it does not allow us to derive new
conditionals from given ones. In fact, credulous inference relations do not derive
much more conditionals than what can be derived already by basic inference
(see [Bochman, 2001]). This indicates that there should be no hope to capture
credulous nonmonotonic reasoning by derivability in some nonmonotonic logic.
Something else should be added to the above logical framework in order to
represent the relevant form of nonmonotonic reasoning. Though less evident,
the same holds for skeptical inference. Both these kinds of inference need to be
augmented with an appropriate globally nonmonotonic semantics that would
provide a basis for the associated systems of defeasible entailment, as described
in the next section.
4.4
Defeasible entailment
A theory of reasoning about default conditionals should occupy an important
place in the general theory of nonmonotonic reasoning. Thus, the question
whether a proposition B is derivable from an evidence A in a default base
is reducible to the question whether the conditional A |∼ B is derivable from
the base, so practically all reasoning problems about default conditionals are
reducible to the question what conditionals can be derived from a conditional
default base. The latter problem constitutes therefore the main task of a theory
of default conditionals (see [Lehmann and Magidor, 1992]).
For a skeptical reasoning, a most plausible understanding of default conditionals is obtained by treating them as skeptical inference rules in the framework
of epistemic states. Accordingly, preferential inference relations of [Kraus et al.,
1990] can be considered as a logic behind skeptical nonmonotonic reasoning; the
rules of the former should be taken for granted by the latter.
This does not mean, however, that nonmonotonic reasoning about default
conditionals is reducible to preferential derivability. Preferential inference is
49
severely sub-classical and does not allow us, for example, to infer “Red birds
fly” from “Birds fly”. In fact, this is precisely the reason why such inference
relations have been called nonmonotonic. Clearly, there are good reasons for not
accepting such a derivation as a logical rule for preferential inference; otherwise
“Birds fly” would imply also “Birds with broken wings fly” and even “Penguins
fly”. Still, this should not prevent us from accepting “Red birds fly” on the basis
of “Birds fly” as a reasonable nonmonotonic (or defeasible) conclusion, namely
a conclusion made in the absence of information against it. By doing this, we
would just follow the general strategy of nonmonotonic reasoning that involves
making reasonable assumptions on the basis of available information. Thus,
the logical core of skeptical inference, preferential inference relations, should be
augmented with a mechanism of making nonmonotonic conclusions. This kind of
reasoning will of course be defeasible, or globally nonmonotonic, since addition
of new conditionals can block some of the conclusions made earlier. Note in
this respect that, though preferential inference is (locally) nonmonotonic with
respect to premises of conditionals, it is nevertheless globally monotonic: adding
new conditionals does not change the validity of previous derivations.
On the semantic side, default conditionals are constraints on epistemic states
in the sense that the latter should make them skeptically valid. Still, usually
there is a huge number of epistemic states that satisfy a given set of conditionals,
so we have both an opportunity and necessity to choose among them. Our
guiding principle in this choice can be the same basic principle of nonmonotonic
reasoning, namely that the intended epistemic states should be as normal as is
permitted by the current constraints. By choosing particular such states, we
thereby will adopt conditionals that would not be derivable from a given base
by preferential inference alone.
The above considerations lead to a seemingly inevitable conclusion that default conditionals possess a clear logical meaning and associated logical semantics based on epistemic states (or possible worlds models), but they still lack a
globally nonmonotonic semantics that would provide an interpretation for the
associated defeasible entailment.
Actually, the literature on nonmonotonic reasoning is abundant with such
theories of defeasible entailment. A history of studies on this subject could
be summarized as follows. Initial formal systems, namely Lehmann’s rational closure [Lehmann, 1989; Lehmann and Magidor, 1992] and Pearl’s system
Z [Pearl, 1990], have turned out to be equivalent. This encouraging development has followed by a realization that both theories are insufficient for representing defeasible entailment, since they do not allow to make certain intended conclusions. Hence, they have been refined in a number of ways, giving such systems as lexicographic inference [Benferhat et al., 1993; Lehmann,
1995], and similar modifications of Pearl’s system [Goldszmidt et al., 1993;
Tan and Pearl, 1995]. Unfortunately, these refined systems have encountered an
opposite problem, namely, together with some desirable properties, they invariably produced some unwanted conclusions. All these systems have been based
on a supposition that defeasible entailment should form a rational inference relation. A more general approach in the framework of preferential inference has
50
been suggested in [Geffner, 1992].
Yet another, more syntactic, approach to defeasible entailment has been pursued in the framework of inheritance hierarchies (see [Horty, 1994]). Inheritance
reasoning deals with a quite restricted class of conditionals constructed from
literals. Nevertheless, in this restricted domain it has achieved a remarkably
close correspondence between what is derived and what is expected intuitively.
Accordingly, inheritance reasoning has emerged as an important test bed for
adjudicating proposed theories.
Despite the diversity, the systems of defeasible entailment have a lot in common, and take as a starting point a few basic principles. Thus, most of them
presuppose that intended models should be described, ultimately, in terms of
material implications corresponding to a given set of conditionals. More exactly, these classical implications should serve as defaults in the nonmonotonic
reasoning sanctioned by a default base. This idea can be made precise as follows:
Default base-generation. The intended epistemic states for a
default base B should be base-generated by the corresponding set
~ of material implications.
B
In other words, the admissible belief states of intended epistemic states for
~ and the preference order
a default base B should be formed by subsets of B,
should be monotonic on these subsets (see Section 4.1.1). In addition, it should
be required that all the conditionals from B should be skeptically valid in the
resulting epistemic state.
Already these constraints on intended epistemic states allow us to derive
“Red birds fly” from “Birds fly” for all default bases that do not contain conflicting information about redness. The constraints also sanction defeasible entailment across exception classes: if penguins are birds that normally do not
fly, while birds normally fly and have wings, then we are able to conclude that
penguins normally have wings, despite being abnormal birds. This excludes, in
effect, Pearl’s system Z and rational closure that cannot make such a derivation. Still, these requirements are quite weak and do not produce problematic
conclusions that plagued some stronger systems suggested in the literature.
Unfortunately, though the above constraints deal successfully with many examples of defeasible entailment, they are still insufficient for capturing some
important reasoning patterns. What is missing in our construction is a principled way of constructing a preference order on default sets. This problem
has turned out to be far from being trivial or univocal. As for now, two most
plausible solutions to this problem, suggested in the literature, are Geffner’s
conditional entailment and inheritance reasoning.
Conditional entailment [Geffner, 1992] determines a prioritization of default
bases by making use of the following relation among conditionals:
Definition. A conditional α dominates a set of conditionals Γ if the set of
implications {~Γ, α
~ } is incompatible with the antecedent of α.
The origins of this relation can be found already in [Adams, 1975], and it has
been used in practically all studies of defeasible entailment, including the notion
51
of preemption in inheritance reasoning. A suggestive reading of dominance says
that if α dominates Γ, it should have priority over at least one conditional in
Γ7 . Accordingly, a priority order on the default base is admissible if it satisfies
this condition. Then the intended models can be identified with epistemic states
that are generated by all admissible priority orders on the default base (using
the lexicographic rule - see above).
Conditional entailment has shown itself as a serious candidate on the role of
a general theory of defeasible entailment. Still, it does not capture inheritance
reasoning. The main difference between the two theories is that conditional
entailment is based on absolute priorities among defaults, while inheritance
hierarchies determine such priorities in a context-dependent way, namely in
presence of other defaults that provide a (preemption) link between two defaults
(see [Dung and Son, 2001]). Indeed, it has been shown in [Bochman, 2001]
that inheritance reasoning is representable by epistemic states that are basegenerated by default conditionals ordered by certain conditional priority orders.
Still, the corresponding construction could hardly be called simple or natural.
A more natural representation of inheritance reasoning has been given in
[Dung and Son, 2001] as an instantiation of an argumentation theory that belongs already to explanatory nonmonotonic formalisms, discussed in the next
section. Furthermore, Geffner himself has shown in [Geffner, 1992] that conditional entailment still does not capture some important derivations, and it
should be augmented with an explicit representation of causal reasoning. In
fact, the causal generalization suggested by Geffner in the last chapters of his
book has served as one of the inspirations for a causal theory of reasoning about
actions and change (see [Turner, 1999]).
5
Explanatory nonmonotonic reasoning
The ‘mainstream’ approach to nonmonotonic reasoning includes default and
modal nonmonotonic logics, logic programming, abductive and causal reasoning. We will call this approach explanatory nonmonotonic reasoning, since explanation can be seen as its basic ingredient. Propositions and facts may be not
only true or false in a model of a problem situation, but some of them are explainable (justified) by other facts and rules that are accepted. In the epistemic
setting, some of the propositions are derivable from other propositions using
rules that are admissible in the situation. In the objective setting, some of the
facts are caused by other facts and causal rules acting in the domain. Furthermore, explanatory nonmonotonic reasoning is based on very strong principles
of Explanation Closure or Causal Completeness (see [Reiter, 2001]), according
to which any fact holding in a model should be explained, or caused, by the
rules that describe the domain. Incidentally, it is these principles that make
explanatory reasoning nonmonotonic.
By the above description, abduction, that is, reasoning from facts to their
explanations, is an integral part of explanatory nonmonotonic reasoning. Ulti7 This
secures that α will be valid in the resulting epistemic state.
52
mate explanations, or abducibles, correspond not to normality defaults, but to
conjectures representing base causes or facts that do not require explanation;
we assume the latter only for explaining some evidence.
In some domains, explanatory formalisms adopt simplifying assumptions
that exempt, in effect, certain propositions from the burden of explanation.
Closed World Assumption [Reiter, 1978] is the most important assumption of
this kind. According to it, negative assertions do not require explanation. Nonmonotonic reasoning in databases and logic programming are domains for which
such an assumption turns out to be most appropriate. It is important to note
that the minimization principle that has been the source of the preferential approach can also be derived as a result of combining Explanation Closure with
the Closed World Assumption (see below). Consequently, it need not be viewed
as a principle of scaled preference of negative information; rather, it could be
seen as a by-product of the stipulation that negated propositions can be accepted without any further explanation, while positive assertions always require
explanation. This understanding allows us to explain why McCarthy’s circumscription, that is based on the principle of minimization, is subsumed also by
explanatory formalisms.
The above principles form an ultimate basis for all formal systems of explanatory nonmonotonic reasoning. They presuppose, however, a richer picture
of what is in the world than what is usually captured in logical models of the
latter. The traditional understanding of possible worlds in logic stems from
the Tractatus’ metaphysics where ‘[e]ach item can be the case or not the case
while everything else remains the same’ ([Wittgenstein, 1961]). Consequently,
there is no way of making an inference from one fact to another, and there is
no causal nexus to justify such an inference. The only restriction on the structure of the world is the principle of non-contradiction. Such worlds leave no
place for dependencies among facts and related notions, in particular for causation. Wittgenstein himself concluded that belief in such dependencies is a
superstition.
An alternative picture depicts the world not as a mere assemblage of unrelated facts, but as something that has a structure. This structure determines
dependencies among occurrent facts that serve as a basis for our explanatory
and causal claims. It is this structure that makes the world intelligible and,
what is especially important for AI, controllable. By this picture, explanatory
and causal relations form an integral part of understanding of and acting in the
world. Consequently, such relations should form an essential part of knowledge
representation, at least in Artificial Intelligence.
5.1
Biconsequence Relations
A uniform account of explanatory nonmonotonic formalisms can be given in the
framework of biconsequence relations described in this section. Biconsequence
relations are specialized consequence relations for reasoning with respect to a
pair of contexts. On the interpretation suitable for nonmonotonic reasoning, one
of these contexts is the main (objective) one, while the other context provides
53
assumptions, or explanations, that justify inferences in the main context. This
separation of inferences and their justifications creates a possibility of explanatory nonmonotonic reasoning.
The two contexts will be termed, respectively, the context of truth and the
context of falsity. In the truth context the propositions are evaluated as being
either true or non-true, while in the falsity context they can be false or nonfalse. As a benefit of this terminological decision, a bi-context reasoning can
also be interpreted as a reasoning with possibly inconsistent and incomplete
information. Furthermore, such a reasoning can also be viewed as a four-valued
reasoning (see [Belnap, 1977]).
A bisequent is an inference rule of the form a : b c : d, where a, b, c, d
are finite sets of propositions. We will employ two informal interpretations of
bisequents. According to the four-valued interpretation, it says
‘If all propositions from a are true and all propositions from b are
false, then either one of the propositions from c is true or one of the
propositions from d is false’.
According to the explanatory (or assumption) interpretation, it says
‘If no proposition from b is assumed, and all propositions from d
are assumed, then all propositions from a hold only if one of the
propositions from c holds’.
A biconsequence relation is a set of bisequents satisfying the rules:
Monotonicity
Reflexivity
Cut
a:bc:d
, if a ⊆ a0 , b ⊆ b0 , c ⊆ c0 , d ⊆ d0 ;
: b0 c0 : d0
a0
A:A:
and
: A : A;
a : b A, c : d A, a : b c : d
a:bc:d
a : b c : A, d a : A, b c : d
.
a:bc:d
A biconsequence relation can be seen as a fusion, or fibring, of two Scott
consequence relations much in the sense of [Gabbay, 1998]. This fusion gives a
syntactic expression to combining two independent contexts.
The definition of a biconsequence relation is extendable to arbitrary sets of
propositions by accepting the compactness requirement:
Compactness
u : v w : z iff a : b c : d,
for some finite sets a, b, c, d such that a ⊆ u, b ⊆ v, c ⊆ w and d ⊆ z.
For a set u of propositions, u will denote the set of propositions that do
not belong to u. A pair (u, v) of sets of propositions will be called a bitheory
of a biconsequence relation if u : v 1 u : v. A set u of propositions is a
(propositional) theory of , if (u, u) is a bitheory of .
54
Bitheories can be seen as pairs of sets that are closed with respect to the
bisequents of a biconsequence relation. A bitheory (u, v) of is positively minimal, if there is no bitheory (u0 , v) of such that u0 ⊂ u. Such bitheories play
an important role in describing nonmonotonic semantics.
By a bimodel we will mean a pair of sets of propositions. A set of bimodels
will be called a binary semantics.
Definition. A bisequent a : b c : d is valid in a binary semantics B, if, for
any (u, v) ∈ B, if a ⊆ u and b ⊆ v, then either c ∩ u 6= ∅, or d ∩ v 6= ∅.
The set of bisequents that are valid in a binary semantics forms a biconsequence relation. On the other hand, any biconsequence relation is determined in this sense by its canonical semantics defined as the set of bitheories of
. Consequently, the binary semantics provides an adequate interpretation of
biconsequence relations.
According to Belnap’s idea, the four truth-values {>, t, f , ⊥} of a four-valued
interpretation can be identified with the four subsets of the set {t, f } of classical
truth-values, namely {t, f }, {t}, {f } and ∅. Thus, > means that a proposition is
both true and false (i.e., contradictory), t means that it is ‘classically’ true (that
is, true without being false), f means that it is classically false, while ⊥ means
that it is neither true nor false (undetermined). This representation allows us
to see any four-valued interpretation ν as a pair of ordinary interpretations
corresponding, respectively, to independent assignments of truth and falsity to
propositions:
ν |= A
iff t ∈ ν(A)
ν =| A
iff
f ∈ ν(A).
Now, a bimodel (u, v) can be viewed as a four-valued interpretation, where
u is the set of true propositions, while v is the set of propositions that are not
false. Biconsequence relations provide in this sense a syntactic formalism for
four-valued reasoning.
A bisequent theory is an arbitrary set of bisequents. For any bisequent
theory ∆ there is a least biconsequence relation ∆ containing it that describes
the logical content of ∆. This allows us to extend the notions of a bitheory and
propositional theory to arbitrary bisequent theories.
Two kinds of bisequents are of special interest for nonmonotonic reasoning.
The first are default bisequents a:b c: without negative conclusions that are
related to rules of default logic. The second are autoepistemic bisequents :b c:d
that are related to autoepistemic logic. A default bisequent says that if no
proposition from b is assumed, then all propositions from a hold only if one
of the propositions from c holds. Such a bisequent can be viewed as a Scott
inference rule a ` c that is conditioned by a set of negative assumptions (i.e.,
absence of b’s). Thus, such bisequents involve full inference capabilities with
respect to the main context, but permit only negative assumptions. In contrast,
an autoepistemic bisequent :b c:d says that if no proposition from b is assumed,
and all propositions from d are assumed, then one of the propositions from c
holds. Such rules have rich assumption capabilities, but allow us to make only
unconditional assertions.
55
In addition to the above distinction, we are often interested in singular bisequents, namely bisequents a:b C:d having a single proposition C as a positive
conclusion. Such bisequents can be seen as counterparts of Tarski rules. The
formalism of default consequence relations, introduced in [Bochman, 1994], provided, in effect, a uniform description of singular default and autoepistemic
bisequents as species of default rules of the form a:b ` C. The difference of
default vs. autoepistemic rules has been reflected, however, as a difference in
corresponding logics for default rules. This common representation has given a
convenient basis for a comparative study of default and modal nonmonotonic
reasoning (see [Bochman, 1998c]).
An important feature of biconsequence relations is a possibility of imposing
structural constraints on the binary semantics by accepting additional structural
rules. Some of them play an important role in nonmonotonic reasoning. Thus,
a biconsequence relation is consistent, if it satisfies
Consistency
A:A
On the explanatory interpretation, Consistency says that no proposition can
be taken to hold without assuming it. This amounts to restricting the binary
semantics to consistent bimodels, that is, bimodels (u, v) such that u ⊆ v. On
the four-valued representation, Consistency requires that no proposition can be
both true and false, so it determines a semantic setting of partial logic (see,
e.g., [Blamey, 1986]) that usually deals only with possible incompleteness of
information.
A biconsequence relation is regular if it satisfies
Regularity
b:aa:b
: a : b
Regularity is a kind of an assumption coherence constraint. It says that a
coherent set of assumptions should be such that it is compatible with taking
these assumptions as actually holding. A semantic counterpart of Regularity is
a quasi-reflexive binary semantics in which, for any bimodel (u, v), (v, v) is also
a bimodel.
Any four-valued connective is definable in biconsequence relations via a pair
of introduction rules and a pair of elimination rules corresponding to the two
valuations of a four-valued interpretation. We are primarily interested, however,
in information a bi-context reasoning can give us about ordinary, classical truth
and falsity, so we restrict attention to classical connectives that are conservative
on the subset {t, f }. Such connectives give classical truth-values when their
arguments receive classical values t or f .
There are four natural connectives that are jointly sufficient for defining all
classical four-valued functions. The first is the well-known conjunction:
ν |= A ∧ B
ν =| A ∧ B
iff
ν |= A and ν |= B
iff ν =| A or ν =| B.
56
Next, there are two negation connectives that can be seen as two alternative
extensions of classical negation to the four-valued setting:
ν |= ¬A
iff ν 6|= A
ν =| ¬A iff
ν 6=| A
ν |= ∼A
iff ν =| A
ν =| ∼A
ν |= A.
iff
We will call ¬ and ∼ a local and global negation, respectively. Each of them
can be used together with the conjunction to define a disjunction:
A ∨ B ≡ ∼(∼A ∧ ∼B) ≡ ¬(¬A ∧ ¬B).
Finally, the unary connective A can be seen as a kind of a modal operator,
closely related to the modal operator L that will be used in the modal extension
of our formalism.
ν |= AA
iff ν 6=| A
ν =| AA
iff ν =| A.
From the classical point of view, the most natural subclass of the classical
four-valued connectives is formed by connectives that behave as ordinary classical connectives with respect to each of the two contexts. Connectives of this
kind will be called locally classical. The conjunction ∧ and local negation ¬
form a functionally complete basis for all such connectives.
Having the four-valued connectives at our disposal, we can transform bisequents into more familiar inference rules, and even to ordinary logical formulas.
Let ¬u denote the set {¬A | A ∈ u}, and similarly for ∼u, etc. Then any
bisequent a : b c : d is equivalent to each of the following:
a, ∼b : c, ∼d :
(1)
¬a, c : d, ¬b
(2)
Bisequents of the form (1) can be seen as ordinary sequents. In fact, this
is a common trick used for representing many-valued logics in the form of a
sequent calculus. If the language contains also conjunction (or, equivalently,
disjunction), the set of premises can be replaced by their conjunction, while
the set of conclusions - by their disjunction. Consequently, we can transform
bisequents into Tarski-type rules A ` B. Actually, the resulting system will
coincide with a (flat) theory of relevant entailment [Dunn, 1976].
An important alternative possibility arises from using the representation
(2). This time we can use the local connectives {∧, ¬} in order to reduce such
bisequents to that of the form A : B, where A and B are classical propositions.
The latter bisequents will correspond to rules B ⇒ A of production and causal
inference relations, discussed later.
5.2
Nonmonotonic Semantics
Nonmonotonic semantics of a biconsequence relation is a certain set of its theories. Namely, such theories are explanatory closed in the sense that presence
and absence of propositions in the main context is explained (i.e., derived) using the rules of the biconsequence relation when the theory itself is taken as the
assumption context.
57
5.2.1
Exact semantics
The notion of an exact theory of a biconsequence relation provides us with a
simplest and most general kind of nonmonotonic reasoning.
Definition. A theory u of a biconsequence relation is exact, if there is no
other set v 6= u such that (v, u) is a bitheory of . The set of exact theories will
be called an exact nonmonotonic semantics of .
Exact theories correspond to bitheories, for which the assumption context
determines itself as a unique objective state compatible with it. Such theories
can be given the following syntactic description.
Lemma. A set u of propositions is an exact theory of a biconsequence relation
if and only if, for any proposition A,
A ∈ u iff : u A : u
and
A∈
/ u iff A : u : u.
The exact semantics is extendable to arbitrary bisequent theories by considering their associated biconsequence relations. Regular biconsequence relations
constitute a maximal logic suitable for the exact nonmonotonic semantics.
The above definition of nonmonotonic semantics leaves us much freedom
in determining nonmonotonic consequences of a bisequent theory. The most
obvious skeptical choice consists in taking propositions that belong to all exact
theories. As a credulous alternative, however, we can consider propositions that
belong to at least one theory. An even more general understanding amounts to
a view that all the information that can be discerned from the nonmonotonic
semantics of a bisequent theory or a biconsequence relation can be seen as
nonmonotonically implied by the latter.
Exact theories determine a truly nonmonotonic semantics, since adding new
bisequents to a bisequent theory may not only eliminate exact theories, but
also add new ones (or both). In other words, the set of exact theories does not
change monotonically with the growth of the set of bisequents.
Finally, if a bisequent theory contains only default bisequents a:b c:, then
any its exact theory will also be a minimal propositional theory.
5.2.2
Default semantics
A more familiar class of nonmonotonic models, extensions, correspond to extensions of default logic and stable models of logic programs.
Definition. A set u is an extension of a biconsequence relation , if (u, u) is
a positively minimal bitheory of . A default nonmonotonic semantics of a
biconsequence relation is the set of its extensions.
u is an extension if it is a theory of a biconsequence relation such that there
is no smaller set v ⊂ u such that (v, u) is a bitheory. Hence any exact theory of
a biconsequence relation is an extension, though not vice versa.
Let P r denote the set of all propositions of the language. The following
lemma provides a syntactic description of extensions.
58
Lemma. A set u is an extension of a biconsequence relation if and only if
u = {A | :u A, u:u} and either u 6= P r, or 1: P r.
By the above description, extensions are theories of a biconsequence relation
that explain only why they have the propositions they have. In other words,
for extensions we are relieved from the necessity of explaining why propositions
do not belong to the intended theory. This agreement constitutes the essence
of Reiter’s Closed World Assumption. Another consequence of the above description is that extensions are determined by the autoepistemic bisequents of
a biconsequence relation.
It turns out that the default nonmonotonic semantics is precisely an exact
semantics under a stronger logic of consistent biconsequence relations. The effect
of Consistency A : A amounts to an immediate refutation of any proposition
that is assumed not to hold. That is why absence of propositions from extensions
does not require explanation, only their presence. This makes the minimality
condition in the definition of extensions a consequence of the logical Consistency
postulate, instead of an independent ‘rationality principle’ of nonmonotonic
reasoning.
5.2.3
Bisequent interpretation of logic programs
Representations of general logic programs and their semantics in the logical
formalism of biconsequence relations establish the ways of providing (or, better,
restoring) a logical basis of logic programming.
Recall that a general logic program is a set of program rules not d, c ←
a, not b, where a, b, c, d are finite sets of propositional atoms. Such program rules
involve disjunctions and default negations in heads and subsume practically all
structural extensions of Prolog rules suggested in the literature.
A program rule not d, c ← a, not b can be directly interpreted as a bisequent
a : b c : d. Let bc(Π) denote the bisequent theory corresponding to a general
program Π under this interpretation. The following result shows that it provides an exact correspondence between stable models of a general program and
extensions of the associated bisequent theory.
Theorem. Stable models of a general logic program Π coincide with the extensions of bc(Π).
Moreover, the correspondence turns out to be bidirectional, since any bisequent in a four-valued language is logically reducible to a set of bisequents
without connectives; such bisequents can already be viewed as program rules.
5.3
Causal and Production Inference
As we mentioned earlier, in languages containing certain (four-valued) connectives, bisequents are reducible to much simpler rules. In particular, in a language with local classical connectives, bisequents are reducible to rules of the
form A ⇒ B, where A and B are classical logical formulas (cf. representation
59
(2) above). Such rules can be given an informal reading ‘A causes, or explains,
B’, and the resulting logical system will turn out to coincide with the system
causal inference described in Section 3.6.2. Speaking more formally, for a biconsequence relation in a local classical language, we can define the following
set of rules, called the production subrelation of :
⇒ = {A ⇒ B | B : A}
Then the production subrelation of a regular biconsequence relation will
form a causal inference relation. Moreover, for any causal inference relation ⇒
there is a unique regular biconsequence relation in the language with the local
connectives that has ⇒ as its production subrelation. By this correspondence, a
causal rule A ⇒ B can be seen as an assumption-based conditional saying that if
A is assumed, then B should hold. In this sense, the system of causal inference
constitutes a primary logical system of explanatory nonmonotonic reasoning,
whereas biconsequence relations form a structural counterpart of this logical
formalism. In particular, in the general correspondence between causal and biconsequence relations, the causal nonmonotonic semantics (see Definition 3.6.2)
corresponds to the exact nonmonotonic semantics of biconsequence relations.
As for biconsequence relations, the default nonmonotonic semantics of causal
theories can be obtained by imposing a causal postulate corresponding to the
Consistency postulate for biconsequence relations:
(Default Negation)
¬p ⇒ ¬p, for any propositional atom p.
Default Negation stipulates that negations of atomic propositions are selfexplanatory, and hence it provides a simple causal expression for Reiter’s Closed
World Assumption. As we already mentioned in Section 3.6.2, this kind of causal
inference can be used as a logical basis for logic programming.
Production inference. A useful generalization of causal inference is obtained
by dropping the postulate Or of causal inference that has allowed for reasoning
by cases. The resulting formalism has been called in [Bochman, 2004a] the
system of production inference. For this formalism, we can generalize also the
corresponding nonmonotonic semantics as follows:
Definition. A nonmonotonic production semantics of a production inference
relation is the set of all its exact theories, namely sets u of propositions such
that u = C(u).
As before, an exact theory describes an informational state in which every
proposition is explained by other propositions accepted in this state. Accordingly, restricting our universe of discourse to exact theories amounts to imposing
a kind of an explanatory closure assumption on intended models.
The nonmonotonic production semantics allows us to provide a formal representation of abductive reasoning that covers the main applications of abduction
in AI.
60
To begin with, abducibles can be identified with self-explanatory propositions
of a production relation, that is, propositions satisfying A ⇒ A. Then it turns
out that traditional abductive systems are representable via a special class of
abductive production inference relations that satisfy
(Abduction) If B ⇒ C, then B ⇒ A ⇒ C, for some abducible A.
It has been shown in [Bochman, 2005] that abductive inference relations provide a generalization of abductive reasoning in causal theories [Konolige, 1992;
Poole, 1994b], as well as of abduction in logic programming. On the other hand,
any production inference relation includes a greatest abductive subrelation, and
in many regular situations (e.g., when the production relation is well-founded)
the latter determines the same nonmonotonic semantics. Summing up, the general nonmonotonic semantics of a production relation is usually describable by
some abductive system, and vice versa.
5.4
Epistemic Explanatory Reasoning
Epistemic formalisms of default and modal nonmonotonic logics find their natural place in the framework of supraclassical biconsequence relations, defined
below. An epistemic understanding of biconsequence relations amounts to treating the main and assumption contexts, respectively, as the contexts of knowledge
and belief: propositions that hold in the main context can be viewed as known,
while propositions of the assumption context form the associated set of beliefs.
This understanding will later receive an explicit expression in the modal extension of the formalism. Even in a non-modal setting, however, the epistemic
reading implies that both contexts should correspond not to complete worlds,
but to incomplete deductive theories.
Definition. A biconsequence relation in a classical language is supraclassical,
if it satisfies
Supraclassicality If a A, then a : A : and : A : a.
Falsity
f:
and
: f.
In supraclassical biconsequence relations both contexts respect the classical
entailment. In addition, sets of positive premises and negative conclusions can
be replaced by their conjunctions, but positive conclusion sets and negative
premise sets are not replaceable in this way by classical disjunctions. Also, the
deduction theorem, contraposition, and disjunction in the antecedent are not
valid, in general, for each of the two contexts.
A semantics of supraclassical biconsequence relations is obtained from the
general binary semantics by requiring that bimodels are pairs of consistent deductively closed sets. Structural rules for biconsequence relations can also be
used in the supraclassical case. As before, Consistency will correspond to the
requirement that u ⊆ v, for any bimodel (u, v). Similarly, regular biconsequence
relations will be determined by quasi-reflexive binary semantics.
61
A supraclassical biconsequence relation will be called saturated, if it is consistent, regular, and satisfies the following postulate:
Saturation
A ∨ B, ¬A ∨ B : B.
For a deductively closed set u, let u⊥ denote the set of all maximal subtheories of u, plus u itself. Then a classical bimodel (u, v) will be called saturated,
if u ∈ v⊥. A classical binary semantics B will be called saturated if it is regular,
and all its bimodels are saturated. Such a semantics provides an adequate
interpretation for saturated biconsequence relations.
5.4.1
Classical nonmonotonic semantics
The notions of an exact theory and extension can be directly extended to supraclassical consequence relations, with the only, though important, qualification
that they form now deductively closed sets. Still, practically all the results
about such semantics remain valid for the supraclassical case.
The default nonmonotonic semantics of supraclassical biconsequence relations forms a generalization of default logic. Supraclassical biconsequence relations that are consistent and regular constitute a maximal logic adequate for
extensions. For such biconsequence relations, extensions are described as sets
satisfying the following fixpoint equality:
u = {A : u A : u}.
Thus, an extension is a set of formulas that are provable on the basis of
taking itself as the set of assumptions. As before, classical extensions of a
default bisequent theory will be minimal theories. Actually, default bisequent
theories under this nonmonotonic semantics give an exact representation for
the disjunctive default logic [Gelfond et al., 1991]. For singular default rules
a:b C:, it reduces to the original default logic of [Reiter, 1980].
The nonmonotonic semantics defined below constitutes an exact non-modal
counterpart of Moore’s autoepistemic logic.
Definition. A theory u of a supraclassical biconsequence relation is a classical
expansion of , if, for any v ∈ u⊥ such that v 6= u, the pair (v, u) is not a
bitheory of . The set of classical expansions determines the autoepistemic
semantics of .
Any extension of a supraclassical biconsequence relation will be a classical
expansion, though not vice versa. In fact, classical expansions can be precisely
characterized as extensions of saturated biconsequence relations.
The next result states important sufficient conditions for coincidence of classical expansions and extensions of a bisequent theory. A bisequent theory ∆
will be called positively simple, if positive premises and positive conclusions of
any bisequent from ∆ are sets of classical literals.
Theorem. If a bisequent theory is autoepistemic or positively simple, then its
classical expansions coincide with classical extensions.
62
Bisequents a:b c:d such that a, b, c, d are sets of classical literals, are logical
counterparts of program rules of extended logic programs with classical negation
(see [Gelfond and Lifschitz, 1991; Lifschitz and Woo, 1992]). The semantics of
such programs is determined by answer sets that coincide with extensions of
respective bisequent theories. Moreover, such bisequent theories are positively
simple, so by Theorem 5.4.1 extended logic programs obliterate the distinction
between extensions and classical expansions. This is the logical basis for a
possibility of representing extended logic programs also in autoepistemic logic
(see [Lifschitz and Schwarz, 1993]).
5.5
Modal Nonmonotonic Logics
A general representation of modal nonmonotonic reasoning can be given in the
framework of modal biconsequence relations. The role of the modal operator L
in this setting consists in reflecting assumptions and beliefs as propositions in
the main context.
Definition. A supraclassical biconsequence relation in a modal language will
be called modal if it satisfies the following postulates:
Positive Reflection
Negative Reflection
A : LA:,
: LA :A,
Negative Introspection
: A ¬LA:.
Any theory of a modal biconsequence relation is a modal stable set in the
sense of [Moore, 1985], and hence extensions and expansions of modal biconsequence relations will always be stable theories.
For a modal logic M, a modal biconsequence relation will be called an Mbiconsequence relation, if A: holds for every modal axiom A of M. A possible
worlds semantics for K-biconsequence relations is obtained in terms of Kripke
models having a last cluster, namely models of the form M = (W, R, F, V ),
where (W, R, V ) is an ordinary Kripke model, while F ⊆ W is a non-empty
last cluster of the model (see [Segerberg, 1971]). We will call such models
final Kripke models. The relevance of such semantics for modal nonmonotonic
reasoning has been shown in [Schwarz, 1992a].
For a final Kripke model M = (W, R, F, V ), let |M | denote the set of modal
propositions that are valid in M , while kM k the set of propositions valid in the
S5-submodel of M generated by F . For a set S of final Kripke models, we define
the binary semantics BS = {(|M |, kM k) | M ∈ S}, and say that a bisequent
a:b c:d is valid in S, if it is valid in BS .
Biconsequence relations described in the next definition play a crucial role
in a modal representation of extension-based nonmonotonic reasoning.
Definition. A modal biconsequence relation is an F-biconsequence relation, if
it is regular and satisfies
F
A, LA→B : B.
63
F-biconsequence relations provide a concise representation for the modal
logic S4F obtained from S4 by adding the axiom (A ∧ M LB)→L(M A ∨ B).
A semantic characterization of S4F [Segerberg, 1971] is given in terms of final
Kripke models (W, R, F, V ), such that αRβ iff either β ∈ F , or α ∈
/ F.
Any bisequentVa:b c:d of an F-biconsequence
relation
is
already
reducible to
W
a modal formula (La ∪ L¬Lb) → (Lc ∪ L¬Ld). Consequently, any bisequent
theory ∆ in such a logic is reducible to an ordinary modal theory that we will
˜
denote by ∆.
5.5.1
Modal nonmonotonic semantics
By varying the underlying modal logic, we obtain a whole range of modal nonmonotonic semantics.
Definition. A set of propositions is an M-extension (M-expansion) of a bisequent theory ∆, if it is an extension (resp. expansion) of the least M-biconsequence
relation containing ∆.
If ∆ is a plain modal theory, then M-extensions of ∆ coincide with Mexpansions in the sense of [Marek et al., 1993]8 . Recall, however, that modal
extensions and expansions are modal stable theories, and hence they are determined by their objective subsets. This opens a possibility of reducing modal
nonmonotonic reasoning to a nonmodal one, and vice versa.
For any set u of propositions, let uo denote the set of all non-modal propositions in u. Similarly, if is a modal biconsequence relation, o will denote its
restriction to the non-modal sub-language. Then we have
Theorem. If is a modal biconsequence relation, and u a stable theory, then
u is an extension of if and only if uo is an extension of o .
According to the above result, the objective biconsequence relation o embodies all the information about the modal nonmonotonic semantics of . In
other words, the net effect of modal reasoning in biconsequence relations can be
measured by the set of derived objective bisequents. Consequently, non-modal
supraclassical biconsequence relations turn out to be sufficiently expressive to
capture modal nonmonotonic reasoning.
In the other direction, in modal F-biconsequence relations any bisequent
˜ This allows us to use ordinary
theory ∆ is reducible to a usual modal theory ∆.
modal logical formalisms for representing non-modal nonmonotonic reasoning.
Thus, the following result generalizes the corresponding result of [Truszczyński,
1991] about a modal embedding of default theories.
Theorem. If ∆ is an objective bisequent theory, then classical extensions of ∆
˜
are precisely objective parts of S4F-extensions of ∆.
8 This creates, of course, an unfortunate terminological discrepancy, which is compensated,
however, by the conformity of our terminology with the rest of this paper.
64
We end with considering modal expansions. Two kinds of expansions are
important for a general description. The first is stable expansions of Moore’s
autoepistemic logic. They coincide with M-expansions for any modal logic
M in the range 5 ⊆ M ⊆ KD45. The second kind of expansions is reflexive
expansions of Schwarz’ reflexive autoepistemic logic [Schwarz, 1992b]. They
coincide with M-expansions for any modal logic in the range KT ⊆ M ⊆ SW5.
Both kinds of expansions can be computed either by transforming a bisequent
theory into a modal theory and finding its modal extensions, or by reducing it to
an objective bisequent theory and computing its classical expansions. Finally,
‘normal’ expansions in general can be viewed as a combination of these two
kinds of expansions:
Theorem. A set of propositions is a K-expansion of a modal bisequent theory
∆ if and only if it is both a stable and reflexive expansion of ∆.
6
Conclusions
The following passage from [Reiter, 1987a] remains surprisingly relevant today,
almost twenty years later:
Nonmonotonicity appears to be the rule, rather than the exception, in much of what passes for human commonsense reasoning.
The formal study of such reasoning patterns and their applications
has made impressive, and rapidly accelerating progress. Nevertheless, much remains to be done. ... [M]any more non-toy examples
need to be thoroughly explored in order for us to gain a deeper
understanding of the essential nature of nonmonotonic reasoning.
The ultimate quest, of course, is to discover a single theory embracing all the seemingly disparate settings in AI where nonmonotonic reasoning arises. Undoubtedly, there will be surprises en route,
but AI will profit from the journey, in the process becoming much
more the science we all wish it to be.
Despite clear success, twenty five years of nonmonotonic reasoning research
have shown that we need deep breath and long term objectives in order to make
nonmonotonic reasoning a viable tool for the challenges posed by AI. There
is still much to be done in order to meet the actual complexity of reasoning
tasks required by the latter. In particular, the quest for a unified theory of
nonmonotonic reasoning still has not been accomplished.
The relation between the two principal paradigms of nonmonotonic reasoning, preferential and explanatory one, has emerged as the main theoretical
problem for a future development of the field. By an inspiring analogy, in
nonmonotonic reasoning we have both a global relativity theory of preferential
reasoning and a local quantum mechanics of explanatory reasoning. So, what
we need is a unified theory of nonmonotonic reality. As in physics, however,
this unified theory is not going to emerge as a straightforward juxtaposition of
these components.
65
References
[Adams, 1975] E. W. Adams. The Logic of Conditionals. Reidel, Dordrecht,
1975.
[Alchourrón et al., 1985] C. Alchourrón, P. Gärdenfors, and D. Makinson. On
the logic of theory change: Partial meet contraction and revision functions.
Journal of Symbolic Logic, 50:510–530, 1985.
[Alferes and Pereira, 1992] J. J. Alferes and L. M. Pereira. On logic program
semantics with two kinds of negation. In K. R. Apt, editor, Proc. Joint Int.
Conf. and Symp. on Logic Programming, pages 574–589, Cambridge, Mass.,
1992. MIT Press.
[Amati et al., 1994] G. Amati, L. Carlucci Aiello, and F. Pirri. Defaults as
restrictions on classical hilbert-style proofs. Journal of Logic, Language and
Information, 3:303–326, 1994.
[Amati et al., 1997] G. Amati, L. Carlucci Aiello, and F. Pirri. Definability and
commonsense reasoning. Artificial Intelligence, 93:169–199, 1997.
[Andreka et al., 2002] H. Andreka, M. Ryan, and P.-Y. Schobbens. Operators
and laws for combining preference relations. Journal of Logic and Computation, 12:13–53, 2002.
[Antonelli, 1999] G. A. Antonelli. A directly cautious theory of defeasible consequence for default logic via the notion of general extension. Artificial Intelligence, 109:71–109, 1999.
[Antoniou and Wang, 2006] G. Antoniou and K. Wang. Default logic, 2006.
This volume.
[Antoniou, 1997] G. Antoniou. Nonmonotonic Reasoning. MIT Press, Cambridge, Mass., 1997.
[Apt et al., 1988] K. Apt, H. Blair, and A. Walker. Towards a theory of declarative knowledge. In J. Minker, editor, Foundations of Deductive Databases
and Logic Programming, pages 89–148. Morgan Kaufmann, San Mateo, CA,
1988.
[Babovich et al., 2000] Y. Babovich, E. Erdem, and V. Lifschitz. Fages’ theorem
and answer set programming. In Proceedings of the Nonmonotonic Reasoning
Workshop, NMR-2000, 2000.
[Baker, 1989] A. B. Baker. A simple solution to the Yale shooting problem. In
R. J. Brachman, H. J. Levesque, and R. Reiter, editors, KR’89: Principles of
Knowledge Representation and Reasoning, pages 11–20. Morgan Kaufmann,
San Mateo, California, 1989.
66
[Baral and Subrahmanian, 1991] C. Baral and V. S. Subrahmanian. Dualities
between alternative semantics for logic programming and nonmonotonic reasoning (extended abstract). In Proc. 1st Int. Worksop on Logic Programming
and Nonmonotonic Reasoning, pages 69–86, 1991.
[Baral, 2003] C. Baral. Knowledge Representation, Reasoning and Declarative
Problem Solving. Cambridge UP, 2003.
[Belnap, 1977] N. D. Belnap, Jr. A useful four-valued logic. In M. Dunn and
G. Epstein, editors, Modern Uses of Multiple-Valued Logic, pages 8–41. D.
Reidel, 1977.
[Benferhat et al., 1993] S. Benferhat, C. Cayrol, D. Dubois, J. Lang, and
H. Prade. Inconsistency management and prioritized syntax-based entailment. In R. Bajcsy, editor, Proceedings Int. Joint Conf. on Artificial Intelligence, IJCAI’93, pages 640–645, Chambery, France, 1993. Morgan Kaufmann.
[Besnard, 1989] P. Besnard. An Introduction to Default Logic. Springer, 1989.
[Blamey, 1986] S. Blamey. Partial logic. In D. M. Gabbay and F. Guenthner,
editors, Handbook of Philosophical Logic, Vol. III, pages 1–70. D. Reidel, 1986.
[Bochman, 1994] A. Bochman. On the relation between default and modal
consequence relations. In Proc Int. Conf. on Principles of Knowledge Representation and Reasoning, pages 63–74, 1994.
[Bochman, 1995a] A. Bochman. Default consequence relations as a logical
framework for logic programs. In Proc. Int. Conf. on Logic Programming
and Nonmonotonic Reasoning, pages 245–258. Springer, 1995.
[Bochman, 1995b] A. Bochman. Modal nonmonotonic logics demodalized. Annals of Mathematics and Artificial Intelligence, 15:101–123, 1995.
[Bochman, 1995c] A. Bochman. On bimodal nonmonotonic logics and their
unimodal and non-modal equivalents. In Proc. IJCAI’95, pages 1518–1524,
1995.
[Bochman, 1996a] A. Bochman. On a logical basis of general logic programs. In
Proc. Workshop on Nonmonotonic Extensions of Logic Programming, Lecture
Notes in AI, 1996. Springer Verlag.
[Bochman, 1996b] A. Bochman. On a logical basis of normal logic programs.
Fundamenta Informaticae, 28:223–245, 1996.
[Bochman, 1998a] A. Bochman. A logical foundation for logic programming
I: Biconsequence relations and nonmonotonic completion. Journal of Logic
Programming, 35:151–170, 1998.
67
[Bochman, 1998b] A. Bochman. A logical foundation for logic programming II:
Semantics of general logic programs. Journal of Logic Programming, 35:171–
194, 1998.
[Bochman, 1998c] A. Bochman. On the relation between default and modal
nonmonotonic reasoning. Artificial Intelligence, 101:1–34, 1998.
[Bochman, 2001] A. Bochman. A Logical Theory of Nonomonotonic Inference
and Belief Change. Springer, 2001.
[Bochman, 2003a] A. Bochman. Brave nonmonotonic inference and its kinds.
Annals of Mathematics and Artificial Intelligence, 39:101–121, 2003.
[Bochman, 2003b] A. Bochman. Collective argumentation and disjunctive logic
programming. Journal of Logic and Computation, 9:55–56, 2003.
[Bochman, 2003c] A. Bochman. A logic for causal reasoning. In Proceedings
IJCAI’03, Acapulco, 2003. Morgan Kaufmann.
[Bochman, 2004a] A. Bochman. A causal approach to nonmonotonic reasoning.
Artificial Intelligence, 160:105–143, 2004.
[Bochman, 2004b] A. Bochman. A causal logic of logic programming. In
D. Dubois, C. Welty, and M.-A. Williams, editors, Proc. Ninth Conference on
Principles of Knowledge Representation and Reasoning, KR’04, pages 427–
437, Whistler, 2004.
[Bochman, 2005] A. Bochman. Explanatory Nonmonotonic Reasoning. World
Scientific, 2005.
[Bonatti and Olivetti, 2002] P. A. Bonatti and N. Olivetti. Sequent calculi for
propositional nonmonotonic logics. ACM Trans. Comput. Log., 3:226–278,
2002.
[Bondarenko et al., 1997] A. Bondarenko, P. M. Dung, R. A. Kowalski, and
F. Toni. An abstract, argumentation-theoretic framework for default reasoning. Artificial Intelligence, 93:63–101, 1997.
[Bossu and Siegel, 1985] G. Bossu and P. Siegel. Saturation, nonmonotonic reasoning and the closed-world assumption. Artificial Intelligence, 25:13–63,
1985.
[Brewka and Gottlob, 1997] G. Brewka and G. Gottlob. Well-founded semantics for default logic. Fundamenta Informaticae, 31:221–236, 1997.
[Brewka and Konolige, 1993] G. Brewka and K. Konolige. An abductive framework for general logic programs and other nonmonotonic systems. In Proc.
IJCAI, pages 9–17, 1993.
[Brewka et al., 1997] G. Brewka, J. Dix, and K. Konolige. Nonmonotonic Reasoning: An Overview. CSLI Publications, Stanford, 1997.
68
[Brewka, 1991] G. Brewka. Cumulative default logic: In defense of nonmonotonic inference rules. Artificial Intelligence, 50:183–205, 1991.
[Brewka, 1992] G. Brewka. A framework for cumulative default logics. Technical
Report TR-92-042, ICSI, Berkeley, CA, 1992.
[Cabalar, 2001] P. Cabalar. Well-founded semantics as two-dimensional hereand-there. In Proceedings of ASP 01, 2001 AAAI Spring Symposium Series,
pages 15–20, Stanford, 2001.
[Chen, 1994] J. Chen. Relating only knowing to minimal belief and negation
as failure. Journal of Experimental and Theoretical Artificial Intelligence,
6:409–429, 1994.
[Chesñevar et al., 2000] C. I. Chesñevar, A. G. Marguitman, and R. P. Loui.
Logical models of argument. ACM Computing Surveys, 32:337–383, 2000.
[Clark, 1978] K. Clark. Negation as failure. In H. Gallaire and J. Minker,
editors, Logic and Data Bases, pages 293–322. Plenum Press, 1978.
[Colmerauer et al., 1973] A. Colmerauer, H. Kanoui, R. Pasero, and P. Roussel.
Un systeme de communication hommemachine. Technical report, Groupe de
Intelligence Artificielle Universitae de AixMarseille II, Marseille, 1973.
[Console et al., 1991] L. Console, D. Theseider Dupre, and P. Torasso. On the
relationship between abduction and deduction. Journal of Logic and Computation, 1:661–690, 1991.
[Cox and Pietrzykowski, 1987] P. T. Cox and T. Pietrzykowski. General diagnosis by abductive inference. In Proc. IEEE Symposium on Logic Programming,
pages 183–189, 1987.
[Darwiche and Pearl, 1994] A. Darwiche and J. Pearl. Symbolic causal networks. In Proceedings AAAI’94, pages 238–244, 1994.
[de Kleer et al., 1992] J. de Kleer, A. K. Mackworth, and R. Reiter. Characterizing diagnoses and systems. Artificial Intelligence, 52:197–222, 1992.
[de Kleer, 1986] J. de Kleer. An assumption-based TMS. Artificial Intelligence,
28:127–162, 1986.
[Delgrande, 1987] J. P. Delgrande. A first-order conditional logic for prototypical properties. Artificial Intelligence, 33:105–130, 1987.
[Delgrande, 1988] J. Delgrande. An approach to default reasoning based on a
firstorder conditional logic: revised report. Artificial Intelligence, 36:63–90,
1988.
[Denecker and Schreye, 1992] M. Denecker and D. De Schreye. SLDNFA: an
abductive procedure for normal abductive programs. In Proc. Joint Int.l
Conf. and Symp. on Logic Programming, pages 686–702, 1992.
69
[Denecker et al., 2001] M. Denecker, M. Bruynooghe, and V. W. Marek. Logic
programming revisited: Logic programs as inductive definitions. ACM Trans.
Comput. Log., 2:623–654, 2001.
[Denecker et al., 2003] M. Denecker, V. W. Marek, and M. Truszczyński. Uniform semantic treatment of default and autoepistemic logics. Artificial Intelligence, 143:79–122, 2003.
[Dix, 1991] J. Dix. Classifying semantics of logic programs. In A. Nerode,
W. Marek, and V. S. Subrahmanian, editors, Proceedings 1st International
Workshop on Logic Programming and Nonmonotonic Reasoning, pages 166–
180, Cambridge, Mass., 1991. MIT Press.
[Dix, 1992] J. Dix. Classifying semantics of disjunctive logic programs. In
K. Apt, editor, Proc. Joint Int. Conf. and Symp. on Logic Programming,
ICSLP’92, pages 798–812. MIT Press, 1992.
[Doherty et al., 1995] P. Doherty, W. Lukaszewicz, and A. Szalas. Computing
circumscription revisited: Preliminary report. In Proceedings Int. Joint Conf.
on Artificial Intelligence, pages 1502–1508, 1995.
[Doyle, 1979] J. Doyle. A truth maintenance system. Artificial Intelligence,
12:231–272, 1979.
[Doyle, 1994] J. Doyle. Reasoned assumptions and rational psychology. Fundamenta Informaticae, 20:3573, 1994.
[Dung and Son, 2001] P. M. Dung and T. C. Son. An argument-based approach
to reasoning with specificity. Artificial Intelligence, 133:35–85, 2001.
[Dung et al., 2006] P. M. Dung, R. A. Kowalski, and F. Toni. Dialectic proof
procedures for assumption-based, admissible argumentation. Artificial Intelligence, 170:114159, 2006.
[Dung, 1992] P. M. Dung. Acyclic disjunctive programs with abductive procedure as proof procedure. In Proc. Int. Conference on Fifth Generation
Computer Systems, pages 555–561. ICOT, 1992.
[Dung, 1995a] P. M. Dung. An argumentation-theoretic foundation for logic
programming. J. of Logic Programming, 22:151–177, 1995.
[Dung, 1995b] P. M. Dung. On the acceptability of arguments and its fundamental role in non-monotonic reasoning, logic programming and n-persons
games. Artificial Intelligence, 76:321–358, 1995.
[Dunn, 1976] J. M. Dunn. Intuitive semantics for first-degree entailment and
coupled trees. Philosophical Studies, 29:149–168, 1976.
[Eiter and Lukasiewicz, 2004] T. Eiter and T. Lukasiewicz. Complexity results
for explanations in the structural-model approach. Artificial Intelligence,
154(1-2):145–198, April 2004.
70
[Elkan, 1990] C. Elkan. A rational reconstruction of nonmonotonic truth maintenance systems. Artificial Intelligence, 43:219–234, 1990.
[Erdem and Lifschitz, 2003] E. Erdem and V. Lifschitz. Tight logic programs.
Theory and Practice of Logic Programming, 3:499–518, 2003.
[Etherington and Reiter, 1983] D. Etherington and R. Reiter. On inheritance
hierarchies with exceptions. In Proceedings of AAAI-83, pages 104–108, 1983.
[Etherington et al., 1985] D. Etherington, R. Mercer, and R. Reiter. On the
adequacy of predicate circumscription for closed world reasoning. Computational Intelligence, 1:11–15, 1985.
[Etherington et al., 1991] D. Etherington, S. Kraus, and D. Perlis. Nonmonotonicity and the scope of reasoning. Artificial Intelligence, 52:221–261, 1991.
[Etherington, 1987] D. Etherington. Reasoning with Incomplete Information.
Research Notes in AI. Pitman, London, 1987.
[Fages, 1994] F. Fages. Consistency of Clark’s completion and existence of stable models. Journal of Methods of Logic in Computer Science, 1:51–60, 1994.
[Finger, 1987] J. J. Finger. Exploiting Constraints in Design Synthesis. Ph.D.
dissertation, Department of Computer Science, Stanford University, Stanford,
California, 1987.
[Fitting et al., 1992] M. C. Fitting, W. Marek, and M. Truszczyński. The pure
logic of necessitation. Journal of Logic and Computation, 2:349–373, 1992.
[Fitting, 1991] M. C. Fitting. Bilattices and the semantics of logic programming.
Journal of Logic Programming, 11:91–116, 1991.
[Gabbay et al., 1994] D. Gabbay, C. J. Hogger, and J. A. Robinson, editors.
Handbook of Logic in Artificial Intelligence and Logic Programming. Volume
3: Nonmonotonic Reasoning and Uncertain Reasoning. Oxford UP, Oxford,
1994.
[Gabbay, 1985] D. M. Gabbay. Theoretical foundations for non-monotonic reasoning in expert systems. In K. R. Apt, editor, Logics and Models of Concurrent Systems. Springer, 1985.
[Gabbay, 1998] D. M. Gabbay. Fibring Logics. Oxford University Press, 1998.
[Gärdenfors and Makinson, 1994] P. Gärdenfors and D. Makinson. Nonmonotonic inference based on expectations. Artificial Intelligence, 65:197–245,
1994.
[Geffner, 1992] H. Geffner. Default Reasoning. Causal and Conditional Theories. MIT Press, 1992.
71
[Gelfond and Lifschitz, 1988] M. Gelfond and V. Lifschitz. The stable model
semantics for logic programming. In R. Kowalski and K. Bowen, editors,
Proc. 5th International Conf./Symp. on Logic Programming, pages 1070–
1080, Cambridge, MA, 1988. MIT Press.
[Gelfond and Lifschitz, 1991] M. Gelfond and V. Lifschitz. Classical negation
and disjunctive databases. New Generation Computing, 9:365–385, 1991.
[Gelfond et al., 1991] M. Gelfond, V. Lifschitz, H. Przymusińska, and
M. Truszczyński. Disjunctive defaults. In Proc. Second Int. Conf. on Principles of Knowledge Representation and Reasoning, KR’91, pages 230–237,
Cambridge, Mass., 1991.
[Gelfond, 1994] M. Gelfond. Logic programming and reasoning with incomplete
information. Annals of Mathematics and Artificial Intelligence, 12:89–116,
1994.
[Ginsberg, 1987] M. L. Ginsberg, editor. Readings in Nonmonotonic Reasoning.
Morgan Kaufmann, Los Altos, Ca., 1987.
[Ginsberg, 1989] M. L. Ginsberg. A circumscriptive theorem prover. Artif.
Intell., 39:209–230, 1989.
[Ginsberg, 1993] M. Ginsberg. Essentials of Artificial Intelligence. Morgan
Kaufmann, 1993.
[Giunchiglia et al., 2004] E. Giunchiglia, J. Lee, V. Lifschitz, N. McCain, and
H. Turner. Nonmonotonic causal theories. Artificial Intelligence, 153:49–104,
2004.
[Goldszmidt et al., 1993] M. Goldszmidt, P. Morris, and J. Pearl. A maximum
entropy approach to nonmonotonic reasoning. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 15:220–232, 1993.
[Gottlob, 1994] G. Gottlob. The power of beliefs or translating default logic into
standard autoepistemic logic. In G. Lakemeyer and B. Nebel, editors, Foundations of Knowledge Representation and Reasoning, volume 810 of LNAI,
pages 133–144. Springer, 1994.
[Grosof, 1991] B. N. Grosof. Generalising prioritization. In J. Allen, R. Fikes,
and E. Sandewall, editors, Proc. Second International Conference on Principles of Knowledge Representation and Reasoning (KR’91), pages 289–300.
Morgan Kaufmann, 1991.
[Halpern and Moses, 1985] Joseph Y. Halpern and Yoram Moses. Towards a
theory of knowledge and ignorance. In Krzysztof R. Apt, editor, Logics and
Models of Concurrent Systems, pages 459–476. Springer-Verlag, Berlin, 1985.
72
[Halpern and Pearl, 2001a] J. Y. Halpern and J. Pearl. Causes and explanations: A structral-model approach—part I: Causes. In Proc. Seventh Conf.
On Uncertainty in Artificial Intelligence (UAI’01), pages 194–202, San Francisco, CA, 2001. Morgan Kaufmann.
[Halpern and Pearl, 2001b] J. Y. Halpern and J. Pearl. Causes and explanations: A structral-model approach—part II: Explanations. In Proceedings Int.
Joint Conf. on Artificial Intelligence, IJCAI-01. Morgan Kaufmann, 2001.
[Halpern, 1997] J. Y. Halpern. A critical reexamination of default logic, autoepistemic logic, and only knowing. Computational Intelligence, 13:144–163,
1997.
[Halpern, 1998] J. Y. Halpern. Axiomatizing causal reasoning. In G. F. Cooper
and S. Moral, editors, Uncertainty in Artificial Intelligence, pages 202–210.
Morgan Kaufmann, 1998.
[Hanks and McDermott, 1987] S. Hanks and D. McDermott. Non-monotonic
logics and temporal projection. Artificial Intelligence, 33:379–412, 1987.
[Hewitt, 1969] C. E. Hewitt. PLANNER: A language for proving theorems in
robots. In First International Joint Conference on Artificial Intelligence,
pages 295–301, 1969.
[Horty, 1994] J. F. Horty. Some direct theories of nonmonotonic inheritance. In
D. M. Gabbay, C. J. Hogger, and J. A. Robinson, editors, Handbook of Logic
in Artificial Intelligence and Logic Programming 3: Nonmonotonic Reasoning
and Uncertain Reasoning. Oxford University Press, Oxford, 1994.
[Imielinski, 1987] T. Imielinski. Results on translating defaults to circumscription. Artificial Intelligence, 32:131–146, 1987.
[Inoue and Sakama, 1994] K. Inoue and C. Sakama. On positive occurrences
of negation as failure. In Proc. 4th Int. Conf. on Principles of Knowledge
Representation and Reasoning, KR’94, pages 293–304. Morgan Kauffman,
1994.
[Inoue and Sakama, 1998] K. Inoue and C. Sakama. Negation as failure in the
head. Journal of Logic Programming, 35:39–78, 1998.
[Israel, 1980] D. J. Israel. What’s wrong with non-monotonic logic. In Proceedings of the First Annual Conference of the American Association for Artificial
Intelligence, Stanford University, 1980.
[Iwasaki and Simon, 1986] Yumi Iwasaki and Herbert Simon. Causality in device behavior. Artificial Intelligence, 29(1):3–32, 1986.
[Jakobovits and Vermeir, 1999] H. Jakobovits and D. Vermeir. Robust semantics for argumentation frameworks. Journal of Logic and Computation, 9:215–
261, 1999.
73
[Janhunen, 1996] T. Janhunen. Representing autoepistemic introspection in
terms of default rules. In Proc. European Conf. on AI, ECAI-96, pages 70–
74, Budapest, 1996. John Wiley & Sons.
[Kakas and Mancarella, 1990] A. C. Kakas and P. Mancarella. Generalized stable models: A semantics for abduction. In Proc. European Conf. on Artificial
Intelligence, ECAI-90, pages 385–391, Stockholm, 1990.
[Kakas and Toni, 1999] A. C. Kakas and F. Toni. Computing argumentation in
logic programming. Journal of Logic and Computation, 9:515–562, 1999.
[Kakas et al., 1992] A. C. Kakas, R. A. Kowalski, and F. Toni. Abductive logic
programming. Journal of Logic and Computation, 2:719–770, 1992.
[Konolige and Myers, 1989] K. Konolige and K. L. Myers. Representing defaults
with epistemic concepts. Computational Intelligence, 5:32–44, 1989.
[Konolige, 1988] K. Konolige. On the relation between default and autoepistemic logic. Artificial Intelligence, 35:343–382, 1988.
[Konolige, 1989] K. Konolige. On the relation between autoepistemic logic and
circumscription. In Proc. IJCAI-89, pages 1213–1218, 1989.
[Konolige, 1992] K. Konolige. Abduction versus closure in causal theories. Artificial Intelligence, 53:255–272, 1992.
[Konolige, 1994] K. Konolige. Using default and causal reasoning in diagnosis.
Annals of Mathematics and Artificial Intelligence, 11:97–135, 1994.
[Kowalski and Toni, 1996] R. A. Kowalski and F. Toni. Abstract argumentation. Artificial Intelligence and Law, 4:275–296, 1996.
[Kratzer, 1981] A. Kratzer. Partition and revision: The semantics of counterfactuals. Journal of Philosophical Logic, 10:201–216, 1981.
[Kraus et al., 1990] S. Kraus, D. Lehmann, and M. Magidor. Nonmonotonic
reasoning, preferential models and cumulative logics. Artificial Intelligence,
44:167–207, 1990.
[Lee and Lifschitz, 2003] J. Lee and V. Lifschitz. Loop formulas for disjunctive
logic programs. In Proc. Nineteenth Int. Conference on Logic Programming,
pages 451–465, 2003.
[Lee and Lin, 2004] J. Lee and F. Lin. Loop formulas for circumscription. In
Proc AAAI-04, pages 281–286, San Jose, CA, 2004.
[Lehmann and Magidor, 1992] D. Lehmann and M. Magidor. What does a conditional knowledge base entail? Artificial Intelligence, 55:1–60, 1992.
74
[Lehmann, 1989] D. Lehmann. What does a conditional knowledge base entail?
In R. Brachman and H. J. Levesque, editors, Proc. 2nd Int. Conf. on Principles of Knowledge Representation and Reasoning, KR’89, pages 212–222.
Morgan Kaufmann, 1989.
[Lehmann, 1995] D. Lehmann. Another perspective on default reasoning. Annals of Mathematics and Artificial Intelligence, 15:61–82, 1995.
[Levesque and Lakemeyer, 2000] H. Levesque and G. Lakemeyer. The Logic of
Knowledge Bases. The MIT Press, 2000.
[Levesque, 1989] H. J. Levesque. A knowledge-level account of abduction. In
Proc. IJCAI, pages 1061–1067, 1989.
[Levesque, 1990] H. J. Levesque. All I know: A study in autoepistemic logic.
Artificial Intelligence, 42:263309, 1990.
[Lewis, 1973] D. Lewis. Counterfactuals. Harvard University Press, Cambridge,
Mass., 1973.
[Lewis, 1981] D. Lewis. Ordering semantics and premise semantics for counterfactuals. Journal of Philosophical Logic, 10:217–234, 1981.
[Lifschitz and Schwarz, 1993] V. Lifschitz and G. Schwarz. Extended logic programs as autoepistemic theories. In L. M. Pereira and A. Nerode, editors,
Proc. Second Int. Workshop on Logic Programming and Nonmonotonic Reasoning, pages 101–114. MIT Press, 1993.
[Lifschitz and Woo, 1992] V. Lifschitz and T. Woo. Answer sets in general nonmonotonic reasoning (preliminary report). In Proc. Third Int. Conf. on Principles of Knowledge Representation and Reasoning, KR‘92, pages 603–614.
Morgan Kauffman, 1992.
[Lifschitz et al., 1999] V. Lifschitz, L. R. Tang, and H. Turner. Nested expressions in logic programs. Annals of Mathematics and Artificial Intelligence,
25:369–389, 1999.
[Lifschitz et al., 2001] V. Lifschitz, D. Pearce, and A. Valverde. Strongly equivalent logic programs. ACM Transactions on Computational Logic, 2:526–541,
2001.
[Lifschitz, 1985] V. Lifschitz. Computing circumscription. In Proc. 9th Int.
Joint Conf. on Artificial Intelligence, IJCAI-85, pages 121–127. Morgan
Kaufmann, 1985.
[Lifschitz, 1987a] V. Lifschitz. Pointwise circumscription. In M. Ginsberg, editor, Readings in Non-Monotonic Reasoning, pages 179–193. Morgan Kaufmann, San Mateo, CA, 1987.
75
[Lifschitz, 1987b] Vladimir Lifschitz. Formal theories of action: Preliminary
report. In John McDermott, editor, Proceedings of the Tenth International
Joint Conference on Artificial Intelligence, Los Altos, California, 1987. Morgan Kaufmann.
[Lifschitz, 1991a] V. Lifschitz, editor. Artificial Intelligence and Mathematical
Theory of Computation, Papers in Honor of John McCarthy. Academic Press,
1991.
[Lifschitz, 1991b] V. Lifschitz. Nonmonotonic databases and epistemic queries.
In Proceedings Int. Joint Conf. on Artificial Intelligence, IJCAI-91, pages
381–386. Morgan Kaufmann, 1991.
[Lifschitz, 1994a] V. Lifschitz. Minimal belief and negation as failure. Artificial
Intelligence, 70:53–72, 1994.
[Lifschitz, 1994b] Vladimir Lifschitz. Circumscription. In Dov Gabbay, C. J.
Hogger, and J. A. Robinson, editors, Handbook of Logic in Artificial Intelligence and Logic Programming, Volume 3: Nonmonotonic Reasoning and
Uncertain Reasoning, pages 298–352. Oxford University Press, 1994.
[Lifschitz, 1997] V. Lifschitz. On the logic of causal explanation. Artificial
Intelligence, 96:451–465, 1997.
[Lin and Shoham, 1989] F. Lin and Y. Shoham. Argument systems: A uniform
basis for nonmonotonic reasoning. In Proceedings of 1st Intl. Conference
on Principles of Knowledge Representation and Reasoning, pages 245–255,
Stanford, CA, 1989.
[Lin and Shoham, 1992] F. Lin and Y. Shoham. A logic of knowledge and justified assumptions. Artificial Intelligence, 57:271–289, 1992.
[Lin and You, 2002] F. Lin and J.-H. You. Abduction in logic programming:
A new definition and an abductive procedure based on rewriting. Artificial
Intelligence, 140:175–205, 2002.
[Lin and Zhao, 2002] F. Lin and Y. Zhao. ASSAT: Computing answer sets of a
logic program by SAT solvers. In Proceedings AAAI-02, 2002.
[Lin, 1995] F. Lin. Embracing causality in specifying the inderect effect of actions. In Proc. Int. Joint Conf. on Artificial Intelligence, IJCAI-95, pages
1985–1991, Montreal, 1995. Morgan Kaufmann.
[Lin, 1996] F. Lin. Embracing causality in specifying the indeterminate effects
of actions. In Proceedings AAAI-96, pages 670–676, 1996.
[Lobo and Uzcátegui, 1997] J. Lobo and C. Uzcátegui. Abductive consequence
relations. Artificial Intelligence, 89:149–171, 1997.
[Lobo et al., 1992] J. Lobo, J. Minker, and A. Rajasekar. Foundations of Disjunctive Logic Programming. MIT Press, Cambridge, Mass., 1992.
76
[Lukaszewicz, 1990] W. Lukaszewicz. Non-Monotonic Reasoning: Formalization of Commonsense Reasoning. Ellis Horwood, New York, 1990.
[Makinson and van der Torre, 2000] D. Makinson and L. van der Torre. Input/Output logics. Journal of Philosophical Logic, 29:383–408, 2000.
[Makinson, 1989] D. Makinson. General theory of cumulative inference. In
M. Reinfrank, editor, Nonmonotonic Reasoning, volume 346 of LNAI, pages
1–18. Springer, 1989.
[Makinson, 1994] D. Makinson. General patterns in nonmonotonic reasoning. In
D. M. Gabbay and Others, editors, Handbook of Logic in Artificial Intelligence
and Logic Programming, Vol. 3, Nonmonotonic and Uncertain Reasoning,
volume 2, pages 35–110. Oxford University Press, Oxford, 1994.
[Makinson, 2003] D. Makinson. Bridges between classical and nonmonotonic
logic. Logic Journal of the IGPL, 11:69–96, 2003.
[Makinson, 2005] D. Makinson. Bridges from Classical to Nonmonotonic Logic.
King’s College Publications, 2005.
[Marek and Subrahmanian, 1992] W. Marek and V. S. Subrahmanian. The relationship between stable, supported, default and autoepistemic semantics for
general logic programs. Theoretical Computer Science, 103:365–386, 1992.
[Marek and Truszczyński, 1989] W. Marek and M. Truszczyński. Relating autoepistemic and default logics. In Int. Conf. on Principles of Knowledge Representation and Reasoning, KR’89, pages 276–288, San Mateo, Calif., 1989.
Morgan Kaufmann.
[Marek and Truszczyński, 1990] W. Marek and M. Truszczyński. Modal logic
for default reasoning. Annals of Mathematics and Artificial Intelligence,
1:275–302, 1990.
[Marek and Truszczyński, 1993] W. Marek and M. Truszczyński. Nonmonotonic Logic, Context-Dependent Reasoning. Springer, 1993.
[Marek et al., 1990] W. Marek, A. Nerode, and J. Remmel. A theory of nonmonotonic rule systems. Annals of Mathematics and Artificial Intelligence,
1:241–273, 1990.
[Marek et al., 1993] V. W. Marek, G. F. Schwarz, and M. Truszchinski. Modal
nonmonotonic logics: ranges, characterization, computation. Journal of
ACM, 40:963–990, 1993.
[McCain and Turner, 1997] N. McCain and H. Turner. Causal theories of action
and change. In Proceedings AAAI-97, pages 460–465, 1997.
77
[McCarthy and Hayes, 1969] J. McCarthy and P. Hayes. Some philosophical
problems from the standpoint of artificial intelligence. In B. Meltzer and
D. Michie, editors, Machine Intelligence, pages 463–502. Edinburg University
Press, Edinburg, 1969.
[McCarthy, 1959] John McCarthy. Programs with common sense. In Proceedings of the Teddington Conference on the Mechanization of Thought Processes,
pages 75–91, London, 1959. Her Majesty’s Stationary Office.
[McCarthy, 1980] J. McCarthy. Circumscription – a form of non-monotonic
reasoning. Artificial Intelligence, 13:27–39, 1980.
[McCarthy, 1986] J. McCarthy. Applications of circumscription to formalizing
common sense knowledge. Artificial Intelligence, 13:27–39, 1986.
[McDermott and Doyle, 1980] D. McDermott and J. Doyle.
logic. Artificial Intelligence, 13:41–72, 1980.
Nonmonotonic
[McDermott, 1982] D. McDermott. Nonmonotonic logic II: Nonmonotonic
modal theories. Journal of the ACM, 29:33–57, 1982.
[McDermott, 1987] Drew McDermott. Critique of pure reason. Computational
Intelligence, 3(3):149–160, 1987.
[Minker, 1982] J. Minker. On indefinite databases and the closed world assumption. In Proceedings of 6th Conference on Automated Deduction, pages
292–308, New York, 1982.
[Minker, 1993] Jack Minker. An overview of nonmonotonic reasoning and logic
programming. Journal of Logic Programming, 17:95–126, 1993.
[Minker, 2000] Jack Minker, editor. Logic-Based Artificial Intelligence. Kluwer
Academic Publishers, Dordrecht, 2000.
[Minsky, 1974] M. Minsky. A framework for representing knowledge. Tech.
Report 306, Artificial Intelligence Laboratory, MIT, 1974.
[Moore, 1985] R. C. Moore. Semantical considerations on non-monotonic logic.
Artificial Intelligence, 25:75–94, 1985.
[Morgenstern and Stein, 1994] L. Morgenstern and L. Stein. Motivated action
theory: a formal theory of causal reasoning. Artificial Intelligence, 71(1):1–42,
1994.
[Morgenstern, 1996] L. Morgenstern. The problem with solutions to the frame
problem. In K. M. Ford and Z. Pylyshyn, editors, The Robot’s Dilemma
Revisited: The Frame Problem in Artificial Intelligence, pages 99–133. Ablex
Publishing Co., Norwood, New Jersey, 1996.
[Morris, 1988] P. H. Morris. The anomalous extension problem in default reasoning. Artificial Intelligence, 35:383–399, 1988.
78
[Nayak, 1994] P. P. Nayak.
70:277–334, 1994.
Causal approximations.
Artificial Intelligence,
[Niemelä, 1992] I. Niemelä. A unifying framework for nonmonotonic reasoning.
In Proc. 10th European Conference on Artificial Intelligence, ECAI-92, pages
334–338, Vienna, 1992. John Wiley.
[Oikarinen and Janhunen, 2005] E. Oikarinen and T. Janhunen. circ2dlp translating circumscription into disjunctive logic programming. In C. Baral,
G. Greco, N. Leone, and G. Terracina, editors, Logic Programming and
Nonmonotonic Reasoning, 8th International Conference, LPNMR 2005, Diamante, Italy, September 5-8, 2005, Proceedings, volume 3662 of Lecture Notes
in Computer Science, pages 405–409. Springer, 2005.
[Pearce, 1997] D. Pearce. A new logical characterization of stable models and
answer sets. In J. Dix, L. M. Pereira, and T. Przymusinski, editors, NonMonotonic Extensions of Logic Programming, volume 1216 of LNAI, pages
57–70. Springer, 1997.
[Pearl, 1990] J. Pearl. System Z: A natural ordering of defaults with tractable
applications to default reasoning. In Proceedings of the Third Conference on
Theoretical Aspects of Reasoning About Knowledge (TARK’90), pages 121–
135, San Mateo, CA, 1990. Morgan Kaufmann.
[Pearl, 2000] J. Pearl. Causality. Cambridge UP, 2000.
[Perlis, 1986] D. Perlis. On the consistency of commonsense reasoning. Computational Intelligence, 2:180–190, 1986.
[Poole, 1985] D. Poole. On the comparison of theories: Preferring the most
specific explanation. In Proceedings Ninth International Joint Conference on
Artificial Intelligence, pages 144–147, Los Angeles, 1985.
[Poole, 1988a] D. Poole. A logical framework for default reasoning. Artificial
Intelligence, 36:27–47, 1988.
[Poole, 1988b] D. Poole. Representing knowledge for logic-based diagnosis. In
Proc. Int. Conf. on Fifth Generation Computer Systems, pages 1282–1290,
Tokyo, 1988.
[Poole, 1989a] D. Poole. Explanation and prediction: An architecture for default and abductive reasoning. Computational Intelligence, 5:97–110, 1989.
[Poole, 1989b] D. Poole. What the lottery paradox tells us about default reasoning. In R. J. Brachman, H. J. Levesque, and R. Reiter, editors, Proceedings
of the First Int. Conf. on Principles of Knowledge Representation and Reasoning, pages 333 – 340. Morgan Kaufmann, 1989.
[Poole, 1994a] D. Poole. Default logic. In D. M. Gabbay, C. J. Hogger, and
J. A. Robinson, editors, Handbook of Logic in Artificial Intelligence and Logic
Programming, volume 3, pages 189–215. Oxford University Press, 1994.
79
[Poole, 1994b] D. Poole. Representing diagnosis knowledge. Annals of Mathematics and Artificial Intelligence, 11:33–50, 1994.
[Przymusinska and Przymusinski, 1994] H. Przymusinska and T. Przymusinski.
Stationary default extensions. Fundamenta Informaticae, 21:67–87, 1994. (In
print).
[Przymusinski, 1988] T. C. Przymusinski. On the declarative semantics of stratified deductive databases and logic programs. In J. Minker, editor, Foundations of Deductive Databases and Logic Programming, pages 193–216. Morgan
Kauffman, 1988.
[Przymusinski, 1990] T. C. Przymusinski. The well-founded semantics coincides
with the three-valued stable semantics. Fundamenta Informaticae, 13:445–
464, 1990.
[Przymusinski, 1991a] T. C. Przymusinski. Semantics of disjunctive logic programs and deductive databases. In Proc. Second Int. Conf. On Deductive and
Object-Oriented Databases, pages 85–107. Springer, 1991.
[Przymusinski, 1991b] T. C. Przymusinski. Three-valued nonmonotonic formalisms and semantics of logic programs. Artificial Intelligence, 49:309–343,
1991.
[Przymusinski, 1994] T. C. Przymusinski. A knowledge representation framework based on autoepistemic logic of minimal beliefs. In Proceedings AAAI94, 1994.
[Ramsey, 1978] F. P. Ramsey. Foundations. Routledge & Kegan Paul, London,
1978.
[Reggia et al., 1985] J. A. Reggia, D. S. Nau, and P. Y. Wang. A formal model
of diagnostic inference. i. problem formulation and decomposition. Inf. Sci.,
37:227–256, 1985.
[Reinfrank et al., 1989] M. Reinfrank, O. Dressler, and G. Brewka. On the
relation between truth maintenance and autoepistemic logic. In Proc. Int.
Joint Conf. on Artificial Intelligence, pages 1206–1212, 1989.
[Reiter and Criscuolo, 1981] R. Reiter and G. Criscuolo. On interacting defaults. In Proceedings of IJCAI81, pages 270–276, 1981.
[Reiter and de Kleer, 1987] R. Reiter and J. de Kleer. Formal foundations for
assumptionbased truth maintenance systems: preliminary report. In Proceedings of AAAI-87, pages 183–188, Seattle, Washington, 1987.
[Reiter, 1978] R. Reiter. On closed world data bases. In H. Gallaire and
J. Minker, editors, Logic and Data Bases, pages 119–140. Plenum Press, 1978.
[Reiter, 1980] R. Reiter. A logic for default reasoning. Artificial Intelligence,
13:81–132, 1980.
80
[Reiter, 1982] R. Reiter. Circumscription implies predicate completion (sometimes). In Proc. AAAI, pages 418–420, 1982.
[Reiter, 1987a] R. Reiter. Nonmonotonic reasoning. Annual Review of Computer Science, 2:147–186, 1987.
[Reiter, 1987b] R. Reiter. A theory of diagnosis from first principles. Artificial
Intelligence, 32:57–95, 1987.
[Reiter, 1991] R. Reiter. The frame problem in the situation calculus: A simple
solution (sometimes) and a completeness result for goal regression. In V. Lifschitz, editor, Artificial Intelligence and Mathematical Theory of Computation:
Papers in Honor of Lohn McCarthy, pages 318–420. Academic Press, 1991.
[Reiter, 2001] R. Reiter. Knowledge in Action: Logical Foundations for Specifying and Implementing Dynamic Systems. MIT Press, 2001.
[Russell, 1957] B. Russell. Mysticism and Logic, chapter On the notion of cause,
pages 180–208. Allen & Unvin, London, 1957.
[Sandewall, 1972] E. Sandewall. An approach to the frame problem and its implementation. In Machine Intelligence, volume 7, pages 195–204. Edinburgh
University Press, 1972.
[Sandewall, 1994] Erik Sandewall. Features and Fluents: A Systematic Approach to the Representation of Knowledge About Dynamical Systems. Oxford
University Press, Oxford, 1994.
[Schlechta, 1997] K. Schlechta. Nonmonotonic Logics: Basic Concepts, Results,
and Techniques, volume 1187 of Lecture Notes in AI. Springer Verlag, 1997.
[Schlechta, 2004] Karl Schlechta. Coherent Systems, volume 2 of Studies in
Logic and Practical Reasoning. Elsevier, Amsterdam, 2004.
[Schubert, 1990] L. Schubert. Monotonic solution of the frame problem in the
situation calculus; an efficient method for worlds with fully specified actions.
In Henry Kyburg, Ronald Loui, and Greg Carlson, editors, Knowledge Representation and Defeasible Reasoning, pages 23–67. Kluwer Academic Publishers, Dordrecht, 1990.
[Schwarz and Truszczyński, 1994] G. Schwarz and M. Truszczyński. Minimal
knowledge problem: a new approach. Artificial Intelligence, 67:113–141, 1994.
[Schwarz, 1990] G. Schwarz. Autoepistemic modal logics. In R. Parikh, editor,
Theoretical Aspects of Reasoning about Knowledge, TARK-90, pages 97–109,
San Mateo, CA, 1990. Morgan Kaufmann.
[Schwarz, 1992a] G. Schwarz. Minimal model semantics for nonmonotonic
modal logics. In Proceedings LICS-92, pages 34–43, Santa Cruz, CA., 1992.
81
[Schwarz, 1992b] G. Schwarz. Reflexive autoepistemic logic. Fundamenta Informaticae, 17:157–173, 1992.
[Schwind, 1999] C. Schwind. Causality in action theories. Linköping Electronic
Articles in Computer and Information Science, 4(4), 1999.
[Segerberg, 1971] K. Segerberg. An Essay in Classical Modal Logic, volume 13
of Filosofiska Studier. Uppsala University, 1971.
[Shanahan, 1997] M. P. Shanahan. Solving the Frame Problem. The MIT Press,
1997.
[Shepherdson, 1988] J. C. Shepherdson. Negation in logic programming. In
J. Minker, editor, Deductive Databases and Logic Programming, pages 19–88.
M. Kaufmann, 1988.
[Shoham, 1988] Y. Shoham. Reasoning about Change. Cambridge University
Press, 1988.
[Simon, 1952] Herbert Simon. On the definition of the causal relation. The
Journal of Philosophy, 49:517–528, 1952.
[Stalnaker, 1968] R. Stalnaker. A theory of conditionals. In N. Rescher, editor,
Studies in Logical Theory. Basil Blackwell, Oxford, 1968.
[Stalnaker, 1993] Robert C. Stalnaker. A note on non-monotonic modal logic.
Artificial Intelligence, 64(2):183–196, 1993. Widely circulated in manuscipt
form, 1980 to 1992.
[Tan and Pearl, 1995] S.-W. Tan and J. Pearl. Specificity and inheritance in
default reasoning. In Proceedings Int. Joint Conf. on Artificial Intelligence,
IJCAI-95, pages 1480–1486, 1995.
[Thiele, 1990] H. Thiele. On generation of cumulative inference operators by
default deduction rules. In Nonmonotonic and Inductive Logic, Lecture Notes
in Computer Science, pages 100–137. Springer, 1990.
[Thielscher, 1997] M. Thielscher. Ramification and causality. Artificial Intelligence, 89:317–364, 1997.
[Thomason, 2003] Richmond Thomason. Logic and artificial intelligence. In
Edward N. Zalta, editor, The Stanford Encyclopedia of Philosophy. 2003.
[Touretzky, 1986] D. S. Touretzky. The Mathematics of of Inheritance Systems.
Morgan Kaufmann, Los Altos, 1986.
[Truszczyński, 1991] M. Truszczyński. Modal interpretations of default logic.
In J. Myopoulos and R. Reiter, editors, Proceedings Int. Joint Conf. on Artificial Intelligence, IJCAI’91, pages 393–398, San Mateo, Calif., 1991. Morgan
Kaufmann.
82
[Turner, 1999] H. Turner. A logic of universal causation. Artificial Intelligence,
113:87–123, 1999.
[van Benthem, 1984] J. van Benthem. Foundations of conditional logic. Journal
of Philosophical Logic, 13:303–349, 1984.
[van Emden and Kowalski, 1976] M. H. van Emden and R. A. Kowalski. The
semantics of predicate logic as a programming language. J. of ACM, 23:733–
742, 1976.
[van Gelder et al., 1991] A. van Gelder, K. A. Ross, and J. S. Schlipf. The wellfounded semantics for general logic programs. J. ACM, 38:620–650, 1991.
[Veltman, 1976] F. Veltman. Prejudices, presuppositions and the theory of conditionals. In J. Groenendijk and M. Stokhof, editors, Amsterdam Papers on
Formal Grammar, volume 1, pages 248–281. Centrale Interfaculteit, Universiteit van Amsterdam, 1976.
[Wittgenstein, 1961] L. Wittgenstein. Tractatus Logico-Philosophicus. Routledge & Kegan Paul, London, 1961. English translation by D. F. Pears and
B. F. McGuinness.
[You et al., 2000] J.-H. You, L. Y. Yuan, and R. Goebel. An abductive apprach
to disjunctive logic programming. Journal of Logic Programming, 44:101–127,
2000.
[Zhang and Foo, 2001] D. Zhang and N. Foo. EPDL: A logic for causal reasoning. In Proc Int. Joint Conf. on Artificial Intelligence, IJCAI-01, pages
131–136, Seattle, 2001. Morgan Kaufmann.
[Zhang and Rounds, 1997] G.-Q. Zhang and W. C. Rounds. Defaults in domain
theory. Theor. Comput. Sci., 177:155–182, 1997.
83