Abduction and Induction - Introduction
Abduction and Induction - Introduction
net/publication/239658957
CITATIONS READS
62 1,112
2 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Antonis C. Kakas on 27 December 2019.
1.1 INTRODUCTION
This collection is devoted to the analysis and application of abductive and inductive
reasoning in a common context, studying their relation and possible ways for integra-
tion. There are several reasons for doing so. One reason is practical, and based on the
expectation that abduction and induction are sufficiently similar to allow for a tight
integration in practical systems, yet sufficiently complementary for this integration to
be useful and productive.
Our interest in combining abduction and induction is not purely practical, however.
Conceptually, the relation between abduction and induction is not well understood.
More precisely, there are several, mutually incompatible ways to perceive this relation.
For instance, Josephson writes that ‘it is possible to treat every good (...) inductive
generalisation as an instance of abduction’ (Josephson, 1994, p.19), while Michalski
has it that ‘inductive inference was defined as a process of generating descriptions that
imply original facts in the context of background knowledge. Such a general definition
includes inductive generalisation and abduction as special cases’ (Michalski, 1987,
p.188).
One can argue that such incompatible viewpoints indicate that abduction and induc-
tion themselves are not well-defined. Once their definitions have been fixed, studying
their relation becomes a technical rather than a conceptual matter. However, it is not
self-evident why there should exist absolute, Platonic ideals of abduction and induc-
tion, waiting to be discovered and captured once and for all by an appropriate defini-
1
2 P.A. FLACH AND A.C. KAKAS
tion. As with most theoretical notions, it is more a matter of pragmatics, of how useful
a particular definition is going to be in a particular context.
A more relativistic viewpoint is often more productive in these matters, looking at
situations where it might be more appropriate to distinguish between abduction and
induction, and also at cases where it seems more useful to unify them. Sometimes
we want to stress that abduction and induction spring from a common root (say hypo-
thetical or non-deductive reasoning), and sometimes we want to take a finer grained
perspective by looking at what distinguishes them (e.g. the way in which the hypothe-
sis extends our knowledge). The following questions will therefore be our guidelines:
When and how will it be useful to unify, or distinguish, abduction and induction?
How can abduction and induction be usefully integrated?
Peirce’s syllogistic theory. In Peirce’s days logic was not nearly as well-developed
as it is today, and his first attempt to classify arguments (which he considers ‘the
chief business of the logician’ (2.619), follows Aristotle in employing syllogisms.
The following syllogism is known as Barbara:
All the beans from this bag are white;
these beans are from this bag;
therefore, these beans are white.
The idea is that this valid argument represents a particular instantiation of a reason-
ing scheme, and that any alternative instantiation represents another argument that is
likewise valid. Syllogisms should thus be interpreted as argument schemas.
Two other syllogisms are obtained from Barbara if we exchange the conclusion (or
Result, as Peirce calls it) with either the major premiss (the Rule) or the minor premiss
(the Case):
Case. — These beans are from this bag.
Result. — These beans are white.
Rule. — All the beans from this bag are white.
1 Referencesto Peirce’s collected papers take the form X :Y , where X denotes the volume number and Y the
paragraph within the volume.
6 P.A. FLACH AND A.C. KAKAS
and the result) Peirce calls making a hypothesis or, briefly, hypothesis – the term
‘abduction’ is introduced only in his later theory. 2
8 Deductive or Analytic
Peirce thus arrives at the following classification of inference (2.623):
< Induction
Inference
: Synthetic
Hypothesis
Comparing this classification with the one obtained in Section 1.2.1, we can point
out the following similarities. That what was called induction previously corresponds
to what Peirce calls synthetic inference (another term he uses is ampliative reason-
ing, since it amplifies, or goes beyond, the information contained in the premisses).
Furthermore, what Peirce calls induction corresponds to what we called inductive gen-
eralisation in Section 1.2.1.3
On the other hand, the motivations for these classifications are quite different in
each case. In Section 1.2.1 we were concentrating on the different kinds of support or
confirmation that arguments provide, and we noticed that this is essentially the same
for all non-deductive reasoning. When we concentrate instead on the syllogistic form
of arguments, we find this to correspond more naturally to a trichotomy, separating
non-deductive reasoning into two subcategories. As Horn clause logic is in some sense
a modern upgrade of syllogistic logic, it is perhaps not surprising that the distinction
between abduction and induction in logic programming follows Peirce’s syllogistic
classification to a large extent. This will be further taken up in Section 1.3.
Peirce’s inferential theory. In his later theory of reasoning Peirce abandoned the
idea of a syllogistic classification of reasoning:
‘(...) I was too much taken up in considering syllogistic forms and the doctrine
of logical extension and comprehension, both of which I made more fundamental
than they really are. As long as I held that opinion, my conceptions of Abduc-
tion necessarily confused two different kinds of reasoning.’ (Peirce, 1958, 2.102,
written in 1902)
Instead, he identified the three reasoning forms – abduction, deduction and induction
– with the three stages of scientific inquiry: hypothesis generation, prediction, and
evaluation (Figure 1.1). The underlying model of scientific inquiry runs as follows.
When confronted with a number of observations she seeks to explain, the scientist
comes up with an initial hypothesis; then she investigates what other consequences
this theory, were it true, would have; and finally she evaluates the extent to which
these predicted consequences agree with reality. Peirce calls the first stage, coming up
with a hypothesis to explain the initial observations, abduction; predictions are derived
from a suggested hypothesis by deduction; and the credibility of that hypothesis is
estimated through its predictions by induction. We will now take a closer look at these
stages.
2 Peirce also uses the term ‘retroduction’, a translation of the Greek word απαγωγή used by Aristotle (trans-
statistical versions.
ABDUCTIVE AND INDUCTIVE REASONING: BACKGROUND AND ISSUES 7
hypothesis
deduction
abduction
predictions
induction
observations
R E A L I T Y
Figure 1.1 The three stages of scientific inquiry.
Let us investigate the logical form of abduction given by Peirce a little closer. About
C we know two things: that it is true in the actual world, and that it is surprising. The
latter thing can be modelled in many ways, one of the simplest being the requirement
that C does not follow from our other knowledge about the world. In this volume,
Aliseda models it by an epistemic state of doubt which calls for abductive reasoning
to transform it into a state of belief.
Then, ‘if A were true, C would be a matter of course’ is usually interpreted as ‘A
logically entails C’.4 Peirce calls A an explanation of C, or an ‘explanatory hypothe-
sis’. Whether or not this is an appropriate notion of explanation remains an issue of
4 Note
!
that interpreting the second premiss as a material implication, as is sometimes done in the literature,
renders it superfluous, since the truth of A C follows from the truth of the observation C.
8 P.A. FLACH AND A.C. KAKAS
debate. In this volume, Console and Saitta also propose to identify explanation with
entailment, but Josephson argues against it.
Besides being explanatory, Peirce mentions two more conditions to be fulfilled
by abductive hypotheses: they should be capable of experimental verification, and
they should be ‘economic’. A hypothesis should be experimentally verifiable, since
otherwise it cannot be evaluated inductively. Economic factors include the cost of
verifying the hypothesis, its intrinsic value, and its effect upon other projects (Peirce,
1958, 7.220). In other words, economic factors are taken into account when choosing
the best explanation among the logically possible ones. For this reason, abduction is
often termed ‘inference to the best explanation’ (Lipton, 1991).
Induction is identified by Peirce as the process of testing a hypothesis against reality
through selected predictions. ‘Induction consists in starting from a theory, deducing
from it predictions of phenomena, and observing those phenomena in order to see how
nearly they agree with the theory’ (Peirce, 1958, 5.170). Such predictions can be seen
as experiments:
‘When I say that by inductive reasoning I mean a course of experimental inves-
tigation, I do not understand experiment in the narrow sense of an operation by
which one varies the conditions of a phenomenon almost as one pleases. (...) An
experiment (...) is a question put to nature. (...) The question is, Will this be the
result? If Nature replies ‘No!’ the experimenter has gained an important piece
of knowledge. If Nature says ‘Yes,’ the experimenter’s ideas remain just as they
were, only somewhat more deeply engrained.’ (Peirce, 1958, 5.168)
This view of hypothesis testing is essentially what is called the ‘hypothetico-deductive
method’ in philosophy of science (Hempel, 1966). The idea that a verified prediction
provides further support for the hypothesis is very similar to the notion of confirma-
tion as discussed in Section 1.2.1, and also refutation of hypotheses through falsified
predictions can be brought in line with confirmation theory, with a limiting degree of
support of zero.5 The main difference from confirmation theory is that in the Peircean
view of induction the hypothesis is, through the predictions, tested against selected
pieces of evidence only. This leads to a restricted form of hypothesis evaluation, for
which we will use the term hypothesis testing.
Peirce’s inferential theory makes two main points. It posits a separation between
hypothesis generation and hypothesis evaluation; and it focuses attention on hypothe-
ses that can explain and predict. Combining the two points, abduction is the process of
generating explanatory hypotheses (be they general ‘rules’ or specific ‘cases’, as in the
syllogistic account), and induction corresponds to the hypothetico-deductive method
of hypothesis testing. However, the two points are relatively independent: e.g., we can
perceive generation of non-explanatory hypotheses. We will come back to this point
in the discussion below.
5 From j j
a Bayesian perspective P(H E ) is proportional to P(E H )P(H ), where P(H ) is the prior probability
j
of the hypothesis; if E is contrary to a prediction P(E H ) = 0. See Poole’s chapter for further discussion of
the Bayesian perspective.
ABDUCTIVE AND INDUCTIVE REASONING: BACKGROUND AND ISSUES 9
1.2.3 Discussion
In the previous two sections we have considered three philosophical and logical per-
spectives on how non-deductive reasoning may be categorised: the inductivist view,
which holds that no further categorisation is needed since all non-deductive reasoning
must be justified in the same way by means of confirmation theory; the syllogistic
view, which distinguishes between inductive generalisation on the one hand and hy-
pothesis or abduction as inference of specific ‘cases’ on the other; and the inferential
view, which holds that abduction and induction represent the hypothesis generation
and evaluation phases in explanatory reasoning. As we think that none of these view-
points provides a complete picture, there is opportunity to come to a partial synthesis.
Hypothesis generation and hypothesis evaluation. The most salient point of Peirce’s
later, inferential theory is the distinction between hypothesis generation and hypothe-
sis evaluation. In most other accounts of non-deductive reasoning the actual hypothe-
sis is already present in the argument under consideration, as can be seen clearly from
the argument forms discussed in Section 1.2.1. For instance, when constructing an
inductive generalisation
X percent of observed Fs are Gs;
therefore, (approximately) X percent of all Fs are Gs.
our job is first to conjecture possible instantiations of F and G (hypothesis generation),
and then to see whether the resulting argument has sufficient support (hypothesis eval-
uation).
One may argue that a too rigid distinction between generation and evaluation of
hypotheses is counter-productive, since it would lead to the generation of many, ulti-
mately useless hypotheses. Indeed, Peirce’s ‘economic factors’, to be considered when
constructing possible abductive hypotheses, already blur the distinction to a certain ex-
tent. However, even if a too categorical distinction may have practical disadvantages,
on the conceptual level the dangers of confusing the two processes are much larger.
Furthermore, the distinction will arguably be sharper drawn in artificial reasoning sys-
tems than it is in humans, just as chess playing computers still have no real alternative
to finding useful moves than to consider all possible ones.
In any case, whether tightly integrated or clearly separated, hypothesis generation
and hypothesis evaluation have quite distinct characteristics. Here we would argue that
it is hypothesis generation, being concerned with possibilities rather than choices, that
is most inherently ‘logical’ in the traditional sense. Deductive logic does not help the
mathematician in selecting theorems, only in distinguishing potential theorems from
fallacious ones. Also, as (Hanson, 1958) notes, if hypothesis evaluation establishes
a logic at all, then this would be a ‘Logic of the Finished Research Report’ rather
than a ‘Logic of Discovery’. An axiomatic formalisation of the logic of hypothesis
generation is suggested by Flach in his chapter in this volume.
We also stress the distinction between generation and evaluation because it provides
a useful heuristic for understanding the various positions of participants in the debate
on abduction and induction. This rule of thumb states that those concentrating on
generating hypotheses tend to distinguish between non-deductive forms of reasoning;
those concentrating on evaluating hypotheses tend not to distinguish between them.
10 P.A. FLACH AND A.C. KAKAS
Not only does the rule apply to the approaches discussed in the previous two sections;
we believe that it can guide the reader, by and large, through the chapters in this
collection.
where P(x) denotes a formula with free variable x. Possible instantiations of P(x) can
be found by pretending that there exist no other objects than those in the sample, and
looking for true universal sentences. For instance, we might note that every object in
the sample is either female or male. This approach is further discussed in the chapter
by Lachiche.
generalisations: the rule ‘every parent of John is a parent of John’s brother’ does not
explain parenthood.
In line with recent developments in inductive logic programming, we would like to
suggest that inductive generalisations like these are not explanatory at all. They sim-
ply are generalisations that are confirmed by the sample. The process of finding such
generalisations has been called confirmatory induction (also descriptive induction).
The difference between the two forms of induction can be understood as follows. A
typical form of explanatory induction is concept learning, where we want to learn a
definition of a given concept C in terms of other concepts. This means that our induc-
tive hypotheses are required to explain (logically entail) why particular individuals are
Cs, in terms of the properties they have.
However, in the more general case of confirmatory induction we are not given a
fixed concept to be learned. The aim is to learn relationships between any of the
concepts, with no particular concept singled out. The formalisation of confirmatory
hypothesis formation thus cannot be based on logical entailment, as in Peirce’s ab-
duction. Rather, it is a qualitative form of degree of confirmation, which explains its
name. We will have more to say about the issue in Section 1.3.2.
Abduction. Turning next to abduction, it may seem at first that Peirce’s syllogistic
and inferential definitions are not easily reconcilable. However, it is possible to per-
ceive a similarity between the two when we notice that the early syllogistic view of
abduction or hypothesis (p. 5) provides a special form of explanation. The Result (tak-
ing the role of the observation) is explained by the Case in the light of the Rule as a
given theory. The syllogistic form of abduction can thus be seen to meet the explana-
tory requirement of the later inferential view of abduction. Hence we can consider
explanation as a characterising feature of abduction. This will be further discussed in
Section 1.3.2.
Even if the syllogistic and inferential view of abduction can thus be reconciled, it is
still possible to distinguish between approaches which are primarily motivated by one
of the two views. The syllogistic account of abduction has been taken up, by and large,
in logic programming and other work in artificial intelligence addressing tasks such
as that of diagnosis and planning. In this volume, the logic programming perspective
on abduction can be found in the contributions by Christiansen, Console and Saitta,
Inoue and Haneda, Mooney, Poole, Lamma et al., Sakama, and Yamamoto. The logic
programming and artificial intelligence perspective will be more closely examined in
the next section. On the other hand, the chapters by Aliseda, Josephson, and Psillos
are more closely related to the inferential perspective on abduction.
earlier, syllogistic theory. In Section 1.3.2 we argue that abductive hypotheses primar-
ily provide explanations, while inductive hypotheses provide generalisations. We then
further investigate abduction and induction from a logical perspective in Section 1.3.3,
pointing out differences in the way in which they extend incomplete theories. In Sec-
tion 1.3.4 we investigate how more complex reasoning patterns can be viewed as being
built up from simple abductive and inductive inferences. Finally, in Section 1.3.5 we
address the computational characteristics of abduction and induction.
accounted for by the theory alone. 6 If our assessment of the distinction between ab-
duction and induction that is usually drawn in AI is correct, we must conclude that
the above specifications are unable to account for this distinction. In the remainder of
Section 1.3 we will try to understand the differences between abduction and induction
as used in AI in modern, non-syllogistic terms. For an account which stays closer to
syllogisms, the reader is referred to the chapter by Wang.
6 Extraelements that are often added to the above definitions are the satisfaction of integrity constraints for
the case of abduction, and the avoidance of negative examples for the case of induction; these can again be
viewed under the same heading, namely as being aimed at exclusion of certain hypotheses.
ABDUCTIVE AND INDUCTIVE REASONING: BACKGROUND AND ISSUES 15
foreground knowledge may also be used. In some cases it may be empty, for instance
when we are learning the definition of a recursive predicate, when we are learning
the definitions of several mutually dependent predicates, or when we are doing data
mining. The observations specify incomplete (usually extensional) knowledge about
the observables, which we try to generalise into new foreground knowledge.
On the other hand, in abduction we are inferring instance knowledge from ob-
servations and other known information. The latter necessarily contains foreground
information pertaining to the observations at hand. Possible abductive hypotheses are
built from specific non-observable predicates called abducibles in ALP. The intuition
is that these are the predicates of which the extensions are not completely known as in-
stance knowledge. Thus, an abductive hypothesis is one which completes the instance
knowledge about an observed individual. This difference between the effect of abduc-
tion and induction on observable and instance knowledge is studied in the chapter by
Console and Saitta.
puts it, inductive hypotheses do not explain particular observations, but they explain
the frequencies with which the observations occur (viz. that non-white beans from this
bag are never observed).
Generalisation. We thus find that inductive hypotheses are not explanatory in the
same way as abductive hypotheses are. But we would argue that being explanatory is
not the primary aim of inductive hypotheses in the first place. Rather, the main goal of
induction is to provide generalisations. In this respect, we find that the ILP definition
of induction (p. 13) is too much focused on the problem of learning classification rules,
without stressing the aspect of generalisation. An explanatory hypothesis would only
be inductive if it generalises. The essential aspect of induction as applied in AI seems
to be the kind of sample-to-population inference exemplified by categorical inductive
generalisation, reproduced here in its more general form from Section 1.2.3:
All objects in the sample satisfy P(x);
therefore, all objects in the population satisfy P(x).
As with Peirce’s syllogisms, the problem here is that P(x) is already assumed to be
given, while in AI a major problem is to generate such hypotheses. The specification
of confirmatory or descriptive induction follows this pattern, but leaves the hypothesis
unspecified:
Given a consistent set of observations O and a consistent background knowledge
B, find a hypothesis H such that: M (B O) j= H
(Helft, 1989; De Raedt and Bruynooghe, 1993; Flach, 1995)
Hence the formal requirement now is that any generated hypothesis should be true
in a certain model constructed from the given knowledge and observations (e.g. the
truth-minimal model).
This specification can be seen as sample-to-population inference. For example, in
Peirce’s bean example (p. 5), B is ‘these beans are from this bag’ (instance knowledge),
O is ‘these beans are white’ (observation), and H – ‘all the beans from this bag are
white’ – is satisfied by the model containing ‘these beans’ as the only beans in the
universe. Under the assumption that the population is similar to the sample, we achieve
generalisation by restricting attention to formulae true in the sample. Note that the
induced hypothesis is not restricted to one explaining the whiteness of these beans:
we might equally well have induced that ‘all white beans are from this bag’.
Above we defined a hypothesis as generalising if it makes a prediction involving
an observable. We have to qualify this statement somewhat, as the following example
shows (taken from the chapter by Console and Saitta, Example ??, p. ??). Let our
background theory contain the following clauses:
measles(X):-brother(X,Y),measles(Y).
red_spots(X):-measles(X).
brother(john,dan).
red spots(dan). Thus, the hypothesis that John has measles also seems to qualify
as a generalisation. We would argue however that this generalisation effect is already
present in the background theory. On the other hand, an inductive hypothesis produces
a genuinely new generalisation effect, in the sense that we can find new individuals for
which the addition of the hypothesis to our knowledge is necessary to derive some ob-
servable property for these individuals (usually this property is that of the observations
on which the induction was based). With an abductive hypothesis this kind of exten-
sion of the observable property to other new individuals does not necessarily require
the a priori addition of the abductive hypothesis to the theory but depends only on the
properties of this individual and the given background theory: the generalisation, if
any, already exists in the background theory.
We conclude that abductive and inductive hypotheses differ in the degree of gen-
eralisation that each of them produces. With the given background theory T we im-
plicitly restrict the generalising power of abduction as we require that the basic model
of our domain remains that of T . The existence of this theory separates two levels
of generalisation: (a) that contained in the theory and (b) new generalisations that are
not given by the theory. In abduction we can only have the first level with no in-
terest in genuinely new generalisations, while in induction we do produce such new
generalisations.
ties for this new individual.7 Given an abductive theory T as above, the process of
abduction is to select one of the abductive extensions T (Δ) of T in which the given
observation to be explained holds, by selecting the corresponding formula Δ. We can
then reason deductively in T (Δ) to arrive at other conclusions. By selecting Δ we are
essentially enabling one of the possible associations between Δ and the observation
among those supplied by the theory T .
It is important here to emphasise that the restriction of the hypothesis of abduction
to abducible predicates is not incidental or computational, but has a deeper representa-
tional reason. It reflects the relative comprehensiveness of knowledge of the problem
domain contained in T . The abducible predicates and the allowed abductive formu-
lae take the role of ‘answer-holders’ for the problem goals that we want to set to our
theory. In this respect they take the place of the logical variable as the answer-holder
when deductive reasoning is used for problem solving. As a result this means that the
form of the abductive hypothesis depends heavily on the particular theory T at hand,
and the way that we have chosen to represent in this our problem domain.
Typically, the allowed abducible formulae are further restricted to simple logical
forms such as ground or existentially quantified conjunctions of abducible literals.
Although these further restrictions may be partly motivated by computational consid-
erations, it is again important to point out that they are only made possible by the
relative comprehensiveness of the particular representation of our problem domain in
the theory T . Thus, the case of simple abduction – where the abducible hypothesis
are ground facts – occurs exactly because the representation of the problem domain in
T is sufficiently complete to allow this. Furthermore, this restriction is not significant
for the purposes of comparison of abduction and induction: our analysis here is inde-
pendent of the particular form of abducible formulae. The important elements are the
existence of an enumeration of the abductive formulae, and the fact that these do not
involve observable predicates.
Inductive extensions. Let us now turn to the case of induction and analyse this
process to facilitate comparison with the process of abduction as described above.
Again, we have a collection of possible inductive hypotheses from which one must be
selected. The main difference now is the fact that these hypotheses are not limited to
a particular subset of predicates that are incompletely specified in the representation
of our problem domain by the theory T , but are restricted only by the language of T .
In practice, there may be a restriction on the form of the hypothesis, called language
bias, but this is usually motivated either by computational considerations, or by other
information external to the theory T that guides us to an inductive solution.
Another essential characteristic of the process of induction concerns the role of the
selected inductive hypothesis H. The role of H is to extend the existing theory T to a
new theory T 0 = T H, rather than reason with T under the set of assumptions H as is
the case for abduction. Hence T is replaced by T 0 to become a new theory with which
we can subsequently reason, either deductively of abductively, to extract information
7 Note that this type of abductive (or open) reasoning with a theory T collapses to deduction, when and if
from it. The hypothesis H changes T by requiring extra conditions on the observable
predicates that drive the induction, unlike abduction where the extra conditions do not
involve the observable predicates. In effect, H provides the link between observables
and non-observables that was missing or incomplete in the original theory T .
Analogously to the concept of abductive extension, we can define inductive ex-
tensions as follows. Consider a common given theory T with which we are able to
perform abduction and induction. That is, T has a number of abductive extensions
T (Δ). Choosing an inductive hypothesis H as a new part of the theory T has the ef-
fect of further conditioning each of the abductive extensions T (Δ). Hence, while in
abduction we select an abductive extension of T , with induction we extend each of
the abductive extensions with H. The effect of induction is thus ‘universal’ on all the
abductive extensions.
If we now consider the new abductive theory T 0 = T H, constructed by induction,
we can view induction as a process of selecting a collection of abductive extensions,
namely those of the new theory T 0 . Hence an inductive extension can be viewed as a
set of abductive extensions of the original theory T that are further (uniformly) condi-
tioned by the common statement of the inductive hypothesis H. This idea of an induc-
tive extension consisting of a set of abductive extensions was used in (Denecker et al.,
1996) to obtain a formalisation of abduction and induction as selection processes in
a space of possible world models over the given theory in each case. In this way the
process of induction can be seen to have a more general form than abduction, able to
select a set of extensions rather than a single one. Note that this does not necessar-
ily mean that induction will yield a more general syntactic form of hypotheses than
abduction.
Analysis . Comparing the possible inductive and abductive extensions of a given the-
ory T we have an essential difference. In the case of abduction some of the predicates
in the theory, namely the observables, cannot be arbitrarily defined in an extension.
The freedom of choice of abduction is restricted to constrain directly (via Δ) only the
abducibles of the theory. The observable predicates cannot be affected except through
the theory: the observables must be grounded in the existing theory T by the choice of
the abductive conditions on the abducible part of the extension. Hence in an abductive
extension the extent to which the observables can become true is limited by the theory
T and the particular conditions Δ on the rest of the predicates.
In induction this restriction is lifted, and indeed we can have inductive extensions
of the given theory T , the truthvalue of which on the observable predicates need not
be attributed via T to a choice on the abducibles. The inductive extensions ‘induce’
a more general change (from the point of view of the observables) on the existing
theory T , and – as we will see below – this will allow induction to genuinely gener-
alise the given observations to other cases not derivable from the original theory T .
The generalising effect of abduction, if at all present, is much more limited. The se-
lected abductive hypothesis Δ may produce in T (Δ) further information on abducible
or other predicates, as in the measles example from the previous section. Assuming
that abducibles and observables are disjoint, any information on an observable derived
in T (Δ) is a generalisation already contained in T .
20 P.A. FLACH AND A.C. KAKAS
What cannot happen is that the chosen abductive hypothesis Δ alone (without T )
predicts a new observation, as Δ does not affect directly the value of the observable
predicates. Every prediction on an observable derived in T (Δ), not previously true in
T (including the observation that drives the abductive process), corresponds to some
further instance knowledge Δ 0 , which is a consequence of T (Δ), and describes the
new situation (or individual) at hand. Such consequences are already known to be
possible in the theory T , as we know that one of its possible extensions is T (Δ0 ).
In the measles example (p. 16), the observation red spots(john) gives rise to the
hypothesis Δ = measles(john). Adopting this hypothesis leads to a new prediction
red spots(dan), corresponding to the instance knowledge Δ 0 = measles(dan),
which is a consequence of T (Δ). This new prediction could be obtained directly from
T (measles(dan)) without the need of Δ = measles(john).
Similarly, if we consider a previously unobserved situation (not derivable from
T (Δ)) described by Δnew with T (Δ) Δnew deriving a new observation, this is also
already known to be possible as T (Δ Δnew ) is one of the possible extensions of T .
For example, if Δnew = measles(mary), then T (Δ) Δnew , and in fact T Δnew
derives red spots(mary), which is again not a genuine generalisation.
In short, abduction is meant to select some further conditions Δ under which we
should reason with T . It concerns only this particular situation described by Δ and
hence, if Δ cannot impose directly any conditions on the observable predicates, the
only generalisations that we can get on the observables are those contained in T under
the particular restrictions Δ. In this sense we say that the generalisation is not genuine
but already contained in T . Hence, as argued in the chapter by Console and Saitta,
abduction increases the intension of known individuals (abducible properties are now
made true for these individuals), but does not have a genuine generalisation effect on
the observables (it does not increase the extension of the observables with previously
unobserved individuals for which the theory T alone could not produce this extension
when it is given the instance knowledge that describes these individuals).
On the other hand, the universal conditioning of the theory T by the inductive
hypothesis H produces a genuine generalisation on the observables of induction. The
extra conditions in H on the observables introduce new information on the relation of
these predicates to non-observable predicates in the theory T , and from this we get
new observable consequences. We can now find cases where from H alone together
with a (non-observable) part of T , describing this case, we can derive a prediction not
previously derivable in T .
The new generalisation effect of induction shows up more when we consider as
above the case where the given theory for induction has some of its predicates as
abducible (different from the observables). It is now possible to have a new individual
described by the extra abducible information Δ new , such that in the new theory T 0 =
T H produced by induction a new observation holds which was not known to be
possible in the old theory T (i.e. it is not a consequence of T Δ new ). Note that we
cannot (as in the case of abduction) combine H with Δ new to a set Δ0new of instance
knowledge, under which the observation would hold from the old theory T . We can
also have that a new observation holds alone from the hypothesis H and Δ new for such
previously unobserved situations not described in the given theory T . These are cases
ABDUCTIVE AND INDUCTIVE REASONING: BACKGROUND AND ISSUES 21
of genuine generalisation not previously known to be possible from the initial theory
T.
Summarising this subsection, induction – seen as a selection of a set of extensions
defined by the new theory T H – has a stronger and genuinely new generalising
effect on the observable predicates than abduction. The purpose of abduction is to
select an extension and reason with it, thus enabling the generalising potential of the
given theory T . In induction the purpose is to extend the given theory to a new theory,
the abductive extensions of which can provide new possible observable consequences.
Finally, we point out a duality between abduction and induction (first studied in
(Dimopoulos and Kakas, 1996)) as a result of this analysis. In abduction the theory T
is fixed and we vary the instance knowledge to capture (via T ) the observable knowl-
edge. On the other hand, in induction the instance knowledge is fixed as part of the
background knowledge B, and we vary the general theory so that if the selected the-
ory T is taken as our abductive theory then the instance knowledge in B will form
an abductive solution for the observations that drove the induction. Conversely, if
we perform abduction with T and we consider the abductive hypothesis Δ explaining
the observations as instance knowledge, the original theory T forms a valid inductive
hypothesis.
by the statement of the observation O. In fact, we can replace the universal quantifica-
tion in ‘all bananas from this shop’ by a typical representative through skolemisation.
More importantly, the link of the observation O with the extra information of H is
known a priori as one of the possible ways of reasoning with the theory T to derive
new observable information.
There is a second way in which to view this reasoning and the hypothesis H above.
We can consider the predicate ‘from Barbados’ as the observable predicate with a set
of observations that each of the observed bananas in the shop is from Barbados. We
then have a prototypical inductive problem (like the white bean example of Peirce)
where we generate the same statement H as above, but now as an inductive hypothe-
sis. From this point of view the hypothesis now has a genuine generalising effect over
the observations on the predicate ‘from Barbados’. But where did the observations on
Barbados come from? These can be obtained from the theory T as separate abductive
explanations for each of the original observations (or a typical one) on the predicate
‘yellow’. We can thus understand this example as a hybrid process of first using (sim-
ple) abduction to translate separately each given observation as an observation on the
abducibles, and then use induction to generalise the latter set of observations, thus
arriving at a general statement on the abducibles.
Essentially, in this latter view we are identifying, by changing within the same
problem the observable and abducible predicates, simple basic forms of abduction and
induction on which we can build more complex forms of non-deductive reasoning.
Referring back to our earlier discussion in Section 1.3, these basic forms are: pure
abduction for explanation with no generalisation effect (over what already exists in the
theory T ); and pure induction of simple generalisations from sample to population.
This identification of basic distinct forms of reasoning has important computational
consequences. It means that we can consider two basic computational models for the
separate tasks of abduction and induction. The emphasis then shifts to the question
of how these basic forms of reasoning and computation can be integrated together to
solve more complex problems by suitably breaking down these problems into simpler
ones.
It is interesting to note here that in the recent framework of inverse entailment as
used by the ILP system Progol (Muggleton, 1995) where we can learn from general
clauses as observations, an analysis of its computation as done in the chapter by Ya-
mamoto reveals that this can be understood as a mixture of abduction and induction.
As described in the above example, the Progol computation can be separated into first
abductively explaining according to the background theory a skolemised, typical ob-
servation, and then inductively generalising over this abductive explanation. The use-
fulness of explicitly separating out abduction and induction is also evident in several
works of theory formation or revision. Basic computational forms of abduction and
induction are used together to address these complex problems. This will be described
further in Section 1.4 on the integration of abduction and induction in AI.
ming. Indeed, when we examine the computational models used for abduction and
induction in AI, we notice that they are very different. Their difference is so wide that
it is difficult, if not impossible, to use the computational framework of one form of
reasoning in order to compute the other form of reasoning. Systems developed in AI
for abduction cannot be used for induction (and learning), and vice versa, inductive AI
systems cannot be used to solve abductive problems. 9 In the chapter by Christiansen a
system is described where the computation of both forms of reasoning can be unified
at a meta-level, but where the actual computation followed by the system is different
for the separate forms of reasoning.
We will describe here the main characteristics of the computational models of the
basic forms of abduction and induction, discussed above, as they are found in practical
AI approaches. According to these basic forms, abduction extracts an explanation
for an observation from a given theory T , and induction generalises a set of atomic
observations. For abduction the computation has the following basic form: extract
from the given theory T a hypothesis Δ and check this for consistency. The search
for a hypothesis is done via some form of enhanced deduction method e.g. resolution
with residues (Cox and Pietrzykowski, 1986; Eshghi and Kowalski, 1989; Kakas and
Mancarella, 1990; Denecker and de Schreye, 1992; Inoue, 1992; Kakas and Michael,
1995), or unfolding of the theory T (Console et al., 1991; Fung and Kowalski, 1997).
The important thing to note is that the abductive computation is primarily based on
the computation of deductive consequences from the theory T . The proofs are now
generalised so that they can be successfully terminated ‘early’ with an abductive for-
mula. To check consistency of the found hypothesis, abductive systems employ stan-
dard deductive methods (these may sometimes be specially simplified and adapted to
the particular form that the abductive formulae are restricted to take). If a hypothesis
(or part of a hypothesis) is found inconsistent then it is rejected and another one is
sought. Note that systems that compute constructive abduction (e.g. SLDNFA (De-
necker and de Schreye, 1998) , IFF (Fung and Kowalski, 1997), ACLP (Kakas and
Michael, 1995)), where the hypothesis may not be ground but can be an existentially
quantified conjunction (with arithmetic constraints on these variables) or even a uni-
versally quantified formula, have the same computational characteristics. They arrive
at these more complex hypotheses by extending the proof methods for entailment to
account for the (isolated) incompleteness on the abducible predicates.
On the other hand, the computational model for the basic form of induction in AI
takes a rather different form. It constructs a hypothesis and then refines this under
consistency and other criteria. The construction of the hypothesis is based on methods
for inverting entailment proofs (or satisfaction proofs in the case of confirmatory in-
duction) so that we can obtain a new theory that would then entail (or be satisfied by)
the observations. Thus, unlike the abductive case, the computation cannot be based
on proof methods for entailment, and new methods such as inverse resolution, clause
generalisation and specialisation are used. In induction the hypothesis is generated
from the language of the problem domain (rather than a given theory of the domain),
9 With the possible exception of Cigol (Muggleton and Buntine, 1988), a system designed for doing unre-
stricted reversed deduction.
24 P.A. FLACH AND A.C. KAKAS
T′ O
T O′
described below. For further discussion on the integration of abduction and induction
in the context of machine learning see the chapter by Mooney in this volume. Also
the chapter by Sakama studies how abduction can be used to compute induction in an
integrated way.
The cycle of abductive and inductive knowledge development. On the one hand,
abduction can be used to extract from the given theory T and observations O abducible
information that would then feed into induction as (additional) training data. One ex-
ample of this is provided by (Ourston and Mooney, 1994), where abduction identifies
points of repair of the original, faulty theory T , i.e. clauses that could be generalised
so that positive observations in O become entailed, or clauses that may need to be
specialised or retracted because they are inconsistent with negative observations.
A more active cooperation occurs when, first, through the use of basic abduction,
the original observations are transformed to data on abducible background predicates
in T , becoming training data for induction on these predicates. An example of this was
discussed in Section 1.3.4; another example in (Dimopoulos and Kakas, 1996) shows
that only if, before inductive generalisation takes place, we abductively transform the
observations into other predicates in a uniform way, it is possible to solve the original
inductive learning task. In this volume, Abe studies this type of integration, employing
an analogy principle to generate suitable data for induction. Similarly, Yamamoto’s
analysis of the ILP system Progol in this volume shows that – at an abstract level – the
computation splits into a first phase of abductively transforming the observations on
one predicate to data on other predicates, followed by a second generalisation phase
to produce the solution.
In the framework of the system RUTH (Adé et al., 1994), we see induction feeding
into the original abductive task. An abductive explanation may lead to a set of required
facts on ‘inducible’ predicates, which are inductively generalised to give a general rule
in the abductive explanation for the original observations, similar to (one analysis of)
the bananas example discussed previously.
These types of integration can be succinctly summarised as follows. Consider a
cycle of knowledge development governed by the ‘equation’ T H j= O, where T is
the current theory, O the observation triggering theory development, and H the new
knowledge generated. Then, as shown in Figure 1.2, on one side of this cycle we
26 P.A. FLACH AND A.C. KAKAS
have induction, its output feeding into the theory T for later use by abduction, as
shown in the other half of the cycle, where the abductive output in turn feeds into the
observational data O for later use by induction, and so on.
Inducing abductive theories. Another way in which induction can feed into abduc-
tion is through the generation of confirmatory (or descriptive) inductive hypotheses
that could act as integrity constraints for the new theory. Here we initially have some
abductive hypotheses regarding the presence or absence of abducible assumptions.
Based on these hypotheses and other data in T we generate, by means of confirmatory
induction, new sentences I which, when interpreted as integrity constraints on the new
theory T , would support the abducible assumptions (assumptions of presence would
be consistent with I, assumptions of absence would now be inconsistent with I).
This type of cooperation between abductive and inductive reasoning is based on a
deeper level of integration of the two forms of reasoning, where induction is perceived
as hypothesising abductive (rather than deductive) theories. The deductive coverage
relation for learning is replaced by abductive coverage, such that an inductive hypoth-
esis H is a valid generalisation if the observations can be abductively explained by
T 0 = T H, rather than deductively entailed. A simple example of this is the exten-
sion of Explanation-Based Learning with abduction (Cohen, 1992; O’Rorke, 1994),
such that deductive explanations are allowed to be completed by abductive assump-
tions before they are generalised.
Inducing abductive theories is particularly useful in cases where the domain theory
is incomplete, and also when performing multiple predicate learning, as also in this
case the background knowledge for one predicate includes the incomplete data for the
other predicates to be learned. In these cases the given theory T is essentially an ab-
ductive theory, and hence it is appropriate to use an abductive coverage relation. On
the other hand, it may be that the domain that we are trying to learn is itself inher-
ently abductive or non-monotonic (e.g. containing nested hierarchies of exceptions),
in which case the hypothesis space for learning is a space of abductive theories.
LAB (Thompson and Mooney, 1994) is one of the first learning systems adopting
this point of view (see also Mooney’s contribution to this volume). The class predi-
cates to be learned are the abducible predicates, and the induced theory H describes the
effects of these predicates on other predicates that we can observe directly with rules
of the form observation class. Then the training examples (each consisting of a set
of properties and its classification) are captured by the induced hypothesis H when the
correct classification of the examples forms a valid abductive explanation, given H, for
their observed properties. Other frameworks for learning abductive theories are given
in (Kakas and Riguzzi, 1997; Kakas and Riguzzi, 1999; Dimopoulos et al., 1997) and
the chapter by Lamma et al. Here, both explanatory and confirmatory induction are
used to generate theories together with integrity constraints. In this volume, Inoue
and Haneda also study the problem of learning abductive logic programs for capturing
non-monotonic theories.
With this type of integration we can perceive abduction as being used to evaluate
the suitability or credibility of the inductive hypothesis. Similarly, abductive expla-
nations that lead to induction can be evaluated by testing the induced generalisation.
ABDUCTIVE AND INDUCTIVE REASONING: BACKGROUND AND ISSUES 27
In this sense, the integration of abduction and induction can help to cross-evaluate the
hypothesis that they generate.
1.5 CONCLUSIONS
The nature of abduction and induction is still hotly debated. In this introductory chap-
ter we have tried to chart the terrain of possible positions in this debate, and also to
provide a roadmap for the contributions to this volume. From a logico-philosophical
perspective, there are broadly speaking two positions: either one holds that abduction
provides explanations and induction provides generalisations; or one can hold that ab-
duction is the logic of hypothesis generation and induction is the logic of hypothesis
evaluation. AI approaches tend to adopt the first perspective (although there are ex-
ceptions) – abduction and induction each deal with a different kind of incompleteness
of the given theory, extending it in different ways.
As stressed in the introduction to this chapter, we do however think that absolute
positions in this debate may be counter-productive. Referring back to the questions
formulated there, we think it will be useful to unify abduction and induction when
concentrating on hypothesis evaluation. On the other hand, when considering hypoth-
esis generation we often perceive a distinction between abduction and induction, in
particular in their computational aspects.
With respect to the second question, abduction and induction can be usefully inte-
grated when trying to solve complex theory development tasks. We have reviewed a
number of AI approaches to such integration. Most of these frameworks of integration
use relatively simple forms of abduction and induction, namely abduction of ground
facts and basic inductive generalisations. Moreover, each of the two is computed sep-
arately and its results transferred to the other, thus clearly recognising two separate
and basic computational problems. From these, they synthesise an integrated form of
reasoning that can produce more complex solutions, following a cyclic pattern with
each form of reasoning feeding into the other.
A central question then arises as to what extent the combination of such basic forms
of abduction and induction is complete, in the sense that it encapsulates all solutions to
the task. Can they form a generating basis for any method for such theory development
which Peirce describes in his later work as ‘coming up with a new theory’? We hope
that the present collection of papers will contribute towards understanding this issue,
and many other issues pertaining to the relation between abduction and induction.
Acknowledgments
Part of this work was supported by Esprit IV Long Term Research Project 20237 (Inductive
Logic Programming 2).
{
References
Adé, H., Malfait, B., and Denecker, M. (1994). RUTH: an ILP theory revision sys-
tem. In Proceedings of the Eighth International Symposium on Methodologies for
28 P.A. FLACH AND A.C. KAKAS