<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Loizos Michael</string-name>
          <email>loizos@ouc.ac.cy</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Open University of Cyprus</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Starting from McCarthy's seminal proposal for the construction of an advice-taking machine, we show how tools from abstract argumentation and computational learning can help disambiguate and formalize that goal. Our focus is on establishing provable guarantees on the realizability of the advice taker, and on implementing an actual prototype system.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The construction of the advice taker as a means towards the
endowment of machines with intelligent behavior was put
forward by McCarthy [1959] both as a complement to the use of
programming — the then-prevalent mode of human-machine
interaction — and as an intermediate milestone to the
eventual use of learning from experiences. Programming has
certainly stuck around as a key way of interacting with machines,
while the recent seemingly uncanny ability of learning-based
machines to outperform human experts across tasks and
domains might suggest that McCarthy’s agenda is now dated.</p>
      <p>This might have been the case had it not been for a new
impetus on the viability of modern machine-learning techniques
(such as deep learning) to produce human-like intelligence:
their lack of operational transparency1, and their brittleness
— which, somewhat ironically, was the issue plaguing
systems based on classical logic (like McCarthy’s advice taker)
that machine learning was supposed to address — as a result
of their subtle dependency on biases in their training data.</p>
      <p>We do not intend to debate whether the currently-trending
machine-learning paradigm will eventually succeed in
overcoming this impetus. We do contend, however, that
McCarthy’s proposal for the development of an advice taker still
remains, today, relevant as one way of thinking about these
issues, and perhaps making progress towards addressing them.</p>
      <p>In this work we revisit McCarthy’s [1959] proposal
seeking to make three contributions: (i) to show that certain key,
but admittedly under-specified, aspects of the original
proposal can be made concrete through formal argumentation and
learning theory, both of which have dealt with the problem of
symbolic knowledge manipulation; (ii) to establish formal
results on the realizability of constructing the advice taker; and
(iii) to present an explicit example of such a construction.
1.1</p>
      <p>Overview
We introduce in Sections 2–4 a view of the advice taker as
an online learning machine that seeks to approximate a target
theory known to an advice giver (e.g., a human) interacting
with the advice taker, in a manner that offers PAC-like
guarantees [Valiant, 1984]. Both the target theory and the advice
taker’s hypothesis are formalized in Sections 6–7 through a
novel structured argumentation framework. The framework
accounts for the representation of arguments that entail not
only inferences, but also actions that the advice taker is
assumed to execute. Unlike work in reinforcement learning,
however, the learning of these action-entailing arguments does
not presuppose the execution of the actions and the receipt of
a reward indicating their appropriateness in a given context.
Instead, a supervised learning approach is used to learn all
arguments, both inference-entailing and action-entailing ones.</p>
      <p>Facilitated by the underlying structured argumentation
framework, and unlike what a straightforward application of the
PAC semantics would suggest: the advice taker does not only
predict the label of an example — which would correspond to
an extension of the advice taker’s hypothesis — but also
provides an explanation on how the prediction is reached;
correspondingly, the advice giver does not return the correct label
— which would correspond to an extension of the target
theory — but instead offers advice, which generally corresponds
to only a part of an argument, and is both smaller than the full
correct label, and cognitively easier for a human to construct.</p>
      <p>Our approach, thus, takes a step back from a pure
learningbased view of the advice taker, and a step towards
programming, by treating each piece of advice as a “code snippet” that
the advice taker integrates into its current hypothesis. We
quantify the necessary information content of these pieces of
advice in Section 5, and show in Section 8 that PAC-like
guarantees can indeed be established for a natural advice giver
and a simple integration policy. An advice taker that adopts
this particular integration policy is fully implemented in
SWIProlog, while a visual web-based interface is under continual
development, with the current version being accessible online
from: http://cognition.ouc.ac.cy/prudens/.</p>
      <p>The Advice Taker and the Advice Giver
[. . . ] its behavior will be improvable merely by making statements
to it [that] will require little if any knowledge of the program or
the previous knowledge of the advice taker. — McCarthy, 1959
Unlike most contemporary work on Knowledge
Representation and Reasoning, which tacitly assumes that knowledge is
available a priori, we consider the involvement of an explicit
advice giver in an online knowledge acquisition process.</p>
      <p>We start with the supposition that the advice taker follows
an elaborate-integrate cycle. Each cycle is initiated by an
external or internal stimulus, such as an incoming query by a
user, an event occurrence, the lapse of time, etc. The advice
taker continues to sense the current context, which it uses as
input to the processes of that cycle. By context we mean
anything visible to the advice giver that might be relevant for the
advice taker to consider during that cycle, including data that
the advice taker may produce through its internal processes.
As this need not be specified more concretely for the
purposes of the current section, we shall simply assume, for now,
that a context x 2 X is a finitely-representable piece of data.</p>
      <p>The advice taker proceeds to elaborate the information in
context x by drawing inferences that follow from its current
knowledge and x. Some of the elaborations may correspond
to actions that the advice taker executes, including the action
of simply reporting to a user some or all of the inferences that
were drawn. Each elaboration is accompanied by an
explanation on how it was reached, which is made available to the
advice giver. With access to those explanations, the advice
giver proceeds to offer advice to the advice taker, and the
latter integrates that advice into its current knowledge, which is
later used to elaborate contexts in subsequent cycles.</p>
      <p>As is the case for context, a high-level abstract view of
the data involved during elaboration and integration suffices
for now, and a more concrete proposal will be made in the
sequel. We shall assume, therefore, that the advice taker and
the advice giver share a set D of finitely-representable
derivations, which can be thought to be determined by a previously
agreed-upon syntax, such as a fragment of first-order logic.</p>
      <p>During the elaboration step, the advice taker maps the
current context x into a set of derivations, through an elaboration
theory t : X ! 2D that it maintains and revises across cycles.
The derivations in t(x) determine the inferences and actions
that follow from x and their corresponding explanations.</p>
      <p>During the integration step, the advice giver computes and
offers a finitely-representable advice 2 A to the advice
taker, by applying a feedback function f : X 2D ! A on the
current context x and the advice taker’s elaboration t(x). In
turn, the advice taker integrates advice into its current
elaboration theory t to produce a revised version. Although we
shall not specify, for the time being, how each piece of advice
is integrated, we shall make provisions for A to include the
symbol ‘?’ indicating that the advice giver has “no advice”
to give during that cycle, or, equivalently, that it finds the
elaboration t(x) of the advice taker acceptable, meaning that the
advice taker is not expected to revise its current elaboration
theory. We shall further assume that for each context x, there
exists at least one elaboration e(x) 2 2D that is acceptable.</p>
      <p>Necessarily, the interaction between the advice taker and
the advice giver ends up endowing the former with subjective
knowledge that reflects the beliefs and biases of the latter. We
will, indeed, embrace this subjectivity of the knowledge for
most of this paper, and we will briefly revisit at the end the
question of how commonsense knowledge can be acquired by
the advice taker through some form of autonomous learning.
3</p>
      <p>How Should the Advice Taker Elaborate?
One will be able to assume that the advice taker will have
available to it a fairly wide class of immediate logical consequences of
anything it is told and its previous knowledge. — McCarthy, 1959
The advice taker elaborates a context by producing inferences
and actions, supported by a set of derivations that are returned
by its elaboration theory. Whether an elaboration is
meaningful boils down to how the elaboration theory is revised after
a piece of advice is received by the advice taker. We shall
require, then, that the advice taker works towards ensuring
that its elaboration theory eventually is such that t(x) is
acceptable to the advice giver (i.e., f (x; t(x)) = ?) sufficiently
frequently. To make this intuition concrete, we proceed to
quantify the terms “frequently”, “sufficiently”, and“eventually”.</p>
      <p>We do not interpret “frequently” narrowly in terms of some
fraction of contexts that needs to be elaborated in an
acceptable fashion. Indeed, it might be the case that certain contexts
occur rarely, and hence are practically inconsequential, while
a small fraction of contexts might occur frequently enough so
that their acceptable elaboration would be critical. To
quantify this situation, we shall assume that contexts are drawn
independently at random from some fixed, but unknown to the
advice taker, probability distribution P over contexts in X .
We acknowledge that this independence assumption might
not always hold in certain realistic scenarios, where contexts
across different cycles of interaction might end up being
correlated. Nonetheless, we adopt this simplifying assumption if
only for the purposes of formal analysis of the interaction.</p>
      <p>Given the above notion of frequency, we say that an
elaboration theory t : X ! 2D is (1 ")-approximately
acceptable under P against an advice giver with feedback function
f if f (x; t(x)) = ? for a context x drawn from P, except
with probability at most " over the randomness of sampling
from P. Instead of fixing “sufficiently” to correspond to a
particular value of ", we shall let the advice taker receive an
arbitrary value for " 2 (0; 1] as a parameter when it is
initialized, and we will expect the advice taker to eventually produce
a (1 ")-approximately acceptable elaboration theory.</p>
      <p>This leads to the question of how long “eventually” is; how
long are we willing to allow the advice taker to be in a
training mode where its elaborations are not sufficiently
frequently accepted, before producing a desirable elaboration theory.
We are not suggesting that the advice taker should ever stop
accepting advice and integrating it in its elaboration theory;
indeed, we view the elaborate-integrate cycle as a perpetual
process. Rather, we are looking only to quantify the length of
an initial period during which the advice taker’s elaborations
should still be under scrutiny by the advice giver, much like
how the actions of a novice human assistant are closely
monitored during an initial training period of their employment.</p>
      <p>On the one hand, the advice taker should be given sufficient
training time to compensate for the burden of its knowledge
acquisition task, which grows with certain task parameters,
such as: the representation size n of the drawn contexts, the
representation size s of the advice giver’s feedback function,
the value of 1=" (since the task becomes harder as "
becomes smaller). We shall assume, therefore, that the training
time is some function g(1="; n; s). On the other hand,
allowing g(1="; n; s) to grow arbitrarily (e.g., exponentially) in its
parameters would entail a computationally overly-powerful
advice taker, and would impose unrealistic demands on the
advice giver. A theoretically-reasonable compromise is to let
g(1="; n; s) grow polynomially in its parameters.
Pragmatically, the degree of the polynomial needs to be a small value.</p>
      <p>Despite our effort to allocate a fair amount of resources to
the advice taker, we have not accounted for the possibility of
the advice taker simply being unlucky! Consider, for
example, a probability distribution P that assigns probability 1=3
to context x1 and probability 2=3 to context x2. Suppose,
further, that we set " := 1=2, and expect the advice taker to
produce a 1=2-approximately acceptable elaboration theory t
under P against some advice giver. It follows that t(x2)
should eventually be acceptable by the advice giver. But what
if the advice taker never gets the opportunity to interact with
the advice giver in context x2 to get advice on what might be
acceptable in that context? Despite x2 being the most
probable context under P, the probability of never drawing it is
non-zero, irrespectively of any finitely-long training period.</p>
      <p>Consequently, we slightly downgrade our expectations, so
that the advice taker may fail to eventually produce a
desirable elaboration theory, but only with a “sufficiently small”
probability . As in our treatment of ", we let the advice taker
receive an arbitrary value for 2 (0; 1] as a parameter when
it is initialized, and we allow it to fail with at most probability
. Correspondingly, we also allow the advice taker’s training
time to depend on (and grow polynomially in) 1= , so that the
advice taker’s training time is a function g(1= ; 1="; n; s).
4</p>
      <p>An Advice Taker with Learning Guarantees
Our ultimate objective is to make programs that learn from their
experience as effectively as humans do. — McCarthy, 1959
In the preceding section we had quantified the type of
guarantees that we expect the advice taker to offer on its operation.
As we have already discussed, these guarantees are
necessarily tied to the advice giver. In a sense, there is no objectively
correct way for the advice taker to draw inferences and act,
since, in particular, there is no objectively correct elaboration
theory (or knowledge) that the advice taker can have. Instead,
the inferences and actions of an elaboration theory (or
knowledge) can, at best, conform to the advice giver’s feedback.</p>
      <p>Even with this shift from objective truth to subjective
conformance, the requirements that we have imposed on the
advice taker — to probably and tractably produce an
approximately acceptable elaboration theory — might seem overly
arduous. In this section we seek to bring these requirements
together, and establish results on their realizability. Our
formalization adopts and extends the celebrated probably
approximately correct semantics for concept learning [Valiant,
1984], emphasizing tractable operation (through the
restrictions on the training time g), provable guarantees on
performance (through the parameters ; " that quantify the quality of
the returned elaboration theory), and a crisp specification of
the interaction between the advice taker and the advice giver.
Definition 4.1 (PAC Advice Takers). An advice taker is
probably approximately conformant (PAC) to an advice giver
with a feedback function in a class F (with fixed sets X , D,
A) if for every real values ; " 2 (0; 1], every probability
distribution P over contexts in X that have a finite
representation size of at most n, and every feedback function f 2 F
with a finite representation size of s, the advice taker is given
access to the parameters ; "; n; F , and proceeds as follows:
a context x is drawn independently at random from P; the
advice taker uses its current (possibly stochastic) elaboration
theory t : X ! 2D to return a set t(x) of derivations; the
advice giver offers the piece of advice f (x; t(x)); the advice
taker integrates the advice to revise the elaboration theory t.</p>
      <p>After time at most g(1= ; 1="; n; s) the advice taker
terminates and returns, except with probability at most , an
elaboration theory t that is (1 ")-approximately acceptable under
P against f . If the running-time function g grows
polynomially in its parameters, then the advice taker is efficient.</p>
      <p>Definition 4.1 captures concisely the requirements on the
interaction and on the elaboration theory of the advice taker
up until the end of the training phase; as we have already
discussed, the interaction and revision of the elaboration theory
might continue after that. For the purposes of our subsequent
formal analysis, we have allowed the elaboration theory to be
a stochastic function, so that the advice taker can choose to
produce random sets of derivations during the training phase,
if this happens to be beneficial towards its goal of being PAC.
In a later section we will offer an explicit construction of a
PAC advice taker without appealing to this stochasticity.</p>
      <p>Although the advice taker interacts with a single
predetermined advice giver, the latter’s feedback function is generally
unknown to the former; had this not been the case, the advice
taker could simply be assumed to have access to all the advice
the advice giver could provide. On the flip side of things, the
advice taker can, and does, have some information about
certain aspects of the advice giver’s feedback function (such as
the set A of all pieces of advice it can possibly return),
depending on the particular scenario in which the interaction takes
place. This partial knowledge of the advice giver’s feedback
function f is captured in Definition 4.1 by assuming that f
belongs in a class F that is known to the advice taker, while
leaving open the possibility that f could be any member of
F , and insisting that the advice taker delivers in all cases.</p>
      <p>With the aforementioned clarifications, we can now
establish formal results on the realizability of a PAC advice taker.</p>
      <p>Our first result shows that a (possibly non-efficient) PAC
advice taker always exists, demonstrating, on the one hand,
the non-vacuousness of Definition 4.1, and highlighting, on
the other hand, the importance of insisting on efficiency.
Theorem 4.1 (Existence of a Universal PAC Advice Taker).
For every class F , there exists an advice taker that is PAC to
an advice giver with any feedback function that belongs in F .
Proof (sketch). Memorize the first m contexts, and map each
through t to the next set of derivations in 2D that have not yet
been paired with it, until all m elaborations are acceptable.
Do not execute any of the actions in the considered
elaborations. The result follows for an appropriate choice of m.</p>
      <p>The advice taker constructed in the proof of Theorem 4.1
does not really utilize any advice from the advice giver other
than for the purpose of checking whether its elaboration of
each given context is acceptable. This minimal reliance on the
advice is compensated by the long running time of the advice
taker, which effectively performs a blind systematic search of
an appropriately selected subspace of all elaboration theories.
5</p>
    </sec>
    <sec id="sec-2">
      <title>How Should the Advice Giver Advise?</title>
      <p>In order for a program to be capable of learning something it must
first be capable of being told it. In fact, in the early versions we
shall concentrate entirely on this point [. . . ] — McCarthy, 1959
In view of Theorem 4.1, a natural next step is to examine how
the advice offered by the advice giver can be utilized more
effectively by the advice taker towards constructing its
elaboration theory. Since we are still approaching the interaction
of the advice taker and the advice giver at an abstract level,
we cannot yet discuss the structure of this advice, but we can,
however, examine the amount of information that it conveys.</p>
      <p>To quantify the amount of such information that might be
necessary, we first introduce an auxiliary definition. Consider
a partition u of 2D of size juj, and denote by u[i] its i-th
element in some order. We shall say that u is shattered by a
class F of feedback functions if there exists a context x 2 X
such that for every i 2 f1; 2; : : : ; jujg, there exists a feedback
function f 2 F for which fe j f (x; e) = ?g u[i]. The
(information) sparsity spr(F ) of a class F is the maximum
value juj for a chosen partition u of 2D that is shattered by F .</p>
      <p>Intuitively, the sparsity of a class F captures how little
information is conveyed to the advice taker, in a worst-case
context x, towards identifying what is acceptable to the advice
giver, after the advice taker has successfully eliminated a
feedback function in F from being the one actually used by the
advice giver. Consequently, the higher the sparsity of F , the
less information the advice taker receives after each
elimination, and the more feedback functions it needs to eliminate
before gathering enough information to meet its goal.2
Theorem 5.1 (Necessary Feedback for Efficient PAC Advice
Takers). If there exists an advice taker that is PAC to an
advice giver with any feedback function f : X 2D ! A
in class F , and runs in time g( ), then jAjg( ) is (spr(F )).
Proof (sketch). An advice taker receives at most g( ) pieces
of advice and, therefore, at most log jAj g( ) bits of
information, each eliminating at most half of the remaining parts of u
from including the acceptable elaborations according to f . If
jAjg( ) (1=2) spr(F ), then at least two parts of u are not
eliminated, giving a t with an expected error of at least 1=2.
The result follows for appropriate choices of and ".</p>
      <p>2This measure of the lack of information in a worst-case example
differs from VC-dimension [Kearns and Vazirani, 1994], which aims
to lower-bound the number of distinct examples needed to learn.</p>
      <p>Depending, then, on the paucity of initial information that
the advice taker has on which elaborations the advice giver
might find acceptable (to the extent that such information
follows from the advice taker’s knowledge of F ), the constraint
on the efficiency of the advice taker has an effect on the
richness of the advice that the advice giver needs to offer.
6</p>
      <p>Knowledge Representation and Reasoning
[. . . ] probably have a single rule of inference which will combine
substitution for variables with modus ponens. — McCarthy, 1959
Our emphasis up to this point was on formalizing and
quantifying certain aspects of the interaction between the advice
taker and the advice giver. Before giving an explicit
construction of the two actors, we must make concrete the language
in which the advice taker represents the knowledge that it
reasons with and the advice giver represents the advice it offers.</p>
      <p>Recall that the role of the advice taker’s reasoning module
is not to identify certain objective truths, nor to deduce what
is logically entailed by a given context and the available
knowledge. Instead, the role of reasoning is to produce
elaborations that are acceptable by the advice giver. This subjectivity
lessens the burden on the consideration of any particular type
of reasoning semantics for the advice taker, since the
interaction with the advice giver can finely-tune the produced
elaborations by customizing the former’s knowledge. This is
decidedly demonstrated by Theorem 4.1, which establishes the
existence of a PAC advice taker without reliance on the
reasoning semantics. The primary role of the reasoning semantics,
then, is to support the flexible revision of the knowledge. Our
proposal below should, therefore, be read in this spirit, as a
reasonable candidate for the representational syntax and the
reasoning semantics, but by no means as the only option.</p>
      <p>We consider a fixed language L that assigns finite names
to the members of the two non-overlapping sets of constants
and variables, and defines the following constructs: An atom
is a constant or a variable. A predicate is a constant followed
by a finite parentheses-surrounded comma-separated list of
atoms; these atoms are the predicate’s arguments, and their
multitude is the predicate’s arity. A literal is a predicate or a
predicate prefixed by the symbol ‘:’. A query is a predicate
prefixed by the symbol ‘?’. An action is a predicate prefixed
by the symbol ‘!’. A term is a literal, a query, or an action.
Two variable-free terms are conflicting if they are the literals
and : for a predicate . Following Prolog convention, we
capitalize the first letter of variables but not of constants.</p>
      <p>A context in X is a collection of variable-free literals. A
rule in R is an expression of the form ' , where the
body ' of the rule is a finite comma-separated list of literals
or queries, and the head of the rule is a literal or an action
whose variables all appear in some of the terms of '. An
inferential rule is a rule with a literal head, while a
production rule is a rule with an action head. An instance of a rule
is an expression obtained from the rule after all variables in
the rule’s terms are replaced with constants, so that all
appearances of the same variable across the rule map to the same
constant. Two (inferential) rule instances are conflicting if
their (variable-free) head terms are conflicting. Two
(inferential) rules are conflicting if they have conflicting instances.</p>
      <p>A knowledge base = h%; i comprises a finite collection
% R of rules, and an irreflexive antisymmetric priority
relation that is a subset of % % and includes pairs of
conflicting (inferential) rules only. For simplicity of exposition,
we may occasionally assume that rules in % are named, and
define over their names. The priority relation is lifted to
apply on rule instances also, in the natural manner. A
knowledge base is totally-prioritized if for every pair of conflicting
(inferential) rules r1; r2 2 % either r1 r2 or r2 r1 holds.
For a totally-prioritized knowledge base, the priority relation
need not be specified explicitly, as it can be determined by the
order in which rules are represented in the knowledge base.</p>
      <p>The advice taker reasons with such a knowledge base with
the aim to elaborate a context. At a high level, reasoning
proceeds by applying a forward chaining policy, starting with an
initial set of variable-free predicates specified by the given
context, and repeatedly applying modus ponens by
interpreting rules as classical implications. Whenever a query ? is
encountered in the body of a rule, this is treated as a call to
an external oracle to decide if the predicate is true on its
given arguments. If not all of the arguments of happen to be
constant at the time of this call, then the call unifies, if
possible, the variables of with constants such that the instantiated
predicate is true. If multiple such instantiations are possible,
this gives multiple instances of the rule, the evaluation of each
of which proceeds independently. Whenever a (variable-free)
action is encountered in the head of a rule, this is treated as
an execution of a correspondingly-named external procedure
with inputs determined by the action’s (constant) arguments.</p>
      <p>To make our reference to an external oracle and to external
procedures concrete, we will stipulate the existence,
alongside the knowledge base , of a Prolog program , and shall,
henceforth, write ( ) when we wish to make explicit the
Prolog program that is linked to the knowledge base . The
encountering of a query ? or an action ! in a rule of
corresponds to the posing of the associated predicate as a goal to
. In the case of an action, the Prolog call is made for its
sideeffects, not for determining whether it succeeds. This unified
treatment allows one to include seamlessly arbitrary pieces of
imperative code in an otherwise declarative knowledge base.
Such imperative code allows the advice taker to gain access to
the sensors and actuators of any device on which the advice
taker is operating, and, in particular, to perceive signals
beyond those that are available in a context. A special kind of
these sensors, and we expect the most oft-used ones, are
those that check for the axiomatic truth of statements involving,
most typically, equality, ordering, or set membership.
Definition 6.1 (Derivation). A derivation for a term 0 in a
context x under a knowledge base ( ) is a singly-rooted
directed acyclic hypergraph such that: each vertex is a distinct
variable-free term; each leaf term is either a literal that
belongs in x or a query that is true according to ; the root
term is 0; each directed hyperedge between term sets B and
H is labeled by an instance 'i i of a rule in such that
H = f ig and B comprises the terms in 'i; no term
appears in the head of more than one labeling rule instance. The
crown rule instance of a derivation, if it exists, is the labeling
rule instance with head 0, and is the only possible use of an
instance of a production rule in the derivation.</p>
      <p>We conclude this section with an example knowledge base.
Example 6.1. A phone assistant implementing a user’s
policy, with rules in increasing order of priority. Queries refer
to built-in or easy-to-implement Prolog predicates, and
actions refer to Prolog predicates that control phone functions.
Policy: Decline calls when busy; e.g., when in a meeting at work.
Calls from family members are important. Send an SMS when
declining important calls. Repeated calls are urgent; unless coming from
“serial” repeated callers. Repeated important calls are urgent.
Answer urgent calls. Answer calls from colleagues that are expected.
my pos(P1), work pos(P2), ?dist(P1,P2,D), ?(D&lt;50)
at work
calendar(From,To), time(T), ?(From&lt;T), ?(T&lt;To)
in meeting
in meeting, at work</p>
      <p>busy
call from(C)</p>
      <p>will answer(C)
call from(C), busy
will answer(C)
:will answer(C)</p>
      <p>:will answer(C)
!answer(C)</p>
      <p>!decline(C)
contact(C,Info), ?member(in group(family),Info)
important(C)
call from(C), important(C), :will answer(C)
!sms(C,‘Busy. . . ’)
contact(C,Info), ?member(log(L),Info), ?last entry(E,L), time(T),
?member(when(W),E), ?diff(T,W,D), ?(D&lt;3) recent caller(C)
call from(C), recent caller(C)</p>
      <p>urgent(C)
“. . . left as an exercise to the reader!”
serial repeated caller(C)
call from(C), serial repeated caller(C)
:urgent(C)
call from(C), recent caller(C), important(C)
urgent(C)
call from(C), urgent(C)</p>
      <p>will answer(C)
contact(C,Info), ?member(in group(work),Info)
colleague(C)
call from(C), colleague(C), is expected(C)
will answer(C)</p>
      <p>The explicit treatment of time, in domains where this is
beneficial, can proceed analogously to our work on story
understanding [Diakidoy et al., 2014; 2015], with extra knowledge
used to address the specific problems associated with
temporal reasoning (e.g., the persistence of observed information).
7</p>
    </sec>
    <sec id="sec-3">
      <title>An Argumentative Elaboration Theory</title>
      <p>In our opinion, a system which is to evolve intelligence of human
order should [at least be such that:] Interesting changes in
behavior must be expressible in a simple way. — McCarthy, 1959
The advice taker elaborates a context x by mapping it through
some elaboration theory t to a set t(x) of derivations.
Definition 6.1 already accounts for several of the features we
require for this elaboration process, including its dual role to
specify drawn inferences (which come through crown
inferential rule instances) and actions to be taken (which come
through crown production rule instances). To complete our
proposal on how a knowledge base gives rise to elaborations,
it remains to specify which derivations should appear in t(x).</p>
      <p>Without loss of generality, we interpret any set of actions
for which derivations exist in t(x) as being intended to be
executed together in an arbitrary order. Whenever a
particular order is required, then a different action corresponding to
an external procedure that orders things appropriately should
be used. In effect, then, derivations can be incompatible only
because of a conflict between their respective instances of
inferential rules. To resolve which derivations appear together
in t(x), we appeal to abstract argumentation [Dung, 1995].</p>
      <p>Priorities in a knowledge base indicate a preference on how
conflicts between rule instance pairs are to be resolved,
effectively stating that if the bodies of both rule instances are
known to hold, then the evidence favors the head of the more
preferred rule instance to be drawn as an inference. In the
absence of a priority between two conflicting rule instances,
then neither of their heads is to be drawn as an inference. This
idea can be lifted to handle conflicts between derivations.
Definition 7.1 (Rebuttal). Consider two derivations, the first
one for term 1 and the second one for some unspecified term,
both in a context x under a knowledge base ( ) with =
h%; i. The first derivation rebuts the second derivation on
rule instance r2 if there exists a non-leaf term 2 in the second
derivation pointed to by a hyperedge labeled by rule instance
r2, such that 2 is conflicting with 1, while term 1 is either:
a leaf in the first derivation; the rebuttal is exogenous,
the head of the first derivation’s crown rule instance r1,
and it holds that r2 6 r1; the rebuttal is endogenous.</p>
      <p>The derivation being rebutted is always attacked on a weak
link: a rule instance r2 in the derivation that either conflicts
with a leaf of the attacking derivation, or is less preferred
than the crown rule instance r1 of the attacking derivation.
The former type of rebuttal is equivalent to saying that r2 is
attacked because the negation of its head appears in context x.
Indeed, the attacking derivation in the former type of rebuttal
necessarily comprises only that single leaf, since this is the
only way that the leaf can also be the derivation’s root 1.</p>
      <p>Each derivation naturally corresponds to an argument, and
the rebuttal relation between derivations corresponds to an
attacking relation between arguments. The argumentation
framework that is induced in the stated manner by a knowledge
base ( ) and a context x can be seen to adopt the following
choices under the general ASPIC+ categorization [Prakken,
2010]: axiomatic premises (which correspond to the context
x), defeasible rules, rebutting attacks, and application of rule
preferences on the last link. We, thus, conclude the
specification of the advice taker’s reasoning semantics by fixing the set
t(x) of derivations that it uses to elaborate x to be the unique
grounded extension of the induced argumentation framework.</p>
      <p>The grounded extension-based semantics is chosen, in part,
because of the tractability of computing the grounded
extension. This tractability is known to hold only when it is
measured against the size of the argumentation framework. The
induced argumentation framework, in our case, can be
exponentially larger than the size of the knowledge base (even for
a “propositional” knowledge base). Fortuitously, we are able
to prove the following, essentially optimal, tractability result.
Theorem 7.1 (Efficient Computation of the Grounded
Extension). Consider a knowledge base ( ) and a context x. Let
jj ( )jj be the number of distinct crown rule instances in all
derivations in x under ( ), and let jxj be the size of x.
Assume, further, that every call to takes unit time. Then, the
grounded extension of the argumentation framework induced
by ( ) and x can be computed in time polynomial in jj ( )jj
and jxj. Furthermore, there exists no algorithm that can
compute the grounded extension of the argumentation framework
induced by ( ) and x in time sub-linear in jj ( )jj and jxj.
Proof (sketch). Starting from x, repeatedly apply modus
ponens to construct a derivation network N that concisely
represents all derivations in x under ( ) [S1]. Mark literals
in x [S2] and repeat the following until convergence: remove
literals that conflict with marked literals [S3]; remove
hyperedges that are not labeled by the crown rule instance of a
derivation in N [S4]; mark hyperedges that are labeled by a
rule instance r1 whose body literals are all currently marked,
and for which any other hyperedge that is labeled by a
conflicting rule instance r2 is such that r1 r2 [S5]; mark the
head term of every hyperedge that is currently marked [S6].
Return the marked part of N [S7], which concisely represents
the grounded extension. The Appendix shows graphically the
execution of the process on an example knowledge base.</p>
      <p>The argumentation-based semantics gives the advice taker
considerable flexibility in revising its elaboration theory. The
addition or removal of rules or priorities, as might be
suggested by the advice giver, can be done in a local and
revisiontolerant3 manner [McCarthy, 1998], while still having a
significant impact on the operating behavior of the advice taker.
8</p>
      <p>Explicit Construction of an Advice Taker
A machine is instructed mainly in the form of a sequence of
imperative sentences; while a human [. . . ] in declarative sentences
describing the situation in which action is required together with
a few imperatives that say what is wanted. — McCarthy, 1959
We conclude the development of our framework for PAC
advice taking by providing explicit, but domain-independent,
constructions of an advice taker and an advice giver.</p>
      <p>The advice taker maintains during each interaction cycle i a
knowledge base i( ), and elaborates a given context x using
the elaboration theory ti induced by i( ), as described in the
preceding section. The advice giver can be thought to hold its
own fixed knowledge base ( ) also, which it uses to identify
whether it finds the inferences drawn and / or actions taken by
the advice taker acceptable. The argumentative nature of
reasoning presents the advice giver with a natural strategy for
offering advice to the advice taker: if the latter’s elaboration
ti(x) is not acceptable by the former in context x, then it must
be because ti(x) does not coincide with the grounded
extension of the argumentation framework induced by ( ) and x;
the former’s advice, then, seeks to reduce this discrepancy,
towards making i( ) and ( ) have induced argumentation
frameworks that share a common grounded extension.</p>
      <p>Consider, therefore, some context x during the i-th cycle
of interaction. The advice taker proposes an elaboration ti(x)
based on its current knowledge base i( ). In response, the
advice giver offers the advice f ( )(x; ti(x)) through a
feedback function f ( ) that depends on the grounded extension
e(x) of the argumentation framework induced by ( ) and
x. The advice taker concludes the cycle by revising i( ) to
3McCarthy [1998] actually calls this property “elaboration
tolerance” — not to be confused with our use of the term “elaboration”.
If: A rule r, an instance of which is in ti(x), is not in ( ).
Advise: “Rule r is invalid and should not be used!”
Revise: The advice taker removes rule r from its
knowledge base, along with all its dependent priorities.
Else if: One of rule r’s instances that is in e(x), is
applicable (i.e., its body is satisfied) in ti(x), but is not in ti(x).
Advise: “Rule r is relevant and should be applied!”
Revise: The advice taker adds rule r to its knowledge base,
with priority over all conflicting rules.</p>
      <p>Else if: A derivation d rebuts, under ( ), a derivation in
ti(x), and it is not rebutted by a derivation in e(x).
Advise: “Derivation d’s rebuttal should be recognized!”
Revise: The advice taker adds rules with instances in d to
its knowledge base, with priority over all conflicting rules.
Else: (None of the preceding conditions holds.)
Advise: “The elaboration of the context is acceptable!”
Revise: The advice taker retains its knowledge base.
i+1( ). In all cases, = h%; i and i = h%i; ii for each
cycle i, with 0 chosen arbitrarily as long as extra time
polynomial in jj 0( )jj is allowed. We can show the following:
Theorem 8.1 (Existence of an Efficient PAC Advice Taker).
Assume unit-time access to an oracle that computes grounded
extensions of argumentation frameworks as discussed in
Theorem 7.1. For every fixed Prolog program , and every class
F f ( ) j is any totally-prioritized knowledge base
of feedback functions that offer advice as described in
Figure 1, there exists an efficient advice taker that is PAC to
an advice giver with any feedback function that belongs in F .
Proof (sketch). The advice taker and the advice giver interact
as in Figure 1, until m consecutive elaborations are
acceptable. The result follows for an appropriate choice of m.</p>
      <p>Thinking of the advice giver as a human wishing to get an
advice-taking machine to operate in a certain manner, we call
this mode of human-machine interaction machine coaching.
Lying between machine learning and machine programming,
in this mode the human coach interacts rather actively, but
still naturally, with the machine, as summarized below.
machine learning:
the interaction is one-sided and can be online or batch
the human labels inputs with outputs of the target theory
the machine generalizes to create a hypothesis theory
machine programming:
the interaction is one-sided and happens at the beginning
the human generates explicit parts of the target theory
the machine blindly adds parts to the hypothesis theory
machine coaching:
the interaction is dialectical and is necessarily online
the human recognizes mistakes in the hypothesis theory
the machine appropriately revises the hypothesis theory
To highlight a key aspect of machine coaching, note that
our example advice giver can, in principle, either tackle by
itself the advice taker’s task of elaborating contexts using the
grounded extension of the former’s induced argumentation
framework, or directly program the advice taker with the
former’s knowledge. Pragmatically, however, machine coaching
allows the advice giver to offload equitably its computational
and cognitive burdens, exploiting the computational
superiority of machines, and acknowledging the lighter human
cognitive load to recognize mistakes in a dialectical setting
rather than to generate arguments [Mercier and Sperber, 2011].
9</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusions</title>
      <p>[. . . ] a system which can be told to make a specific improvement
in its behavior [. . . ] Once this is achieved, we may be able to tell
the advice taker how to learn from experience. — McCarthy, 1959
McCarthy [1959] had envisioned much more that the advice
taker could do, most intriguing of which is perhaps the
specification of its elaboration and integration procedures through
machine coaching. We believe that our proposal offers a
general enough basis to investigate such possible extensions, by
considering, for example, the case of knowledge bases with
production rules whose associated actions operate not on the
advice taker’s environment, but on the advice taker itself.</p>
      <p>With an eye towards developing argumentation-based
systems compatible with human cognitive capabilities and
limitations [Michael et al., 2015; Kakas and Michael, 2016],
machine coaching is most usefully viewed not as a primary mode
of human-machine interaction — except for simple everyday
tasks (cf. Example 6.1) — but as a debugging and
personalization mode subsidiary to programming and learning, which
currently serve better the goal of acquiring imperative and
general (commonsense) declarative knowledge, respectively.</p>
      <p>On the matter of declarative knowledge, our earlier work
shows that it can be learned autodidactically [Michael, 2008]
without human interaction, and just by reading text [Michael,
2009] as found on the Web. Noting that learning cannot
proceed independently from the manner in which knowledge is
used to reason [Michael, 2014], our recently-proposed NERD
algorithm [Michael, 2016] for learning in a certain
argumentative setting seems adaptable to this work’s reasoning
semantics. The role of machine coaching in learning-driven
knowledge acquisition is exemplified by observing that text on the
Web encodes what can be called websense [Michael, 2013],
with whatever biases and mistakes this entails. Its inevitable
discrepancies with commonsense or user-specific knowledge
can be ironed out by using machine coaching alongside
machine learning. Whether deep learning can likewise be
integrated with machine coaching remains an intriguing prospect.
A</p>
      <p>Example Construction from Theorem 7.1
C
r01
C
r01
C
r01
C
r01</p>
      <p>Y
Y
Y</p>
      <p>Y
r03
P
r03
P
P
r03
r03
P
r04
r04
r04
r04</p>
      <p>C
Q
C
Q
C
Q
C</p>
      <p>Q
r05 r12
istoteemrp.r:: ---- r11 Z r13</p>
      <p>r05 r12
istoteemrp.r:: 61 r11 Z r13</p>
      <p>r05 r12
istoteemrp.r:: 42 r11 Z r13</p>
      <p>r05 r12
istoteemrp.r:: 63 r11 Z r13</p>
      <p>C
r01
C
r01
C
r01
C
r01</p>
      <p>Y
Y
Y
Y</p>
      <p>X
X
X
X</p>
      <p>P
P
P
P
r04
r04
r04
r04</p>
      <p>C
Q
C
Q
C
Q
C</p>
      <p>Q</p>
      <p>X
X
X
X</p>
      <p>R
R
R
R</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Diakidoy et al.,
          <year>2014</year>
          ]
          <string-name>
            <surname>Irene-Anna</surname>
            <given-names>Diakidoy</given-names>
          </string-name>
          , Antonis Kakas,
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Rob</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Story Comprehension through Argumentation</article-title>
          .
          <source>In Proceedings of the 5th International Conference on Computational Models of Argument</source>
          , pages
          <fpage>31</fpage>
          -
          <lpage>42</lpage>
          ,
          <string-name>
            <given-names>Scottish</given-names>
            <surname>Highlands</surname>
          </string-name>
          , U.K.,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Diakidoy et al.,
          <year>2015</year>
          ]
          <string-name>
            <surname>Irene-Anna</surname>
            <given-names>Diakidoy</given-names>
          </string-name>
          , Antonis Kakas,
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Rob</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>STAR: A System of Argumentation for Story Comprehension and Beyond</article-title>
          .
          <source>In Proceedings of the 12th International Symposium on Logical Formalizations of Commonsense Reasoning</source>
          , pages
          <fpage>64</fpage>
          -
          <lpage>70</lpage>
          , Palo Alto, CA,
          <string-name>
            <surname>U.S.A.</surname>
          </string-name>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[Dung</source>
          , 1995]
          <string-name>
            <surname>Phan</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Dung</surname>
          </string-name>
          .
          <article-title>On the Acceptability of Arguments and its Fundamental Role in Nonmonotonic Reasoning, Logic Programming, and n-Person Games</article-title>
          .
          <source>Artificial Intelligence</source>
          ,
          <volume>77</volume>
          (
          <issue>2</issue>
          ):
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Kakas and Michael</source>
          , 2016]
          <string-name>
            <given-names>Antonis</given-names>
            <surname>Kakas</surname>
          </string-name>
          and
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <source>Cognitive Systems: Argument and Cognition. IEEE Intelligent Informatics Bulletin</source>
          ,
          <volume>17</volume>
          (
          <issue>1</issue>
          ):
          <fpage>14</fpage>
          -
          <lpage>20</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <source>[Kearns and Vazirani</source>
          , 1994]
          <string-name>
            <given-names>Michael</given-names>
            <surname>Kearns</surname>
          </string-name>
          and
          <string-name>
            <given-names>Umesh</given-names>
            <surname>Vazirani</surname>
          </string-name>
          .
          <article-title>An Introduction to Computational Learning Theory</article-title>
          . The MIT Press, Cambridge, MA, U.S.A.,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>[McCarthy</source>
          ,
          <year>1959</year>
          ]
          <article-title>John McCarthy. Programs with Common Sense</article-title>
          .
          <source>In Proceedings of the Teddington Conference on the Mechanization of Thought Processes</source>
          , pages
          <fpage>75</fpage>
          -
          <lpage>91</lpage>
          , London, England, U.K.,
          <year>1959</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>[McCarthy</source>
          ,
          <year>1998</year>
          ]
          <string-name>
            <given-names>John McCarthy. Elaboration</given-names>
            <surname>Tolerance</surname>
          </string-name>
          .
          <source>In Proceedings of the 4th International Symposium on Logical Formalizations of Commonsense Reasoning</source>
          , pages
          <fpage>198</fpage>
          -
          <lpage>216</lpage>
          , London, England, U.K.,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>[Mercier and Sperber</source>
          , 2011]
          <string-name>
            <given-names>Hugo</given-names>
            <surname>Mercier</surname>
          </string-name>
          and
          <string-name>
            <given-names>Dan</given-names>
            <surname>Sperber</surname>
          </string-name>
          .
          <article-title>Why Do Humans Reason? Arguments for an Argumentative Theory</article-title>
          .
          <source>Behavioral and Brain Sciences</source>
          ,
          <volume>34</volume>
          (
          <issue>2</issue>
          ):
          <fpage>57</fpage>
          -
          <lpage>74</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [Michael et al.,
          <year>2015</year>
          ]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          , Antonis Kakas,
          <string-name>
            <given-names>Rob</given-names>
            <surname>Miller</surname>
          </string-name>
          , and
          <article-title>Gyo¨rgy Tura´n. Cognitive Programming</article-title>
          .
          <source>In Proceedings of the 3rd International Workshop on Artificial Intelligence and Cognition</source>
          , pages
          <fpage>3</fpage>
          -
          <lpage>18</lpage>
          , Turin, Italy,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>[Michael</source>
          , 2008]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <source>Autodidactic Learning and Reasoning. Doctoral Dissertation</source>
          , Harvard University, Cambridge, MA, U.S.A.,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[Michael</source>
          , 2009]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <article-title>Reading Between the Lines</article-title>
          .
          <source>In Proceedings of the 21st International Joint Conference on Artificial Intelligence</source>
          , pages
          <fpage>1525</fpage>
          -
          <lpage>1530</lpage>
          , Pasadena, CA,
          <string-name>
            <surname>U.S.A.</surname>
          </string-name>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[Michael</source>
          , 2013]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <article-title>Machines with WebSense</article-title>
          .
          <source>In Proceedings of the 11th International Symposium on Logical Formalizations of Commonsense Reasoning</source>
          , Ayia Napa, Cyprus,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <source>[Michael</source>
          , 2014]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <article-title>Simultaneous Learning and Prediction</article-title>
          .
          <source>In Proceedings of the 14th International Conference on Principles of Knowledge Representation and Reasoning</source>
          , pages
          <fpage>348</fpage>
          -
          <lpage>357</lpage>
          , Vienna, Austria,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Michael</source>
          , 2016]
          <string-name>
            <given-names>Loizos</given-names>
            <surname>Michael</surname>
          </string-name>
          .
          <article-title>Cognitive Reasoning and Learning Mechanisms</article-title>
          .
          <source>In Proceedings of the 4th International Workshop on Artificial Intelligence and Cognition</source>
          , pages
          <fpage>2</fpage>
          -
          <lpage>23</lpage>
          , New York City, NY,
          <string-name>
            <surname>U.S.A.</surname>
          </string-name>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>[Prakken</source>
          , 2010]
          <string-name>
            <given-names>Henry</given-names>
            <surname>Prakken</surname>
          </string-name>
          .
          <article-title>An Abstract Framework for Argumentation with Structured Arguments</article-title>
          .
          <source>Argument and Computation</source>
          ,
          <volume>1</volume>
          (
          <issue>2</issue>
          ):
          <fpage>93</fpage>
          -
          <lpage>124</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>[Valiant</source>
          , 1984]
          <string-name>
            <given-names>Leslie G.</given-names>
            <surname>Valiant</surname>
          </string-name>
          .
          <article-title>A Theory of the Learnable</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>27</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1134</fpage>
          -
          <lpage>1142</lpage>
          ,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>