<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>On Efficiency of Learning: A Framework and Justification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>JLQG LFK B$FKD</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>A conceptual framework, whose goal is the improvement of efficiency of machine learning, is presented. The framework is designed in a broader context of problem solver (PS). The design is solved as an integration of all basic cognitive functions and as a software-engineering problem. Many (one hundred) requirements imposed on PS are considered. The most important of them are the object-oriented nature of the PS environment, reflexivity of PS, and central role of tool and shifted border.1</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        This is a position paper, presenting a framework for machine
learning (ML), reacting on some reviewers’ reamrks. The
framework is very briefly described and justified. More details can
be found in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        My extensive evaluation of current ML reveals that the
outstandingly dominant part of ML is devoted to simple learning. I
consider this harmful and the framework is a trail to remedy this
situation. I would roughly characterize simple learning as a
oneshot creation of knowledge, describing one function of a small part
of an environment, and applied in a manually selected part of
environment. On the other hand, I would characterize complex
learning as a cumulative creation of knowledge with many parts,
which describes many related functions of whole environment. It is
structured, preferably in object-oriented way, applied in different
parts of different cognitive functions, applied in the whole real
world. These are two extremes. The state-of-the-art is, of course,
somewhere in-between, much closer to simple learning. It contrasts
with some other AI areas, like knowledge engineering, where we
do use complex knowledge structures [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        The application of this framework to learning ontologies is
highly relevant: 1) Both in ontologies and in the framework, there
is a common stress put on knowledge structure, reuse, object
orientation. 2) Learning ontologies can be approached both top-down
and bottom-up. If there are no worked out firm foundations then
they should be done first. This is also the aim of the framework. 3)
Both learning and ontology areas can be integrated. For example,
learning could accept complex knowledge structures; ontologies
could accept an approach to approximate knowledge. This would
modify e.g. some design “principles” that “have been proved useful
in the development of ontologies”, like clarity, completeness, and
coherence [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. On the other hand, this may explain e.g. one
______________________________________________________
1 ,QVWLWXWH RI ,QIRUPDWLRQ 7KHRU\ DQG $XWRPDWLRQ 3RG YRGi UHQVNRX Y t
182 08 Prague 8, Czech Republic, email: bucha@utia.cas.cz
often-mentioned problem, that “concrete elements are in many
cases practically more usable then abstract ones” [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>My approach to learning can be characterized concisely as
"learn from learning". This means that it is suitable to use
knowledge, which we have gathered during the exploration of
learning (phenomenon), and apply it as meta-knowledge in the
design of improved learning (algorithm). The body of this
knowledge is vast. What parts should be used? My approach is to use all
meta-knowledge that can be integrated into an efficient learning
system. It is also the solution of the efficiency of learning: Each
piece of this meta-knowledge should support somehow this
efficiency. To design and build a corresponding system is a
formidable task. However, realize four things: 1) There is a great
redundancy in AI knowledge. 2) There is a difference between the
building of a learning system and fully developed problem solving
system. I am concentrating on problem solving (PS), which has
not yet gathered a massive amount of knowledge, but which can
gather it. I am using the frame of PS to stress the importance of the
context of learning. 3) There are approaches to the design of
complex systems; I am using the software-engineering approach and
the object-oriented methodology. 4) There is no evidence that this
task is unsolvable.</p>
      <p>The main objection against the proposed framework is why
should we deal with something what is so unjustified, unproven,
unimplemented? Haven’t we done it already many times without
any result? Is not here a rule that every proposal should be at least
partially verified by some prototype? I understand this but do not
agree. Why? 1) Critical claims are usually very vague. 2) Even
negative experience is useful, it exists, and I’m trying to use it. 3)
Prototype verification has its value; however, it should not be
overrated. 4) Not only a detailed analysis, the synthesis can bring new
knowledge too. To prototype a synthesized system is harder and
needs more careful design. 5) It is in contradiction with the
experience of software engineering: design should be verified
step-bystep. 6) AI should realize the shortage of synthesis, its reasons and
consequences, and should try to solve it.</p>
      <p>
        In the design, I primarily use the viewpoint of efficiency of
learning [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. There are many other possibilities, e.g. the viewpoint
of design of problem solver, integration of cognitive functions,
design of agents, essences of intelligence. These viewpoints will
manifest in a requirement analysis. This is usually the first design
step of a software-engineering approach. Doing this, I have
analyzed many projects, approaches and surveys, and I have extracted
and classified key requirements either implemented or gained as
conclusions from project experiences. I have gathered more then
one hundred requirements. Majority can be found in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. For some
projects and approaches, I estimated the following fulfillment of
these requirements: Minsky [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] 11%, CYC [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] 7%, PRODIGY
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] 18%, Soar [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] 23%, Brooks methodology [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] 33%, all these
together 65%; my framework covers 100%. This is one reason why
I consider the framework original.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>FRAMEWORK</title>
      <p>First, I am presenting basic assumptions about the learning
context. They form a skeleton for the requirements:
• Environment is a network of many heterogeneous, dynamic
and even uncertain objects.
• PS is an object of environment. PS has a pre-specified goal. If
PS is not in goal state, it has problem, and that should be
solved by PS. PS should control environment to reach goal.
PS control is accomplished by means of its cognitive
functions (implementation, identification, reasoning,
learning, knowledge base (KB), self-control, and initialization).</p>
      <p>The cognitive functions have also their (sub) goals.
• Knowledge is an approximate description (of the behavior) of
object; knowledge usage increases the average probability of
success of reaching PS goal.
• Environment is relatively very stable. During an interval of PS
work, very little part of environment is changing.
• PS (learning) should cope with its complexity.</p>
      <p>
        The framework itself is outlined in two figures. To fulfill the
requirements, PS should be a knowledge-based system and KB
should have the structure outlined in Figure 1. (For graphical
notation see [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]).
      </p>
      <p>Learning is not a stand-alone function. It cannot work without
cooperation with other PS cognitive functions. PS should have the
structure outlined in Figure 2. Notice that it is still a partial view,
e.g. relations to important self-control and initialization functions
are not rendered here. Notice also, there is only one common
concept, description, for description of various PS entities, i.e.
classes of environment objects, specific objects like model, plan,
value, variable, state, goal, meta-model, meta-plan, etc.
3</p>
    </sec>
    <sec id="sec-3">
      <title>JUSTIFICATION</title>
      <p>Here I show two examples of requirements, tool and shifted border,
,GHQWLILFDWLRQ
5HODWLRQ W\SHV
,QVWDQWLDWLRQ
6LPSOLILFDWLRQ
$JJUHJDWLRQ
$VVRFLDWLRQ
(TXLYDOHQFH
$QFKRU
,V JRDO RI
DWWULEXWH</p>
      <p>LQVWDQWLDWLRQ
5HODWLRQ
DWWULEXWH</p>
      <p>’HVFULSWLRQ</p>
      <p>VXESDUW
LQVWDQWLDWLRQ</p>
      <p>LQSXW</p>
      <p>RXWSXW</p>
      <p>VWDWH
,QWHUIDFH</p>
      <p>
        UXOH
5XOH
and homogeneity, and briefly explain how they are implemented in
the framework. More can be found in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] where e.g. justifications
cover differences between my approach and that of Brooks [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        In a simple environment, to identify (the only one and simple)
object and its state need not be a problem. To learn description of
such object, plan a solution in this environment described as such
object, and implement plan need not be a problem. Self-control
would not be necessary and KB would be trivial. To implement it,
the existing cognitive functions or ontologies from AI can be used
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. However, the real environment is not simple. It is
necessary to extend the simple cognitive functions to cover the
work with object-oriented environment, both deterministic and
stochastic objects, with both
static and dynamic objects etc.
      </p>
      <p>Some aspects of object-oriented
nature of environment are
solved by existing techniques,
0RGHO e.g. instantiation, simplification,
and anchor relations in
learning. To solve other aspects, e.g.</p>
      <p>association relations, the
con5HDVRQLQJ cept of tool is used: Let us
consider PS, its environment,
some tool, as an object of
environment, and rest of
environment. Let PS have inputs resp.
3ODQ outputs to perceive resp.
influence environment; the part of
them connected with tool is x
resp. y. x and y together form
the part of the real border
be,QSXW
2XWSXW
2XWSXW
(QYLURQPHQW
tween PS and environment, i.e. border between PS and tool. x' and
y' form the part of border between tool and rest of environment. x'
and y' are accessible to PS during learning but not fully accessible
to PS later. Let us assume that PS can learn characteristics of tool
in such a way that it can later 1) identify tool, 2) model (specify)
the sequence of values of x' and y', if it knows the sequence of
values of x and y. x' and y' are called modeled inputs and outputs.
The equality of modeled inputs and outputs, both in environment
and in PS, can be interpreted as a shift of border between
environment and PS. We can interpret e.g. teacher, society,
communication with similar PS's, and an approach to learning as
tools.</p>
      <p>
        There are approaches using e.g. genetic algorithms for
self-control [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. They offer evidence of improved efficiency. They use
simple learning, in my terminology. Why not generalize this and
use complex learning? It should be even more efficient. This is
enabled by my assumption: PS is part of environment.
Consequently, PS should control itself. It can do it with all its power. For
self-control to be practicable, PS must be homogeneous.
Therefore, it must not be necessary to design and use special control
mechanism for each PS part. KB should be homogeneous; this is
most important. However, some homogeneity is beneficial and
possible for cognitive functions also.
      </p>
      <p>Let us consider the possibility to implement, in the framework,
two types of descriptions, i.e. description of object of environment
and description of PS plan. These two descriptions should have
had many similar properties, e.g. they should have been general,
dynamic, complex, hierarchically organized, object-oriented, might
have been approximate, uncertain, can utilize tools etc. How might
they have differentiated? The only difference is that (sub-)plan
should be directly related to some (sub-)goal. There is also another
viewpoint: PS and its parts have their behaviors. They are
described - generated by plans. PS and its parts are parts of
environment. For both, it is suitable to have common description.
As a description of external object and plan share the majority of
features, it is suitable to describe them in a common way.
4</p>
    </sec>
    <sec id="sec-4">
      <title>CONCLUSION</title>
      <p>This paper proposes a framework to improve the current level of
efficiency of learning. The approach shows feasibility in a sense
that, on the conceptual level, the PS could be designed to satisfy all
challenging requirements. The efficiency of learning has been
analyzed and designed on a conceptual level. This is the only
possible (and necessary) beginning of a long-term project. My
intention is to continue with a detailed analysis and design of
cognitive functions and KB (knowledge base), and their
implementation using a contemporary software-engineering approach, i.e.
utilizing object-oriented, computer-aided software-engineering
approaches and tools and using contemporary knowledge about
cognitive functions.</p>
      <p>The idea to integrate all AI concepts into one system is not new.
However, integration is often preached, but seldom practiced. My
contribution is in doing the "first (small) step" toward such
integration. The framework uses many concepts, but not many new
ones. The exceptions are “learn from learning”, shifted border, that
learning can create all parts of KB, and knowledge
homogenization.</p>
      <p>In higher cognitive processes, self-reflection plays an important
role. AI is a cognitive process of incremental understanding the
phenomenon of intelligence. What is the state of self-reflection in
AI? According to my assessment, the field of AI is just at the
beginning of its conscious self-reflection. What do I mean by this?
I mean that parts of AI community - yet still small - are just
starting to use specific benefits of AI work, i.e. they are starting to
interpret the results of the AI process - the knowledge produced by
AI - to control the AI process itself. I have tried to use this
knowledge already in this paper.</p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGEMENTS</title>
      <p>7KLV ZRUN KDV EHHQ SDUWLDOO\ VXSSRUWHG E\ WKH JUDQW *$ 5 1R
102/99/1564. I also acknowledge the help of the reviewers, my
colleagues J. Grim and A. Halousková, M. Kárný, my daughters,
and niece.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <article-title>- % $FKD 2Q (IILFLHQF\ RI / HDUQLQJ $ )UD PHZRUN DQG Justification</article-title>
          , Research Report No.
          <year>1983</year>
          , UTIA, Prague,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <article-title>- % $FKD 2Q (IILFLHQF\ RI / HDUQLQJ $</article-title>
          )
          <source>UD PHZRUN 5HVHDUFK Report No</source>
          .
          <year>1969</year>
          , UTIA, Prague,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3] -
          <fpage>%</fpage>
          $FKD ,
          <string-name>
            <surname>QWHJUDWLRQ RI</surname>
            <given-names>DOO</given-names>
          </string-name>
          &amp;
          <article-title>RJQLWLYH )XQFWLRQV 5HTXL UHPHQWV http://www</article-title>
          .utia.cas.cz/AS_dept/bucha/ recent.html,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. R.</given-names>
            <surname>Benjamins</surname>
          </string-name>
          ,
          <article-title>Overview of Knowledge Sharing and Reuse Components: Ontologies and Problem-Solving Methods</article-title>
          .
          <source>In [18], 1-1 - 1-15</source>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Console</surname>
          </string-name>
          , and
          <string-name>
            <given-names>O.</given-names>
            <surname>Dressler</surname>
          </string-name>
          ,
          <article-title>Model-based Diagnosis in the Real World: Lessons Learned and Challenges Remaining</article-title>
          . In T. Dean (Ed.),
          <source>IJCAI-99, Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence</source>
          , Stockholm, Sweden. San Francisco, Morgan Kaufmann Publishers,
          <fpage>1393</fpage>
          -
          <lpage>1400</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Gruber</surname>
          </string-name>
          ,
          <article-title>Towards principles for the design of ontologies used for knowledge sharing</article-title>
          .
          <source>International Journal of Human-Computer Studies</source>
          ,
          <volume>43</volume>
          ,
          <fpage>907</fpage>
          -
          <lpage>928</lpage>
          , (
          <year>1995</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Barnaras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Laresgoiti</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Corera</surname>
          </string-name>
          ,
          <article-title>Building and Reusing Ontologies for Electrical Network Applications</article-title>
          , In [19],
          <fpage>288</fpage>
          -
          <lpage>302</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.S.</given-names>
            <surname>Pinto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Pérez</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.P.</given-names>
            <surname>Martins</surname>
          </string-name>
          , Some Issues on Ontology Integration.
          <source>In [18], 7-1 - 7-12</source>
          , (
          <year>1999</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Thrun</surname>
          </string-name>
          , C. Faloutsos, T. Mitchell, L. Wasserman,
          <source>Automated Learning and Discovery: State-Of-The-Art and Research Topics in a Rapidly Growing Field</source>
          , CMU, Pittsburgh,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Minsky</surname>
          </string-name>
          ,
          <source>Steps toward Artificial Intelligence, Proc. of IRE</source>
          ,
          <volume>49</volume>
          ,
          <fpage>8</fpage>
          -
          <lpage>30</lpage>
          (
          <year>1961</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>D. B. Lenat</surname>
            , The Dimensions of Context-Space,
            <given-names>CYCORP</given-names>
          </string-name>
          , Austin, http://www.cyc.com/,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Veloso</surname>
          </string-name>
          et al.,
          <article-title>Integrating Planning and Learning: The PRODIGY Architecture</article-title>
          ,
          <source>Journal of Experimental and Theoretical Artificial Intelligence</source>
          ,
          <volume>7</volume>
          , no.
          <issue>1</issue>
          , (
          <year>1995</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13] The Soar Project at Carnegie Mellon University, http://www.cs. cmu.edu/afs/cs.cmu.edu/project/soar/public/www/home-page.html,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Brooks</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Breazeal (Ferrell)</article-title>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Irie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kemp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marjanovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Scassellat</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Williamson</surname>
          </string-name>
          , Alternate Essences of Intelligence, http://www.ai.mit.edu/people/brooks/papers/groupAAAI-98.pdf.
          <source>Also in AAAI-98</source>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>G.</given-names>
            <surname>Booch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rumbaugh</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Jacobson</surname>
          </string-name>
          ,
          <article-title>The Unified Modeling Language User Guide</article-title>
          , Reading: Addison-Wesley,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Mitchell</surname>
          </string-name>
          , Machine Learning,
          <string-name>
            <surname>McGraw-Hill Companies</surname>
            ,
            <given-names>N.Y.</given-names>
          </string-name>
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>P.</given-names>
            <surname>Collard</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Gaspard</surname>
          </string-name>
          ,
          <article-title>"Royal-Road" Landscapes for a Dual Genetic Algorithm</article-title>
          .
          <source>In [19]</source>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>V. R.</given-names>
            <surname>Benjamins</surname>
          </string-name>
          (Ed.),
          <source>Proceedings of the IJCAI-99 Workshop on Ontologies and Problem-Solving Methods: Lessons Learned and Future Trends</source>
          , Stockholm,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wahlster</surname>
          </string-name>
          (Ed.),
          <source>ECAI 96, 12th European Conference on Artificial Intelligence, J</source>
          . Wiley &amp; Sons, Ltd,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>