<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Human Like Incremental Decision Tree Algorithm</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Combining Rule Learning</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pattern Induction</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Storing Examples</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cognitive Systems Group, University of Bamberg An der Weberei 5</institution>
          ,
          <addr-line>96045 Bamberg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <abstract>
        <p>Early machine learning research was strongly interrelated with research on human category learning while later on the focus shifted to the development of algorithms with high performance. Only recently, there is a renewed interest in cognitive aspects of learning. Machine learning approaches might be able to model and explain human category learning while cognitive models might inspire new, more human like, approaches to machine learning. In cognitive science research there exist di erent theories of category learning, especially, rule-based approaches, prototypes, and exemplar-based theories. To take account of the exibility of human learning and categorization we propose a human like learning algorithm. In the algorithm we combine incremental decision tree learning, least general generalization, and storing examples for similarity-based categorization. In this paper we present rst ideas of this algorithm.</p>
      </abstract>
      <kwd-group>
        <kwd>(human) supervised category learning</kwd>
        <kwd>cognitive modeling</kwd>
        <kwd>incremental decision trees</kwd>
        <kwd>least general generalizations</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Early machine learning research strongly related to human category learning (cf.
Michalski, Carbonell, &amp; Mitchell, 1983). For example, decision tree algorithms
were inspired by research on human category learning by
        <xref ref-type="bibr" rid="ref1">Bruner, Goodnow, and
Austin (1956</xref>
        ). They investigated how humans learned conjunctive or
disjunctive categorization rules for a xed set of arti cial stimuli. It could be shown
that most humans learned such rules in an incremental way using for example
the wholist strategy, that is, they generate an initial rule which is sequentially
modi ed for new examples which do not con rm to the current rule. Based on
these ndings
        <xref ref-type="bibr" rid="ref6">Hunt, Marin, and Stone (1966</xref>
        ) developed the rst decision tree
algorithms. Subsequently, Unger and Wysotzki (1981) introduced the
incremental decision tree algorithm Cal 2 as an ideal model for human category learning
and Quinlan (1986) developed the well-known ID3.
      </p>
      <p>Starting in the 1990s machine learning research was focused on the
development of new algorithms with high performance and not on cognitive plausibility.
Therefore, machine learning and cognitive modeling research separated (Langley,
2016). For example, a main characteristic of human learning is life-long
learning, that is, incremental and cumulative learning, while in the eld of machine
learning many batch learning algorithms were developed (e.g., ID3, neuronal
network approaches, Bayes classi er; Mitchell, 1997). The lack of research in the
eld of incremental and cumulative and therefore life-long machine learning was
addressed and motivated by human learning recently but it did not focus on
cognitive plausibility of the algorithms (Thrun, 1998).</p>
      <p>
        Another main characteristic of human learning is the exible use of di erent
strategies while learning and while using the learned knowledge in
categorization tasks. That is, in human category learning there is evidence for rule-based
learning, prototype learning, and exemplar-based learning which are combined in
hybrid theories of human category learning (Kruschke, 2008). In machine
learning multistrategy learning has been focused in the 1980s and 1990s (Michalski,
1993; Langley, 2016) and Langley (2016) has suggested that the research should
continue in this line. Multistrategy learning shares characteristics with
ensemble learning. However while in ensemble learning focus is on combining di erent
models to improve classi cation performance, in multistrategy learning, focus is
on achieving human-level exibility in application of di erent knowledge
structures in di erent contexts
        <xref ref-type="bibr" rid="ref3">(Dietterich, 2002)</xref>
        .
      </p>
      <p>Besides, Langley (2016) has recommended that researcher should work
interdisciplinary because cognitive science and machine learning can mutually pro t
from each other. Machine learning approaches might be able to model and
explain human category learning while cognitive models might inspire new, more
human like, approaches to machine learning. Following these thoughts, we
currently are developing a human-inspired learning approach based on Cal 2.
Consequently, we focus on learning in supervised classi cation settings, or in other
words in categorization learning from labeled examples.</p>
      <p>Therefore, in the following section we describe Cal 2 as the incremental
rulebased basis of our algorithm and ID3 which o ers an idea for exemplar-based
human category learning. Then we introduce least general generalizations (Mitchell,
1977) which inspire a prototypical view of learning. The algorithm combining
these ideas is presented and applied to a small example in Section 3. In the last
section we discuss further steps like possible extensions of our learning algorithm
and the evaluation of the proposed approach.
2</p>
      <p>Basic Symbolic Approaches to Category Learning
In this section we rst describe relevant decision tree learning algorithms and how
they connect to human category learning theories. However, not all aspects of
human category learning can be explained with rules as generated with decision
trees. Therefore, in the second part of this section we take a closer look on least
general generalizations which could relate to the prototype theories of human
category learning.
2.1</p>
      <p>Decision Tree Learning, Rules and Exemplars
Algorithm 1 shows the decision tree learner Cal 2 (Unger &amp; Wysotzki, 1981).
The algorithm handles examples in an incremental way where the information
of previous examples is stored implicitly in the tree structure, but there is no
knowledge of previous or later examples given explicitly in each step. The
algorithm terminates successfully if all examples are classi ed correctly. Successful
termination is guaranteed for linear separable examples, that is, disjunctive
categories. A typical reason for non-disjunctivness is that the given set of features is
not su cient to discriminate the training examples. If examples are not linearly
separable, the algorithm fails in Line 9.</p>
      <p>The algorithm does not provide a strategy for feature selection. However,
the next feature (cf. Line 9) can make use of a prede ned selection strategy.
In the simplest case the algorithm could choose a feature randomly or use a
prede ned feature order. The well-known information gain as proposed in ID3
(Quinlan, 1986) is no feasible feature selection criterion for Cal 2 because to
calculate the information gain of a speci c feature all examples have to be known
in advance and therefore the learning is not incremental anymore. However, we
currently investigate incremental variants of information gain, for example, the
igain measure (Zeller &amp; Schmid, 2016). Calculating an incremental variant of
the information gain implies that the already used examples are stored explicitly
and not only implicitly in the tree structure. Storing examples explicitly refers to
the exemplar-based theories in cognitive psychology (Nosofksy &amp; Palmeri, 1997;
Jakel, Scholkopf, &amp; Wichmann, 2008).</p>
      <p>Algorithm 1 Incremental Decision Tree Algorithm Cal 2 (Unger &amp; Wysotzki,
1981)
1: procedure Cal 2(examples)
2: Start with a tree containing only * as node
3: while at least one of the examples changes the tree structure do
4: if current example is classi ed correct then
5: do nothing
6: if current example is classi ed as * then
7: replace * with the class of current example
8: if current example is classi ed wrong then
9: add the next feature in the tree
10: set for the branch with the feature value of current example
the class of current example and set * for all other branches</p>
      <p>
        The combination of Cal 2 and igain was used to model human categorization
behavior obtained in an experiment with stimuli in form of lamps (Lafond,
Lacouture, &amp; Cohen, 2009). The lamps were described by four features with two
discrete feature values each. Features are de ned for the base, upright, shade,
and top of each lamp and are given as F 1; : : : ; F 4. The categorization task in
the experiment is to decide whether a speci c feature combination belongs to
Category A or B. The knowledge structure for categorization has to be induced
by the humans from nine training examples. Category distribution follows the
in psychology often used 5-4 structure (see Table 1; Medin &amp; Scha er, 1978).
The igain of the features helped to explain di erences in decision trees when the
presentation order of input examples where di erent for the learner (Zeller &amp;
Schmid, 2016). While ordering e ects in incremental learning were addressed in
early machine learning
        <xref ref-type="bibr" rid="ref4">(cf. Fisher, 1993; Langley, 1995)</xref>
        it only recently comes
into focus in the eld of cognitive modeling of human category learning
        <xref ref-type="bibr" rid="ref2">(cf.
Carvalho &amp; Goldstone, 2015; Mathy &amp; Feldman, 2016)</xref>
        .
In human category learning there is strong evidence that, especially for natural
categories, humans do not learn feature combinations (rules) but prototypes
(Rosch &amp; Mervis, 1975). Categories over entities with discrete features are formed
by family resemblance, that is, a prototype is given by the most frequent feature
value for each feature. It could be shown that humans often search for similarity
by (implicit) counting of the occurrence of feature values within a category and
compare it with the occurrence of feature values in the other categories.
      </p>
      <p>A similar idea is described as least general generalizations in machine
learning where a pattern for a set of examples of a class is generated (cf. Mitchell,
1977). This pattern includes for each feature either a concrete feature value that
matches with all examples or a wild card if there exists more than one feature
value for this feature in the set of examples. Least general generalizations can be
formed for symbolic features, but also for structural features, terms and graphs
(cf. Schmid, Hofmann, Bader, Haberle, &amp; Schneider, 2010; Siebers, Schmid, Seu ,
Kunz, &amp; Lautenbacher, 2016).
3</p>
      <p>An Incremental Decision Tree Algorithm Including
Most Speci c Patterns and Examples
Algorithm 2 shows our human-inspired decision tree learning approach which
realizes incremental learning of rules. Furthermore, to take account of the
human exibility of the categorization process, it integrates storage of generalized
patterns and of examples. That is, dependent on the context of a
categorization task the learned knowledge structure allows to categorize a new object by
applying a rule, by pattern matching, or by similarity to known examples.</p>
      <p>Core of our algorithm is Cal 2 which can be seen in the procedure INCLUDE
(cf. Line 5) where a new example is included to a given tree. To take account
of the exibility of human categorization, the decision tree is enriched by least
general generalizations at each decision node and the leafs (cf. most speci c
pattern, for example, in Line 23). Besides, the examples are stored at the leafs
(cf., for example, Line 24). In di erence to Cal 2 our algorithm considers every
input example only once and stores it in the matching current leaf. For this rst
version of the algorithm which is restricted to disjunctive categories, termination
conditions are inherited from Cal 2. We have realized this algorithm in Prolog
and are currently investigating its ability to model reported ndings of human
category learning.</p>
      <p>The result of the algorithm depends on the category structure, the order
of the input examples, and the feature selection criterion. Figure 1 shows the
generated tree, with the 5-4 category structure mentioned above and the order
of the input examples given in Table 1 when using the igain for feature selection.
Left branches show the branches with feature value 1, right branches show the
branches with feature value 0. The least general patterns contain the feature
values with the following structure: &lt; F 1; F 2; F 3; F 4 &gt;. The ? stands for the
wildcard.
4</p>
      <p>Future Work
The proposed algorithm is work in early progress and further aspects need to be
considered: extension of the learning algorithm and evaluation of the proposed
approach.</p>
      <p>The current algorithm fails if a new example is categorized erroneously and
there is no feature available to extend the tree such that correct categorization
becomes possible. This category structure can, for example, be generated by
nondisjunctive categories. Learning scenarios involving overlapping categories are
often used in psychological studies to demonstrate that humans do not learn rules
(Kruschke, 2008). We propose to keep the rule-based structure but to sample
examples at the leaf nodes and postpone branching decisions until a certain
Algorithm 2 Incremental Decision Tree Algorithm Including Most Speci c
Patterns and Examples
1: procedure incrDTmsp(examples)
2: tree empty tree
3: for each example e 2 examples do
4: tree include(e, tree)</p>
      <p>return: tree
5: procedure include(e, tree)
6: new tree empty tree
7: if tree is empty then
8: new tree (most speci c pattern of e, label of e, e)
9: else if tree is leaf with the same class as e then
10: all examples all examples of the leaf and e
11: new tree same(all examples)
12: else if tree is leaf with a di erent class as e then
13: all examples all examples of the leaf and e
14: new tree different(all examples)
15: else
16: branch the branch of tree matching with e
17: node msp update msp of root node of branch with e
18: new branch (node msp, include(e, branch))
19: new tree substitute branch with new branch in tree
20: return: new tree
21: procedure same(all examples)
22: branch empty tree
23: msp most speci c pattern of all examples
24: branch (msp, label of all examples, all examples)
25: return: branch
26: procedure different(all examples)
27: branch branch all examples by a feature not yet used
28: for each attribute in branch do
29: if attribute has no examples then
30: node empty leaf
31: else if all examples in attribute have the same class then
32: branched examples examples of attribute
33: node same(branched examples)
34: else
35: branched examples examples of attribute
36: node msp most speci c pattern of branched examples
37: node (node msp, different(branched examples ))
38: branch add node for the attribute of the branch
39: return: branch</p>
      <p>F 4
&lt;?; ?; ?; ? &gt;
1</p>
      <p>0
amount of evidence is reached. That is, we introduce \delayed branching" that
can reduce but not solve the problem of failing.</p>
      <p>Delayed branching typically leads to decisions based on stronger evidence,
because a larger sample of examples o ers a better basis for pattern
generalization as well as for similarity-based categorization. Besides, delayed branching
might help to reduce over tting. While in the tree given in Figure 1 most leafs
are associated with patterns representing feature vectors and single examples,
trees constructed with delayed branching will less often specialize in such a way.</p>
      <p>
        A simple approach to sampling is just to de ne a threshold for the number
of examples needed before a new branch is introduced. This idea is realized in
Cal 3 (Unger &amp; Wysotzki, 1981), using a parameter for sample size. We plan to
investigate more sophisticated criteria such as prior probabilities
        <xref ref-type="bibr" rid="ref5">(Frermann &amp;
Lapata, 2016)</xref>
        , or similarity-based coherence of examples (Michalski &amp; Stepp,
1983).
      </p>
      <p>Besides, the least general generalizations does not re ect the prototype
approach in detail. That is, prototypes are build by the most frequent feature
values, while the least general generalizations represents the feature value that
are common among all examples. We plan a variant of the algorithm where
at each node the feature values with the highest frequency instead of the least
general generalizations is annotated.</p>
      <p>Additionally, it might be interesting to make restructuring of the tree and
forgetting of examples possible. In human category learning there is evidence that
humans completely reject partially learned rules and start with a new rule (Unger
&amp; Wysotzki, 1981). If we assume that wrong information of categories, that is
noise, is seldom we could handle noise by forgetting seldom seen examples and
restructure the tree. This procedure re ects representational shifts (Johansen &amp;
Palmeri, 2002).</p>
      <p>Currently, categorization of a new example|while learning and while using
the tree|is strictly guided by the branching in the decision tree. That is, patterns
and exemplars have no in uence on the learning and categorization. Hybrid
theories in human category learning take into account that humans are exible
in using di erent strategies in di erent situations (Kruschke, 2008; Nosofsky,
Palmeri, &amp; McKinley, 1994; Rosch, 1983). The factors for using one or another
information in the tree need to be investigated, and incorporated in the current
algorithm.</p>
      <p>Since our goal is to develop a human like machine learning algorithm as
well as a machine learning inspired cognitive model of categorization learning,
evaluation of the algorithm has to address performance characteristics as well
as validity as a cognitive model. We plan to take a closer look on e ciency
(time, number of training examples) and accuracy in comparison with other
machine learning approaches, such as (other) decision tree learners, random forests,
and inductive logic programming (Schmid, Zeller, Besold, Tamaddoni-Nezhad,
&amp; Muggleton, 2017).</p>
      <p>Furthermore, we want to compare the learning steps as well as the resulting
knowledge structure of our algorithm with human behavior and we plan to
predict human behavior with our algorithm. For this aim we want to select several
learning domains which have been introduced in cognitive science literature. We
are especially interested in domains where it was shown that humans make use
of di erent strategies for learning and categorization. Among these domains is
the lamp domain introduced above but also arti cial domains of letter strings,
or natural categories such as fruit (Rosch &amp; Mervis, 1975).</p>
      <p>Jakel, F., Scholkopf, B., &amp; Wichmann, F. A. (2008). Generalization and
similarity in exemplar models of categorization: Insights from machine learning.</p>
      <p>Psychonomic Bulletin &amp; Review , 15 (2), 256{271.</p>
      <p>Johansen, M. K., &amp; Palmeri, T. J. (2002). Are there representational shifts
during category learning? Cognitive Psychology , 45 , 482{553.</p>
      <p>Kruschke, J. K. (2008). Models of categorization. In R. Sun (Ed.), Computational
psychology (pp. 267{301). Cambridge University Press.</p>
      <p>Lafond, D., Lacouture, Y., &amp; Cohen, A. L. (2009). Decision-tree models of
categorization response times, choice proportions, and typicality judgments.</p>
      <p>Psychological Review , 116 , 833{855.</p>
      <p>Langley, P. W. (1995). Order e ects in incremental learning. In P. Reimann &amp;
H. Spada (Eds.), Learning in humans and machines: Towards and
interdisciplinary learning science (pp. 154{167). Oxford: Elsevier.</p>
      <p>Langley, P. W. (2016). The central role of cognition in learning. Advances in</p>
      <p>Cognitive Systems , 4 , 3{12.</p>
      <p>Mathy, F., &amp; Feldman, J. (2016). The in uence of presentation order on category
transfer. Experimental Psychology , 63 (1), 59{69.</p>
      <p>Medin, D. L., &amp; Scha er, M. M. (1978). Context theory of classi cation learning.</p>
      <p>Psychological Review , 85 , 207{238.</p>
      <p>Michalski, R. S. (1993). Multistrategy learning. Boston: Kluwer Academic</p>
      <p>Publishers.</p>
      <p>Michalski, R. S., Carbonell, J. G., &amp; Mitchell, T. M. (1983). Machine learning:</p>
      <p>An arti cial intelligence approach. Springer.</p>
      <p>Michalski, R. S., &amp; Stepp, R. (1983). Automated construction of classi cations:
Conceptual clustering versus numerical taxonomy. IEEE Transactions on
Pattern Analysis and Machine Intelligence , 5 , 219{243.</p>
      <p>Mitchell, T. M. (1977). Version spaces: A candidate elimination approach to
rule learning. In Fifth International Joint Conference on AI (pp. 305{310).</p>
      <p>Cambridge, MA: MIT Press.</p>
      <p>Mitchell, T. M. (1997). Machine learning. McGraw-Hill.</p>
      <p>Nosofksy, R. M., &amp; Palmeri, T. J. (1997). An exemplar-based random walk
model of speeded classi cation. Psychological Review , 104 (2), 266{300.
Nosofsky, R. M., Palmeri, T. J., &amp; McKinley, S. C. (1994). Rule-plus-exception
model of classi cation learning. Psychological Review , 101 (1), 53{79.
Quinlan, J. R. (1986). Induction of decision trees. Machine Learning , 1 (1),
81{106.</p>
      <p>Rosch, E. (1983). Prototype classi cation and logical classi cation: The two
systems. In E. K. Scholnick (Ed.), New trends in conceptual representation:
Challenges to Piaget's theory? (pp. 73{85). Hillsdale, NJ: Erlbaum.
Rosch, E., &amp; Mervis, C. B. (1975). Family resemblances: Studies in the internal
structure of categories. Cognitive Psychology , 7 , 573{605.</p>
      <p>Schmid, U., Hofmann, M., Bader, F., Haberle, T., &amp; Schneider, T. (2010).
Incident mining using structural prototypes. In N. Garc a-Pedrajas, F.
Herrera, C. Fyfe, J. M. Ben tez, &amp; M. Ali (Eds.), Trends in Applied
Intelligent Systems: The Twenty Third International Conference on Industrial,
Engineering &amp; Other Applications of Applied Intelligent Systems,
IEAAIE 2010, Cordoba, Spain, 01.{04. June 2010 (Vol. 6097, pp. 327{336).</p>
      <p>Springer.</p>
      <p>Schmid, U., Zeller, C., Besold, T., Tamaddoni-Nezhad, A., &amp; Muggleton, S.
(2017). How does predicate invention a ect human comprehensibility?
In J. Cussens &amp; A. Russo (Eds.), Inductive Logic Programming { 26th
International Conference, ILP 2016, London, UK, 04.-06. September 2016,
Revised Selected Papers (Vol. 10326, pp. 52{67). Springer.</p>
      <p>Siebers, M., Schmid, U., Seu , D., Kunz, M., &amp; Lautenbacher, S. (2016).
Characterizing facial expression by grammars of action unit sequences: A rst
investigation using ABL. Information Sciences , 329 , 866{875.</p>
      <p>Thrun, S. (1998). Lifelong learning algorithms. In S. Thrun &amp; L. Pratt (Eds.),</p>
      <p>Learning to learn (pp. 181{209). Boston, MA: Springer US.</p>
      <p>Unger, S., &amp; Wysotzki, F. (1981). Lernfahige Klassi zierungssyssteme (Classi
cation Systems Being Able to Learn). Berlin, Germany: Akademie-Verlag.
Zeller, C., &amp; Schmid, U. (2016). Rule learning from incremental presentation
of training examples: Reanalysis of a categorization experiment. In 13th
Biannual Conference of the German Cognitive Science Society; Bremen,
Germany, 26.{30. September 2016 (pp. 39{42).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bruner</surname>
            ,
            <given-names>J. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodnow</surname>
            ,
            <given-names>J. J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Austin</surname>
            ,
            <given-names>G. A.</given-names>
          </string-name>
          (
          <year>1956</year>
          ).
          <article-title>A study of thinking: With an appendix on language by Roger W</article-title>
          . Brown. New York, NY: Wiley.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Carvalho</surname>
            ,
            <given-names>P. F.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Goldstone</surname>
            ,
            <given-names>R. L.</given-names>
          </string-name>
          (
          <year>2015</year>
          ).
          <article-title>What you learn is more than what you see: What can sequencing e ects tell us about inductive category learning? Frontiers in Psychology , 6</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Dietterich</surname>
            ,
            <given-names>T. G.</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Ensemble learning</article-title>
          .
          <source>The handbook of brain theory and neural networks</source>
          ,
          <volume>2</volume>
          , 110{
          <fpage>125</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Fisher</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>1993</year>
          ).
          <article-title>Ordering e ects in incremental learning</article-title>
          .
          <source>In Training Issues in Incremental Learning: AAAI Spring Symposium</source>
          at Stanford University (pp.
          <volume>35</volume>
          {
          <fpage>42</fpage>
          ). Menlo Park, CA: AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Frermann</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Lapata</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>Incremental Bayesian category learning from natural language</article-title>
          .
          <source>Cognitive Science</source>
          ,
          <volume>40</volume>
          , 1333{
          <fpage>1381</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Hunt</surname>
            ,
            <given-names>E. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Stone</surname>
            ,
            <given-names>P. J.</given-names>
          </string-name>
          (
          <year>1966</year>
          ).
          <article-title>Experiments in induction</article-title>
          . New York, NY: Academic Press.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>