<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>A. Szabari);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Generalized decision directed acyclic graphs and their connection with Formal Concept Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Šimon Horvát</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>L'ubomír Antoni</string-name>
          <email>lubomir.antoni@upjs.sk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ondrej Krídlo</string-name>
          <email>ondrej.kridlo@upjs.sk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexander Szabari</string-name>
          <email>alexander.szabari@upjs.sk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stanislav Krajči</string-name>
          <email>stanislav.krajci@upjs.sk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Computer Science, Faculty of Science, Pavol Jozef Šafárik University in Košice</institution>
          ,
          <addr-line>041 80 Košice</addr-line>
          ,
          <country country="SK">Slovakia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>We propose a generalization of decision jungles from a binary decision directed acyclic graph to a generalized decision directed acyclic graph. We describe the classification method based on a generalized decision directed acyclic graph. Moreover, we explore the properties of our proposed method and illustrate it by example. We present our experiments on several datasets and provide the comparison of our results with related studies. The comparison of the generalized decision directed acyclic graph with the concept lattice of approval for a loan concludes our paper.</p>
      </abstract>
      <kwd-group>
        <kwd>Classification task</kwd>
        <kwd>Decision directed acyclic graphs</kwd>
        <kwd>Concept lattice</kwd>
        <kwd>Generalization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>Decision jungles (Shotton et al. [30]) provide the compact ensemble method for classification</title>
        <p>
          tasks which extend the popular classification methods of decision trees [
          <xref ref-type="bibr" rid="ref4">27, 28, 29, 4</xref>
          ] and
random forests [
          <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
          ]. In particular, decision jungles represented by the ensemble of decision
directed acyclic graphs require less memory in comparison with binary decision trees.
        </p>
      </sec>
      <sec id="sec-1-2">
        <title>Decision directed acyclic graphs for classification tasks were explored in several recent</title>
        <p>
          studies [
          <xref ref-type="bibr" rid="ref8">8, 17, 16, 21, 22, 23, 24, 26</xref>
          ] including their connections with operation of convolution.
Laptev and Buhmann [21] proposed a novel classification method for images, which is based
on so-called transformation-invariant convolutional jungles. Ioannou et al. [17] explored the
connections between decision jungles and convolutional neural networks. They proposed the
hybrid conditional network which is based on trees and convolutional neural networks. Ignatov
and Ignatov [16] proposed an algorithm of decision streams given by tree structure which is
generated by merging nodes from various branches based on their two-sample test statistics
similarity.
        </p>
      </sec>
      <sec id="sec-1-3">
        <title>Moreover, the connections between trees structures and Formal concept analysis were recently investigated in [20, 15, 2, 12, 19]. Kuznetsov [20] described the model of learning from positive</title>
        <p>
          and negative examples in the terms of Formal concept analysis. Guillas et al. [15] presented
the first structural links between dichotomic lattices and decision trees. Bělohlávek et al. [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]
proposed a novel method of induction of decision trees. The concept lattice is seen as the
collection of overlapping decision trees in this approach. Links between random forests and
pattern structures in Formal Concept Analysis were investigated by Krause et al. [19], as well.
        </p>
      </sec>
      <sec id="sec-1-4">
        <title>Dudyrev and Kuznetsov [12] proposed a combination of concept lattices and decision trees.</title>
      </sec>
      <sec id="sec-1-5">
        <title>In this paper, we propose a generalization of decision jungles from the binary decision directed</title>
        <p>acyclic graphs to the generalized decision directed acyclic graphs. In Section 2, we present
the basic notions of classifications tasks for categorical attributes. Our novel method for the
construction of a classifier for a generalized oriented acyclic graph is proposed in Section 3.</p>
      </sec>
      <sec id="sec-1-6">
        <title>We present the experiments on our novel method and a comparison of our results with other related studies in Section 4. Connections of our approach with Formal Concept Analysis are illustrated in Section 5.</title>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Classification tasks with nominal attributes</title>
      <sec id="sec-2-1">
        <title>In this section, we will formally describe the basic notions of classification tasks with nominal (categorical) input attributes.</title>
        <p>Let  be a set and let  ∉  . Consider a set   for each  ∈  and a set   for  ∉  . Then,
∏∈   denotes a set of all functions  with a domain  such that  () ∈   for all  ∈  .</p>
      </sec>
      <sec id="sec-2-2">
        <title>For classification tasks with nominal input attributes, the set  represents the set of input</title>
        <p>attributes and  is a target attribute. The system of sets (  ∶  ∈ ) represents the unordered
sets of values (or so-called classes, or categories) of input attributes. The set   is an unordered
set of values (or so-called classes, or categories) of the target attribute. The number of values in
  is usually low in classification tasks.</p>
        <p>In Table 1, the example for  = { Income, Savings, Self-employed, Married} is shown. For
 ∉  , the target attribute of loan approval is assigned, whereby the sets  Income =
{low, average, high},  Savings = {low, high},  Self-employed = {yes, no},  Married = {yes, no}, and
 LoanApproved = {1, 0} are considered.</p>
        <p>For  ∈ ℕ , a set  given by
 = {(  1,  1), … , (  ,   )} ⊆ ∏   ×  
∈
is called the training set with  objects (also called the labeled objects). In our example from
Table 1, we have  = 16 . In some cases, it can hold that   =   for some ,  ∈ {1, 2, … , },  ≠  ,
but   ≠   . It means that we have two people with the same values of input attributes. However,
both have diferent values of the target attribute. This is not the case in our example.</p>
        <p>For  ∈ ℕ such that  &lt;  , a set  given by
 = {  +1 , … ,   } ⊆ ∏  
∈
is called the testing set with  −  objects (also called the unlabeled objects). In Table 2, we
illustrate the possible testing set for our example of loan approval.</p>
        <p>
          ∶  →  
 ∶  → [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ]   .
is called a classifier. For its possible fuzzy modification, we can consider the function
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>In the following sections, we will present the construction of a novel classifier based on generalized decision acyclic graphs and describe the possible connections of this classifier with concept lattice in Formal Concept Analysis.</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Generalized decision directed acyclic graph for classification tasks</title>
      <sec id="sec-3-1">
        <title>Shotton at al. [30] and Pohlen [26] provide the ensemble of classifiers that is based on the rooted</title>
        <p>binary decision directed acyclic graphs [24]. A rooted binary decision directed acyclic graph is
a directed graph with one root node (i.e., in-degree 0), multiple split nodes (in-degree at least 1,
out-degree 2), and multiple leaf nodes (in-degree at least 1, out-degree 0).</p>
        <p>For constructing a classifier based on a decision directed acyclic graph, Shotton et al. [ 30]
optimize the parameters of the attribute for split node, the numerical threshold for split node,
left child of a split node, and right child node of a split node for root and each internal node of the
decision directed acyclic graph. They optimize these parameters by LSearch or ClusterSearch
algorithms regarding the numerical attributes. Pohlen [26] noted that an attribute map can be
used to process the categorical data in this environment, as well. However, it can lead to the
growing depth of a binary decision directed acyclic graph.</p>
      </sec>
      <sec id="sec-3-2">
        <title>These approaches inspire us to propose the generalized version of decision directed acyclic</title>
        <p>graph to process the categorical data without attribute transformation. Moreover, we aim to
eliminate the growing depth of a binary decision directed acyclic graph by a generalized decision
directed acyclic graph. Finally, we will compare the performance and memory eficiency of our
proposed solution with those from [30, 26].</p>
      </sec>
      <sec id="sec-3-3">
        <title>First, the definition of a rooted generalized directed acyclic graph follows.</title>
        <p>Definition 1.
that:</p>
        <p>A rooted generalized directed acyclic graph is a directed graph  = ( , )
such
(1) A unique root node  ∈  has zero incoming edges and at least two outgoing edges;
(2) There exist no directed cycles in  ;
(3) There exists a directed path from a root node  ∈  to each other node  ∈  ;
(4) Each node  ∈  ⧵ { } has either zero outgoing edges (then it is called a leaf) or has at least
two outgoing edges (then it is called an internal node), and at least one incoming edge.</p>
      </sec>
      <sec id="sec-3-4">
        <title>The comparison of a rooted generalized directed acyclic graph with a rooted binary directed acyclic graph applied by [26, 30] is shown in Table 3.</title>
      </sec>
      <sec id="sec-3-5">
        <title>In the following part, we will define a generalized decision directed acyclic graph.</title>
        <p>Definition 2. Let  be a set of  categorical attributes and let  ∉  is a target attribute. A
generalized decision directed acyclic graph is a rooted generalized directed acyclic graph  = ( , )
such that
(1) All directed paths from the root node  to a node  have the same lengths;
(2) The triple (  ,   ,   ) is assigned for the root node and for each internal node  ∈  , whereby
  ∶   →     is an injective mapping;
(a)   ∈ {1, 2, … , } is an index of the attribute,
(b)   ∈ {1, 2, … , |  |} is a class index of target attribute  ,
(c)   = { ∈  ∶ ( , ) ∈ }
is a set of children nodes of  such that |  | ≤ |   
| and
(3) A class index   ∈ {1, 2, … , |  |} of the target attribute  is assigned for each leaf node  ∈  .</p>
      </sec>
      <sec id="sec-3-6">
        <title>In the following subsection, we will describe the algorithm for the generation of the general</title>
        <p>ized decision directed acyclic graph from a given training set  described in Section 2.
3.1. Algorithm for learning a generalized decision directed acyclic graph</p>
      </sec>
      <sec id="sec-3-7">
        <title>Suppose that we have a training set  and we aim to learn a generalized decision directed acyclic</title>
        <p>graph from  . By Definition</p>
      </sec>
      <sec id="sec-3-8">
        <title>2, we aim to find a triple</title>
        <p>internal node  ∈  . Moreover, we aim to find a class index 
(  ,   ,   ) for the root node and each
 for each leaf node  ∈  .</p>
      </sec>
      <sec id="sec-3-9">
        <title>In Algorithm 1, we describe our proposed procedure for learning a generalized decision</title>
        <p>directed acyclic graph from a training set  . Due to a limited number of pages, we will explain
only the substantial parts of our algorithm in this paper. Additional important notions, the
formal descriptions of mappings and their properties will be included in the extended version
of our paper.</p>
      </sec>
      <sec id="sec-3-10">
        <title>Without loss of generality, we can suppose that  is a set of attributes,  is a target attribute,</title>
        <p>and   is a set of all parent nodes at some fixed level  ∈  (see Figure 1). Let   be a set of all
child nodes for all  ∈</p>
        <p>.</p>
        <sec id="sec-3-10-1">
          <title>Let   denote the set of all labeled objects from a training set  that reaches a node  ∈</title>
          <p>.</p>
          <p>The procedure AttributeIndexOfNode assigns the attribute index   to  (last but one line
of Figure 1). This assignment is based on the entropy function ℎ with a discrete probability
distribution    . Note that for  ∈   and  ∈    
, a set

 = {( ,  ) ∈   ∶    = }
is a subset of a training set  corresponding to a given child node  and an attribute value 
(see Algorithm 1). The procedure TargetClassIndexOfNode assigns the class index   of the

target attribute  to  (last line of Figure 1). The class index of the target attribute with the
maximal number of objects in   is selected. If there exist two or more classes with the same
maximal number, we consider the class with the least index.</p>
        </sec>
        <sec id="sec-3-10-2">
          <title>Finally, the parent nodes  +1 for level  + 1 are obtained from a system (  ∶  ∈   ) by</title>
          <p>procedure MergeChildNodes with the same attribute and class indices at a level  (Figure 2).</p>
        </sec>
      </sec>
      <sec id="sec-3-11">
        <title>We present the full procedure for learning a generalized decision directed acyclic graph in</title>
        <sec id="sec-3-11-1">
          <title>Algorithm 1. Note that the set of parents nodes that are internal at a level  is denoted   .</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>We conducted several experiments to provide the performance and memory eficiency of our proposed method in comparison with other recent studies, mainly with binary decision directed</title>
        <p>acyclic graphs and decision forests [30, 26]. For comparison of memory consumption of our
method with other studies, we use the estimation of the model sizes which are described in
[30, 26]. The complete table of estimated memory consumption for nodes and leaves of various
structures and approaches (including our approach) is shown in Table 4.</p>
      </sec>
      <sec id="sec-4-2">
        <title>We tested the performance of our method and its memory eficiency on four datasets (Connect4, Letter Recognition, Shuttle, and USPS) from UCI Machine Learning Repository [10] to compare our results with [30, 26].</title>
      </sec>
      <sec id="sec-4-3">
        <title>The final results of random forest, the ensemble of 15 decision directed acyclic graphs, and our ensemble of 15 generalized decision directed acyclic graphs on memory consumption</title>
      </sec>
      <sec id="sec-4-4">
        <title>Algorithm 1: Procedure for learning a generalized decision directed acyclic graph</title>
      </sec>
      <sec id="sec-4-5">
        <title>Data: Set of attributes  , target  , training set</title>
        <p>Result: Generalized decision directed acyclic graph  matching 
while   ≠ ∅ do
  ← ∅;
for  ∈   do
if ℎ(   ) ≠ 0 then
 ← |{ ∈ 
  ← { 
for  ∈   do</p>
        <p>∶   ≠ ∅}|;
∶  ∈ {1, … , } };


 ← AttributeIndexOfNode() ;
 ← TargetClassIndexOfNode() ;
end
end
if |{  ∶  ∈   }| =  then</p>
        <p>←   ∪ {} ;
end
end
 ←  + 1 ;
end
 +1 ← MergeChildNodes({  ∶  ∈   });
/Level of graph/
/Parent node at level 1/</p>
        <p>/Internal nodes/
/Number of child nodes/
/New child nodes/
and test accuracy are shown in Table 5. For Accuracy, we present the arithmetic mean of
the test accuracies which were obtained by ensembles of 15 classifiers with a random 5-fold
cross-validation.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Connections with Formal Concept Analysis</title>
      <sec id="sec-5-1">
        <title>The generalized decision directed acyclic graph for our loan approval example (see Table 1) is</title>
        <p>illustrated in Figure 3. Note that the proportion of two numbers inside the nodes corresponds
to the proportion of the count of labeled objects by two classes of target attribute (categories 1
or 0). The sets described in the text on the left side of the nodes in Figure</p>
      </sec>
      <sec id="sec-5-2">
        <title>3 correspond to the</title>
        <p>indices of people (first column in Table 1) with category 1 of loan approval.</p>
        <p>
          Regarding the interesting research results in the area of Formal Concept Analysis [
          <xref ref-type="bibr" rid="ref13">14, 13</xref>
          ],
be considered as (binary) formal context ⟨, 
× ∪ {⟨, 1⟩}, ⟩
the positive examples from a training set  (i.e., the labeled objects with   = 1 in Table 1) can
such that 1 ∈   , (, ⟨, 1⟩) = 1
 × = ⋃ ({} ×   ).
        </p>
        <p>∈</p>
      </sec>
      <sec id="sec-5-3">
        <title>Note that a conceptual scale [14] or several fuzzy extensions in Formal Concept Analysis</title>
        <p>
          [
          <xref ref-type="bibr" rid="ref1 ref11 ref3 ref7 ref9">1, 3, 7, 9, 11, 18, 25</xref>
          ] can be also applied here.
Variable
        </p>
        <p>Size
in bytes</p>
        <p>Total size
in bytes
















 1 ∈  
neighbor
  ( 1)
4
8
4
4
8
4
4
4
4
4
4
Ensemble of
15 decision
directed acyclic
graphs [26, 30]</p>
        <p>Ensemble of
15 generalized decision
directed acyclic
graphs (our paper)
82.0%
1.51 MB
95.7%
1.50 MB
99.9%
0.11 MB
96.0%
0.30 MB
16
20
16
4.|  |
82.1%
0.22 MB
96.3%
0.21 MB
99.9%
0.005 MB
90.0%
0.035 MB</p>
      </sec>
      <sec id="sec-5-4">
        <title>In our example from Table 1, the set of objects  contains people with loan approval</title>
        <p>equals 1 (rows 1 – 9), and the set of attributes includes pairs ⟨income, high⟩, ⟨income, low⟩,
⟨savings, high⟩, ⟨savings, low⟩, ⟨self-employed, yes⟩, ⟨self-employed, no⟩, ⟨married, yes⟩,
⟨married, no⟩, ⟨loan approved, 1⟩. The corresponding concept lattice of loan approval is shown
in Figure 4.</p>
        <p>The circles of concept lattice filled with black color (Figure 4) correspond to the nodes of
generalized decision directed acyclic graph (Figure 3) with the most of node people who are
approved for a loan. For example, the extent of formal concept {1, 2, 3, 4} (highlighted by a
black circle on the right part of concept lattice) corresponds to the path of high income and
high savings in the generalized decision directed acyclic graph. Moreover, the intent of the
mentioned formal concept is {⟨income, high⟩, ⟨savings, high⟩}, as well. However, the concept
lattice provides the possibility to explore other possible groups of people which are not obtained
by generalized directed acyclic graphs since the entropy function returns only the best splitting
attribute for each node.</p>
      </sec>
      <sec id="sec-5-5">
        <title>The second concept lattice of Table 1 can be constructed by taking the formal context</title>
        <p>⟨,  × ∪ {⟨, 0⟩}, ⟩ such that  is a set of labeled objects with   = 0. Then, we obtain the
formal concepts that can correspond to the paths of people who are not approved for a loan by
generalized decision directed acyclic graph.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and future work</title>
      <p>In this paper, we proposed a classification method of generalized decision directed acyclic
graphs which is a generalization of decision jungles from the rooted binary directed acyclic
graph to rooted generalized directed acyclic graph. In the extended version of our paper, we will
present the properties of generalized decision directed acyclic graphs in general. Moreover, we
will emphasize the description of the functions used in our algorithm in more detail and their
importance for our experimental results. The more detailed connections between generalized
decision directed acyclic graphs and Formal Concept Analysis will be explored in the extended
version of the paper, as well.</p>
      <sec id="sec-6-1">
        <title>Our future research aims to propose the algorithms for the generalized decision directed acyclic graphs which can be applied for regression tasks with a numerical target attribute. Moreover, another possibility is to apply the genetic algorithms in the search of optimal parameters, optimize the learning process, and improve the performance of algorithms for testing datasets.</title>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <sec id="sec-7-1">
        <title>This work was partially supported by the Scientific Grant Agency of the Ministry of Education,</title>
      </sec>
      <sec id="sec-7-2">
        <title>Science, Research and Sport of the Slovak Republic under contract VEGA 1/0645/22 Proposal</title>
        <p>of novel methods in the field of Formal Concept Analysis and their application. This article
was partially supported by the project KEGA 012UPJŠ-4/2021. This work was supported by the</p>
      </sec>
      <sec id="sec-7-3">
        <title>Slovak Research and Development Agency under contract No. APVV-21-0468.</title>
        <p>[14] Ganter, G., Wille, R.: Formal Concept Analysis, Mathematical Foundation. Springer Verlag
(1999)
[15] Guillas, S., Bertet, K., Ogier, J.M., Girard, N.: Some links between decision tree and
dichotomic lattice. In: Bělohlávek, R., Kuznetsov, S. O. (eds.) CLA 2008, Proceedings of the</p>
      </sec>
      <sec id="sec-7-4">
        <title>Sixth International Conference on Concept Lattices and Their Applications, pp. 193–205.</title>
      </sec>
      <sec id="sec-7-5">
        <title>Palacký University, Olomouc (2008)</title>
        <p>[16] Ignatov, D., Ignatov, A.: Decision stream: Cultivating deep decision trees. In: IEEE 29th</p>
      </sec>
      <sec id="sec-7-6">
        <title>International Conference on Tools with Artificial Intelligence 2017, pp. 905–912. IEEE (2017)</title>
        <p>[17] Ioannou, Y., Robertson, D., Zikic, D., Kontschieder, P., Shotton, J., Brown, M.,
Criminisi, A.: (2016). Decision forests, convolutional networks and the models in-between,
https://arxiv.org/pdf/1603.01250.pdf. Last accessed 27 March 2022.
[18] Krajči, S.: A generalized concept lattice, Logic J. IGPL 13, 543–550 (2005)
[19] Krause, T., Lumpe, L., Schmidt, S.: A link between pattern structures and random forests. In:</p>
      </sec>
      <sec id="sec-7-7">
        <title>Valverde-Albacete, F. J., Trnečka, M (eds.) Proceedings of the 15th International Conference</title>
        <p>on Concept Lattices and Their Applications, CLA 2020, pp. 131–143. CEUR-WS.org (2020)
[20] Kuznetsov, S.O.: Machine learning and formal concept analysis. In: Eklund, P. (ed.) ICFCA
2004: Concept Lattices, pp. 287–312. Springer, Berlin, Heidelberg (2004)
[21] Laptev, D., Buhmann, J. M.: Transformation-invariant convolutional jungles. In:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3043–3051.</p>
        <p>IEEE (2015)
[22] Oliver, J. J.: Decision graphs – an extension of decision trees. In: Proceedings of the</p>
      </sec>
      <sec id="sec-7-8">
        <title>Fourth International Workshop on Artificial Intelligence and Statistics, pp. 343–350. Fort</title>
      </sec>
      <sec id="sec-7-9">
        <title>Lauderdale, Florida (1993)</title>
        <p>[23] Peterson, A.H., Martinez, T.R.: Reducing decision trees ensemble size using parallel decision</p>
        <p>DAGs. Int. J. Artif. Intell. Tools 18(4), 613–620 (2009)
[24] Platt, J. C., Cristianini, N, Shawe-Taylor, J.: Large margin DAGs for multiclass classification.</p>
      </sec>
      <sec id="sec-7-10">
        <title>In: NIPS’99: Proceedings of the 12th International Conference on Neural Information</title>
      </sec>
      <sec id="sec-7-11">
        <title>Processing Systems, pp. 547–553. MIT Press, Cambridge (1999)</title>
        <p>[25] Pócs, J., Pócsová, J.: Basic theorem as representation of heterogeneous concept lattices,</p>
        <p>Front. Comput. Sci. 9(4), 636–642 (2015)
[26] Pohlen, T.: Decision Jungles, In: Seminar Report, Fakultät für Mathematik, Informatik und</p>
      </sec>
      <sec id="sec-7-12">
        <title>Naturwissenschaften Lehr – und Forschungsgebiet Informatik (2014)</title>
        <p>[27] Quinlan, J. R.: Induction of Decision Trees. Mach. Learn. 1(1), 81–106 (1986)
[28] Quinlan, J. R.: Unknown attribute values in induction. In: Proceedings of the Sixth
International Machine Learning Workshop, pp. 164–168. Morgan Kaufmann (1989)
[29] Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann (1992)
[30] Shotton, J., Sharp, T., Kohli, P., Nowozin, S., Winn, J., Criminisi, A.: Decision jungles:</p>
      </sec>
      <sec id="sec-7-13">
        <title>Compact and rich models for classification. In: NIPS’13 Proceedings of the 26th International</title>
      </sec>
      <sec id="sec-7-14">
        <title>Conference on Neural Information Processing Systems, pp. 234–242. Curran Associates Inc.</title>
        <p>(2013)</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Bělohlávek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Fuzzy concepts and conceptual structures: induced similarities</article-title>
          .
          <source>In: JCIS '98 proceedings, International Conference on Computer Science and Informatics</source>
          , pp.
          <fpage>179</fpage>
          -
          <lpage>182</lpage>
          .
          <article-title>Association for Intelligent Machinery (</article-title>
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Bělohlávek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Baets</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Outrata</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vychodil</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Inducing decision trees via concept lattices</article-title>
          ,
          <source>Int. J. Gen. Syst</source>
          .
          <volume>38</volume>
          (
          <issue>4</issue>
          ),
          <fpage>455</fpage>
          -
          <lpage>467</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Ben</given-names>
            <surname>Yahia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Jaoua</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>Discovering knowledge from fuzzy concept lattice</article-title>
          . In: Kandel,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Last</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Bunke</surname>
          </string-name>
          , H. (eds.)
          <source>Data Mining and Computational Intelligence</source>
          , pp.
          <fpage>169</fpage>
          -
          <lpage>190</lpage>
          . Physica-Verlag (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Friedman</surname>
            ,
            <given-names>J. H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Olshen</surname>
            ,
            <given-names>R. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stone</surname>
            ,
            <given-names>C. J.</given-names>
          </string-name>
          :
          <article-title>Classification and regression trees</article-title>
          .
          <source>Chapman and Hall/CRC</source>
          (
          <year>1984</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Bagging predictors</article-title>
          .
          <source>Mach. Learn</source>
          .
          <volume>24</volume>
          (
          <issue>2</issue>
          ),
          <fpage>123</fpage>
          -
          <lpage>140</lpage>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Random Forests</article-title>
          .
          <source>Mach. Learn</source>
          .
          <volume>45</volume>
          ,
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Burusco</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fuentes-González</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>The study of  -fuzzy concept lattice</article-title>
          .
          <source>Mathw. Soft Comput</source>
          .
          <volume>3</volume>
          ,
          <fpage>209</fpage>
          -
          <lpage>218</lpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Busa-Fekete</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benbouzid</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kegl</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Fast classification using sparse decision DAGs</article-title>
          .
          <source>In: Proceedings of the 29th International Conference on Machine Learning</source>
          <year>2012</year>
          ,
          <volume>8</volume>
          pages.
          <source>Omnipress</source>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Cordero</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Enciso</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mora</surname>
          </string-name>
          , Á.,
          <string-name>
            <surname>Ojeda-Aciego</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rossi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>A Formal Concept Analysis Approach to Cooperative Conversational Recommendation</article-title>
          .
          <source>Int. J. Comput. Intell. Syst</source>
          .
          <volume>13</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1243</fpage>
          -
          <lpage>1252</lpage>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Dua</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graf</surname>
            ,
            <given-names>C.:</given-names>
          </string-name>
          <article-title>UCI Machine Learning Repository</article-title>
          . Irvine, CA: University of California, School of Information and Computer Science (
          <year>2019</year>
          ), http://archive.ics.uci.
          <source>edu/ml. Last accessed 27 March 2022</source>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Dubois</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Medina</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prade</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramírez-Poussa</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Disjunctive attribute dependencies in formal concept analysis under the epistemic view of formal contexts</article-title>
          .
          <source>Inf. Sci</source>
          .
          <volume>561</volume>
          ,
          <fpage>31</fpage>
          -
          <lpage>51</lpage>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Dudyrev</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S. O.</given-names>
          </string-name>
          :
          <article-title>Decision Concept Lattice vs. Decision Trees and Random Forests</article-title>
          . In: Braud,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Buzmakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Hanika</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Le Ber</surname>
          </string-name>
          ,
          <string-name>
            <surname>F</surname>
          </string-name>
          . (eds.)
          <source>ICFCA</source>
          <year>2021</year>
          ,
          <article-title>LNCS (LNAI)</article-title>
          , vol.
          <volume>12733</volume>
          , pp.
          <fpage>252</fpage>
          -
          <lpage>260</lpage>
          . Springer, Heidelberg (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Ferré</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huchard</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaytoue</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kuznetsov</surname>
            ,
            <given-names>S. O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Napoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Formal concept analysis: from knowledge discovery to knowledge processing</article-title>
          .
          <source>In: A Guided Tour of Artificial Intelligence Research</source>
          , Vol. II:
          <source>AI Algorithms</source>
          , pp.
          <fpage>411</fpage>
          -
          <lpage>445</lpage>
          . Springer (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>