<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Contexts as a Lattice of Decision Trees for Machine Reading Comprehension</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Boris Galitsky</string-name>
          <email>bgalitsky@hotmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dmitry Ilvovsky</string-name>
          <email>dilvovsky@hse.ru</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Elizaveta Goncharova</string-name>
          <email>egoncharova@hse.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AIRI</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Higher School of Economics</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Knowledge Trail Inc</institution>
          ,
          <addr-line>San Jose, CA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <abstract>
        <p>Supported decision trees that have been first proposed to boost the performance and the explainability of the expert systems built upon the texts can become a great basis for the machine reading comprehension (MRC) systems. The supported decision tree is based on building and combining the corresponding discourse trees for the text passage. In this work, we build an environment of supported decision trees for the MRC task. Each answer is represented by a path of a supported decision tree and the whole corpus of answers is then form a lattice of supported decision trees. This environment gives a boost to MRC performance, handling cases where it is nontrivial to determine which document/passage MRC needs to be applied to.</p>
      </abstract>
      <kwd-group>
        <kwd>Comprehension</kwd>
        <kwd>decision tree</kwd>
        <kwd>Machine reading comprehension</kwd>
        <kwd>discourse structure</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Machine reading comprehension (MRC) is a question answering task where the goal of the
model is to read and understand text passages and answer the question about them. MRC is
designed to check the language model’s ability to understand text written in natural language,
thus, in some cases, an answer cannot be retrieved from the given text passage, or it requires
some world knowledge to answer a question. In such cases, a model should have access to
the external knowledge base to retrieve the correct answer. The most promising technique to
answer such questions is to augment the language model with the external knowledge database,
such as Wikipedia, and to ensemble it with the additional retrieval component that supports
the system with the relevant documents [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ].
      </p>
      <p>
        Thus, in the cases, when an answer cannot be retrieved directly from a text, the system is split
into two core components [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], where the former is an information retrieval system designed to
identify useful pieces of text from the knowledge source (the retriever); and a system to produce
the answer given the retrieved documents and the question (the reader).
      </p>
      <p>
        The existing models that perform this problem are based on the transformer architecture
and retrieve the relevant text passages based on the constructed latent representations. The
main drawback of these techniques is the lack of explainability and limitation of the external
documents that a model has access to (e.g., Wikipedia only). In this work, we propose to use
supported decision trees DecTSup, first presented in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] for expert systems as the basis for
the relevant passages retrieval. The system allows to enrich the current textual corpora with
external documents and use the organized rule-based structure in order to retrieve the relevant
text passages that contain an answer to the asked question.
      </p>
      <p>In comparison to our previous work, we show that the DecTSups can be built for the texts of
the arbitrary genre. For the experimental evaluation, we refer to the MRC problem in the medical
domain and show that utilizing DecTSup-based retrieval procedure improves the performance
of the QA model over the standard MRC pipelines by up to 6% for the F1-score.</p>
      <p>In a conventional MRC architecture, documents are not organized in any structure, and
once a passage deemed most relevant is retrieved, the other documents are ignored. However,
when humans answer questions, the answer is backed up by supporting documents so that
the answer can be explained. In the real world question answering, the corpus of available
documents serves the purpose of providing additional information and clarification on the
answer topic. In this work we attempt to reproduce this feature of systematized MRC and
organize the documents and passages into such structure as lattice. As multiple questions for a
set of documents are answered, the involved documents are embedded into this lattice to form
chains of explanation for obtained answer. We analyze the advantages of this lattice-based MRC
architecture for delivering precise and explainable answers in a systematic way.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <sec id="sec-2-1">
        <title>2.1. Supported Decision Trees</title>
        <p>
          DecTSups have been first presented in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] as the basis for the expert systems that should retrieve
some instructions from the textual data based on the input query. There, the authors claim
that a flow of potential recommendations can be easily retrieved from the textual data and
organized in the format of the decision tree (DecT) as for the standard numerical attributes. This
lfow of recommendations or instructions is extracted in the form of a discourse tree [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], where
the nodes of the DecT are the elementary discourse units (EDUs) obtained from the discourse
tree expressing some condition, and the edges are some type of the rhetorical relations that
connect EDUs corresponding to diferent nodes. The DecT constructed from textual data can be
enriched with the additional knowledge retrieved from the specific rhetorical relation and, thus,
to construct the supported decision tree (DecTSup). DecTSup can be easily integrated into the
expert system as a basis to perform dialogue management that improves information retrieval
procedure.
        </p>
        <p>
          To turn a DecT into a corresponding DecTSup, each edge should be labeled with the
information extracted from text for the given decision step:
1. The extracted entity;
2. The extracted phrase for the attribute for this entity;
3. The rhetorical relation;
4. The full nucleus and satellite EDUs.
3. Decision Chains and Discourse Structure
3.1. Rhetorical Structure Theory and Decision Chains Construction
A discourse tree (DT) is a basis for the DecT and DecTSup respectively. DT is a hierarchical
structure that describes the relations that hold between text units in a document. Several
theories have been proposed in the past to describe the discourse structure, among which RST
[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] is one of the most developed. During discourse parsing, one can segment a document into
non-overlapping text spans (contiguous units for clauses) called EDUs. Each of these EDUs can
be tagged as either a nucleus or a satellite, where nucleus nodes are more central and satellite
nodes more peripheral. Nucleus units consist of the main information the author expresses
in the text, and satellite units contain additional information supporting the one presented in
a nucleus. The EDU itself can be of diferent lengths, i.e., it can contain just one word or a
word sequence. The discourse units are organized in the hierarchy by rhetorical relations (e.g.,
Antithesis, Elaboration, List, etc.) that reflect the function of these EDUs. As a result, we can
represent the discourse structure of the text as a DT, where the relations at the top level cover
the relations at the bottom.
        </p>
        <p>
          To build the corresponding DecT, first, a decision chain should be retrieved from the DT. A
decision chain is defined as a sequence of EDUs with rhetorical relations between sequence elements
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Elements of a decision chain are connected with  ℎ 


between a premise and a
decision. It can be read as “If ⟨ ⟩
then make ⟨⟩
according to  ℎ 
_ 
”. In
a decision chain, each consecutive member starting from the second one is a ⟨⟩.
        </p>
        <p>An example of the decision chains and a fragment of DecTSup for a text passage presented
below is given in Figure 1.</p>
        <p>“Although there is no cure for type 2 diabetes, studies show it is possible for some people to
reverse it. Through diet changes and weight loss, you may be able to reach and hold normal blood
sugar levels without medication. This does not mean you are completely cured. Type 2 diabetes
is an ongoing disease. Even if you are in remission, which means you are not taking medication
and your blood sugar levels stay in a healthy range, there is always a chance, that symptoms will
return. But it is possible for some people to go years without trouble controlling their glucose and
the health concerns that come with diabetes.”
elaboration
explanation
contrast</p>
        <p>TEXT: Although there is no cure for type 2 diabetes,
attribution</p>
        <p>TEXT: studies show
enablement
evaluation</p>
        <p>TEXT: it is possible for some people</p>
        <p>TEXT: to reverse it.
condition</p>
        <p>TEXT: Through diet changes and weight loss,
manner-means</p>
        <p>TEXT: you may be able to reach and hold normal blood sugar levels</p>
        <p>TEXT: without medication.</p>
        <p>TEXT: This does not mean you are completely cured.
elaboration</p>
        <p>TEXT: Type 2 diabetes is an ongoing disease.
contrast (RightToLeft)
elaboration(RightToLeft)
same-unit
condition</p>
        <p>TEXT: Even if you are in remission,
joint</p>
        <p>TEXT: which means you are not taking medication</p>
        <p>TEXT: and your blood sugar levels stay in a healthy range ,</p>
        <p>TEXT: there is always a chance ,</p>
        <p>TEXT: that symptoms will return .
background</p>
        <p>TEXT: But it is possible for some people to go years
elaboration</p>
        <p>TEXT: without trouble
elaboration</p>
        <p>TEXT: controlling their glucose and the health concerns</p>
        <p>When a text is represented as a discourse tree, it is split into elementary discourse units
(EDUs), denoted by ‘TEXT’ tag. EDUs are organized hierarchically according to rhetorical
relations between them. Using the constructed DT, for an arbitrary rhetorical relation, we can
define some patterns defining, how EDUs are connected to each other. In particular, relation
of Elaboration, ⟨satellite⟩ elaborates (provides additional information) on ⟨nucleus⟩. Certain
rhetorical relations have an obvious interpretations in terms of what decision ⟨satellite⟩ can
be made by means of ⟨nucleus⟩. For example, Enablement(result) ⇒ possible to achieve result
⟨nucleus⟩ by the way of ⟨satellite⟨. Based on this logical connection identified by the specific
EDUs, we can retrieve possible decision chains for the read passage given bellow:
diet changes and weight loss ⇒ without medication ⇒− sugar(normal)
Remission ⇒ not taking medications sugar(normal) ⇒ chance symptoms(yes)
OR Remission ⇒ control(sugar(normal)) ⇒ symptoms(no)</p>
        <p>We denote sugar(normal) as a formal representation of target values. Formal representations
are shown in italic, and original text – in a regular font.</p>
        <p>chance, possibility are the modalities which do not change the configuration of a DecTSup
but control the probability of navigation of the given decision chain.</p>
        <p>
          Formally, a decision chain is defined as a sequence of EDUs with rhetorical relations between
sequence elements [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Each element is a whole original EDU or its representation as a logic
form that can be obtained as a result of a semantic parsing: it depends whether an entity from
this EDU occurs in an available ontology or not. We encourage the readers to refer to [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] for a
detailed explanation of how to obtain semantic-based representation of the EDU. For formalized
elements of decision chains, it is easier to establish a correspondence or synonymy between
entities to form a decision navigation graph.
        </p>
        <p>Elements of a decision chain are connected with ⇒ℎ _ between a premise and a
decision. It can be read as “If ⟨premise⟩ then make ⟨decision⟩ according to rhetorical_relation”.
In a decision chain, each consecutive member starting from the second one is a ⟨decision⟩. Each
previous member is a ⟨premise⟩.</p>
        <p>Figure 1 shows two sections of decision chains extracted from the two texts above. Arrows
connect the same (or corresponding) entities, (possibly, parameterized diferently) such as
 ( ( _)) →  ( )) .</p>
        <p>In the first formalized decision expression  ( ( _)), the outermost predicate is
_) that ranges over control subjects such as  ( _) with an anonymized variable
4. DecTSup Environment for Machine Reading Comprehension</p>
      </sec>
      <sec id="sec-2-2">
        <title>4.1. Basic Example</title>
        <p>MRC is a task to retrieve the correct answer span from the given textual context. However, the
context which is given to MRC system usually restricts its applicability in real-life applications.
As, in many cases, the given passage may not contain the necessary information. Eforts made
in multi-passage MRC research have somewhat broken the limitation of the given context, but
there is still a long way to go as how to nd the most relevant resources for MRC systems
efectively determines the performance of answer prediction. Moreover, the existing DL MRC
systems often fail to capture long-range dependencies existing in a text, thus, even if an answer
can be retrieved from a context, the system may not be able to find it. It calls for a deeper
combination of information retrieval and machine reading comprehension, which we do in this
work.</p>
        <p>Let us consider the example, where in order to answer a simple question, MRC model should
be aware of the additional information.</p>
        <p>Passage: There is an ice cube in a glass of water. When the ice cube melts, will the water
level have risen, fallen, or remained the same? We have an ice cube floating in the water. If it is
lfoating in equilibrium, then it will have to displace enough water to support its weight. When
the ice has melted, it turns into exactly the same volume as it displaced before. So the added
volume is the same, so the level of the water will not change.</p>
        <p>Question: Into what does ice melt?
Answer: The same volume as it displaced before.</p>
        <p>Relying on this text, the transformer-based MRC system cannot answer a very basic question
about melting ice into water. Required knowledge is not spelled out in explanation but instead
is assumed to be known to the reader. As MRC has no means to acquire it, this knowledge
should be added in some way or another. A complete hypothetical decision tree for inferring a
solution to a physics problem contains hints on which texts and/or ontology expressions are
needed to answer all question about the physical system described in the formulated problem.</p>
        <p>We postulate a hypothesis that answering a question, given a passage constitutes navigating
a fragment of a decision tree build from this passage. While this approach seems natural when
this passage describes some form of a decision explicitly, and certain abstraction is required for
an arbitrary text genre where decisions are hypothetical.</p>
      </sec>
      <sec id="sec-2-3">
        <title>4.2. DecTSup MRC Architecture</title>
        <p>We propose to treat any text as some kind of problem formulation so that a respective decision
tree would tell the MRC system which knowledge is necessary to have complete domain
coverage. Usually, a text describes just a single path in a decision tree. Mining for other texts
helps reconstruct texts for other paths. Otherwise, it is unclear if even basic questions can
be answered (see 4.1). Hence forming a DecTSup for a given passage assure better domain
coverage so that the user can expect any relevant question to be answered reasonably well.</p>
        <p>If a question requires building a logical chain of facts or forming a decision, the construction
of a decision tree in the course of question-answering seems natural. The assumption we make
in this study is that a logical chain of facts and respective decision structure is associated with
any text, of an arbitrary genre. Considering a given text passage within a decision structure is
necessary to determine which other passages need to be involved. This approach is expected
to be much more robust and precise than a traditional information retrieval (IR) based, where
candidate passages are determined based on keywords.</p>
        <p>Under regular MRC architecture, an IR component first finds a document and a passage, and
then MRC finds an exact answer in this passage. In the proposed MRC architecture, the IR
system finds candidate documents along with DecTSup built for these documents, if available.
Then the DecTSup-MRC components use the lattice of DecTSups to decide on which passages
need to be involved in the answer.</p>
        <p>We draw the feature-by-feature comparison of three MRC environments: default,
retrievalaugmented [9] and DecTSup-based in Table 1.</p>
        <p>In Figure 2, we present the architecture for the standard MRC pipeline and MRC with the
DecTSup.
4.3. An Algorithm for Building DecTSup Environment
The essence of building DecTSup environment for MRC is organizing mapping between set of
passages  and DecTSups. This algorithm is given a set of documents/passages   in a corpus
and a set of external documents/passages   , and also given a sequence of queries   against   ,
builds a set of supported decision trees DecTSups.</p>
        <p>An answer  is obtained as a function MRC(⟨, ⟩) . A true answer   may need another
passage, not necessarily  , including other passages  ′ external document ℎ belong to   . The
whole passage  corresponds to a DecTSup( ) and an answer  – to a path of this DecTSup( ) ←
ℎ  () () . There is a many-to-many mapping between multiple passages and DecTSups.</p>
        <p>An algorithm for building this DecTSup is as follows.</p>
        <p>Initially, DecTSup = ∅. Given a corpus of passages   , iterate through all available queries  
and build a set ∪∈  () .</p>
        <p>For each ⟨, ?⟩
1. Compute  =  (⟨, ⟩) .
2. Compute   checking if passage  is suitable. If not find another  ′ or a suitable external
doc  = ℎ , using the lattice  . Decide whether to substitute or to augment  .
3. Build a chain of phrases from   for DecTSup( ).
4. Build ℎ  () from this chain of phrases. Only ℎ  () is currently available,
not the whole tree.
5. Extend ℎ  () towards a DecTSup by turning all phrases into negations.
6. Update the DecTSups system, identifying a location for ℎ  () :
a. Find existing path ℎ  () for ⟨, ⟩ and verify ℎ  () covers ℎ  () .
b. If not, find current DecTSup(  ) and augment it with ℎ  () . Check for overlapping
nodes and do the path merge if appropriate;
c. If no such DecTSup( ) exists, form a fragment of new DecTSup( ) from ℎ  () .</p>
        <p>Once all available queries  ∈   are ran, organize a set of DecTSup( ) into a lattice (see 4.4).
When a new batch of queries arrive, re-apply this algorithm and then recompute the lattice.</p>
        <p>The procedure of extension of ℎ  () towards a   is implemented by turning
all phrases into negations works as follows. For each phrase associated with a path node, we
consider its negation and branch it of this node. If a phrase in ℎ  () is already a
negation, remove this negation.</p>
        <p>This algorithm is also naturally applied when  is a book split into passages, the result can
be viewed as a book hyperlink graph.</p>
        <p>During MRC environment construction, each query produces an answer that maps into a
path in the DecTSup- based MRC environment. Initially, most queries form new paths which
are then merged into a part of the future DecTSups. Once new batches of queries does not
construct new DecTSups, we form a lattice from them.</p>
        <p>In Figure 3, we show the end of the learning session for building the lattice of DecTSups.
Most individual DecTSups are now built and new queries do not initiate additions of new paths
to DecTSups. Most of the DecTSups are completed and now form a lattice (shown by dotted
lines). Black nodes of DecTSup show nodes that have been formed by the time queries q1001, ...
are launched.</p>
        <p>Parts of the DecTSups corresponding to the given queries are shown in color. For example,
query q1001 is connected by the blue arrow with its answer and then with its blue path in
the top-left DecTSup. This blue path has just been added. On the contrary, query q1002 is
mapped into the existing part of the DecTSup (shown in dotted red). Query q1001 requires a
new external passage h1004.</p>
      </sec>
      <sec id="sec-2-4">
        <title>4.4. Decision Trees and Concept Lattices</title>
        <p>That is,  ′ is the set of all attributes (text entities) from Y shared by all objects (DecTSups)
= { ∈  |</p>
        <p>for each  ∈  ∶ ℎ,   ∈  }
= { ∈  |
for each  ∈  ∶ ℎ,   ∈  }
′
,
The following phase is responsible for retrieving the relevant text passages from the obtained
DecTSupp of the documents. In this work, we propose to form a concept lattice of all DecTSups
for a corpus of documents to enable this corpus with a structure assuring efective MRC. For a
text and its DecTSup, we extract entities and their attributes used for decisions and form the
formal context.</p>
        <p>We use the essential ideas of Lindig [10] algorithm which eficiently generates formal concepts
together with their subconcept-superconcept hierarchy in the form of concept lattice. The
algorithm builds the concept lattice by iteratively generating the neighbor concepts of a concept
, either top-down the lattice by adding new attributes to concept intents or bottom-up by
adding new objects to concept extents.</p>
        <p>For each  ⊆ 
and  ⊆ 
we denote by  a subset of  and by  a subset of  defined by
′
 = ( ∪ {})
by two.
concept  .
from A (and similarly for  ′).</p>
        <p>A formal concept consists of a set A (so-called extent) of objects which fall under the concept
and a set B (so-called intent) of attributes that fall under the concept such that A is the set of all
objects sharing all attributes from B and, conversely, B is the collection of all attributes from Y
shared by all objects from A.</p>
        <p>The algorithm is based on the fact that a concept ⟨, ⟩
is a neighbor of a given concept
if  is generated by  ∪ { }
, i.e.  = ( ∪ { })
″, where  ∈  \
is an attribute such that for
all attributes  ∈ 
⟨, ⟩ = {⟨, ⟩| = ( ∪ { })
−-  it holds that  ∪ {}
generates the same concept ⟨, ⟩
, i.e. neighbors of
″
,  ∈  \</p>
        <p>″
such that ( ∪ {})
=  for all  ∈ }</p>
        <p>Then a selection of a tree of concepts from the part of the concept lattice occurs. First, for
each concept  = ⟨, ⟩</p>
        <p>the number   of all of its lower concepts is computed. Each lower
concept is counted for each diferent attribute added to the concept  ,   . For instance, if a
concept  = ⟨, ⟩</p>
        <p>is generated from concept  by adding either attribute  or attribute  (i.e.
″ or  = ( ∪ { })</p>
        <p>″, respectively), the concept  is counted twice and   is increased</p>
        <p>Then a tree of concepts is chosen from the part of the concept lattice by iteratively going
from the greatest concept (generated by no attributes or, equivalently, by all objects) to minimal
concepts. The selection is based on the number   of lower concepts of the currently considered</p>
      </sec>
      <sec id="sec-2-5">
        <title>4.5. Online Selection of Passages for a Query</title>
        <p>At search time, to select the passage, we match the query with a chain of phrases for a path in
a DecTSup. We try to find such DecTSup so that as many chain phrases path(i) match query
phrases as possible.</p>
        <p>∶
∑  ( ∧ ℎ()

 
) → 
∧ is a syntactic generalization operator [11].</p>
        <p>Once a single DecTSup is identified, we select the passages   associated with matched nodes
of DecTSup.
maintained:</p>
        <p>In some cases, the best match between the query and path phrases can be distributed through
multiple DecTSups. Then the condition of connectedness in the lattice  of DecTSups must be
 () ∶
∑  ( ∧ ℎ() ()) → 
,
&amp; ( ()
∈  ()) ∈ .</p>
        <p>&amp;</p>
        <p>To summarize the preference for the occurrence of an answer in a corpus of documents, we
prefer it to be in a single passage. If it is not possible, we extend this passage towards other
passages connected vis a DecTSup. If it is still insuficient for answer identification, we further
extend passages towards foreign DecTSups linked in a lattice. Finally, in the worst case scenario,
passages come from unrelated documents associated with DecTSups anywhere in the lattice.</p>
        <p>Linked lattice nodes correspond to passages containing the same entities plus minus one or
two, therefore, closely related. This is maintained by how the lattice is defined on the set of all
DecTSups.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>5. Evaluation</title>
      <p>We check the MRC model performance on the collection of medical instruction collected from
the WebMD website and syntactically generate the question that could be asked regarding the
retrieved passages. We select ten classes of diseases to diversify the experiments, and track
each processing step to identify the performance bottleneck. It should be mentioned that the
F1-score reported to assess the model’s performance is applied for each topic independently
assessing whether a retrieved answer is correct or not.</p>
      <p>In the baseline, we rely on IR to identify passages in the fixed corpus of documents, so the
main source of errors is an improperly selected passage or a lack of appropriate passage in the
corpus. Our second baseline in the third column is syntactic generalization between query 
and passage  which is better than keyword frequency maintained by IR but lags behind the
approach developed in this paper. In the fourth column, we show F1 of DecTSup-enabled MRC
environment. The fith column shows the contribution of ideal, corrected DecTSups and the
last, sixth column should the contribution of the lattice of DecTSups, where individual trees are
organized to systematically identify additional passages which need to be involved in finding
the exact answer.</p>
      <p>The second baseline is only 0.6% better then the IR-based document identification, which
makes the syntactic generalization insuficient for the robust identification of relevant passages
to form the exact comprehensive answer. We observe that an unstructured MRC can have a
boost of 4% F1-score by finding the best document by DecTSup environment. Proceeding from
IR passage selection to the one based on independent DecTSups turns out to be an ultimate win
for the MRC environment. A further upgrade from a set of DecTSups to a lattice gives further
2% boost in search performance.</p>
    </sec>
    <sec id="sec-4">
      <title>6. Conclusion</title>
      <p>In this work, we explored a way to build decision trees from text relying on discourse analysis and
use this environment to boost the MRC model’s performance. We use DecTSup environment
for retrieving the relevant documents and passages, where the answer can be found. We
compare the DecTSup-based retrieval procedures with several baselines including syntactic
generalization for matching questions and passages and keywords matching on the set of the
medical instructions texts. We show that utilizing the DecTSup as the description of texts and
combining them into the concept lattice improves the performance of MRC by up to 6% on
average. In future studies, we will consider building a concept lattice from a textual description
of data [11], instead of a decision tree for authors’ instructions on how to do things and make
decisions in the course of it.
[9] D. S. Sachan, S. Reddy, W. Hamilton, C. Dyer, D. Yogatama, End-to-end training of
multi-document reader and retriever for open-domain question answering, in: NeurIPS,
2021.
[10] C. Lindig, Fast concept analysis, in: Working with Conceptual Structures – Contributions
to ICCS 2000, Shaker Verlag, 2000, pp. 152–161.
[11] B. Galitsky, G. Dobrocsi, J. Rosa, S. Kuznetsov, Using generalization of syntactic parse
trees for taxonomy capture on the web, volume 6828, 2011, pp. 104–117. doi:10.1007/
978-3-642-22688-5_8.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Dhingra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaheer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Mazaitis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Salakhutdinov</surname>
          </string-name>
          , W. Cohen,
          <article-title>Open domain question answering using early fusion of knowledge bases and text</article-title>
          ,
          <source>in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Brussels, Belgium,
          <year>2018</year>
          , pp.
          <fpage>4231</fpage>
          -
          <lpage>4242</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D18</fpage>
          - 1455.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Liu</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>W.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <article-title>Wang, Multi-passage machine reading comprehension with cross-passage answer verification, in: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics</article-title>
          , Melbourne, Australia,
          <year>2018</year>
          , pp.
          <fpage>1918</fpage>
          -
          <lpage>1927</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P18</fpage>
          - 1178.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <article-title>Reading Wikipedia to answer open-domain questions</article-title>
          ,
          <source>in: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Vancouver, Canada,
          <year>2017</year>
          , pp.
          <fpage>1870</fpage>
          -
          <lpage>1879</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P17</fpage>
          - 1171.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Galitsky</surname>
          </string-name>
          ,
          <article-title>Chapter 3 - Obtaining supported decision trees from text for health system applications</article-title>
          , Academic Press,
          <year>2022</year>
          . doi:https://doi.org/10.1016/B978- 0
          <source>- 12- 824521- 7</source>
          .
          <fpage>00013</fpage>
          -
          <lpage>2</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Galitsky</surname>
          </string-name>
          ,
          <article-title>Matching parse thickets for open domain question answering</article-title>
          ,
          <source>Data Knowledge Engineering</source>
          <volume>107</volume>
          (
          <year>2017</year>
          )
          <fpage>24</fpage>
          -
          <lpage>50</lpage>
          . doi:https://doi.org/10.1016/j.datak.
          <year>2016</year>
          .
          <volume>11</volume>
          .002.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Thompson</surname>
          </string-name>
          ,
          <article-title>Rhetorical structure theory: Toward a functional theory of text organization</article-title>
          ,
          <source>Text</source>
          <volume>8</volume>
          (
          <year>1988</year>
          ). doi:
          <volume>10</volume>
          .1515/text.1.
          <year>1988</year>
          .
          <volume>8</volume>
          .3.243.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B.</given-names>
            <surname>Galitsky</surname>
          </string-name>
          ,
          <source>Managing Customer Relations in an Explainable Way</source>
          , Springer International Publishing, Cham,
          <year>2020</year>
          , pp.
          <fpage>309</fpage>
          -
          <lpage>377</lpage>
          . doi:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>030</fpage>
          - 52167-
          <issue>7</issue>
          _
          <fpage>8</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Galitsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ilvovsky</surname>
          </string-name>
          ,
          <article-title>Validating correctness of textual explanation with complete discourse trees</article-title>
          ,
          <source>in: FCA4AI@IJCAI</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>