<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>UH-MAJA-KD at eHealth-KD Challenge 2019</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jorge Mederos Alvarado</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ernesto Quevedo Caballero</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alejandro Rodr guez Perez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roc o Cruz Linares</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Havana</institution>
          ,
          <country country="CU">Cuba</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>85</fpage>
      <lpage>94</lpage>
      <abstract>
        <p>This paper describes the solution presented by the UH-MAJAKD team in IberLEF eHealth-KD 2019: eHealth Knowledge Discovery challenge. Separate strategies were developed to solve substasks A and B, both based on deep learning models using domain-speci c word embeddings, and architectures using Bidirectional Long-Short Term Memory (BiLSTM) cells. In the case of Subtask A, Conditional Random Field was used to produce an output in BMEWO-V tag system to extract keyphrases. For Subtask B, two stacked BiLSTM layers are used along with Shortest Dependency Path in-between a pair of keyphrases to determine possible relationships between them.</p>
      </abstract>
      <kwd-group>
        <kwd>eHealth</kwd>
        <kwd>Knowledge discovery</kwd>
        <kwd>Keyphrase extraction</kwd>
        <kwd>Keyphrase classi cation</kwd>
        <kwd>Relationships extraction</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In the health domain, the large number of research and publications every year
makes nearly impossible for doctors and biomedical researchers to keep up to
date with the literature in their elds. Thus, nding ways to e ectively manage
the vast amounts of information and extract knowledge from it is really
important nowadays. This could help in the task of obtaining new and better scienti c
results or in the diagnosis of complex diseases. Due to all of these reasons, a high
interest around the scienti c community has aroused in developing systems to
automatically extract knowledge from medical texts.</p>
      <p>
        There is an increasing amount of e orts oriented towards this direction. One
of them is the IberLEF eHealth-KD 2019: eHealth Knowledge Discovery
challenge [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], in which context this paper was developed. The goal of this challenge
was the discovery of knowledge in medical texts, via the extraction and
classi cation of keyphrases, as well as the determination of semantic relationships
between pairs of keyphrases. The challenge was divided into two subtasks: A and
B, one for keyphrase extraction and classi cation, and the other oriented to the
extraction of semantic relationships.
      </p>
      <p>
        This paper describes the solution presented by the UH-MAJA-KD team in
IberLEF eHealth-KD 2019: eHealth Knowledge Discovery challenge. It proposes
a strategy using a hybrid model that combines a Bidirectional Long Short
Memory (BiLSTM) layer with a Conditional Random Field (CRF) layer for Subtask
A. This model is inspired on the model presented by UCM team [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] in the past
edition of the challenge; in addition, domain-speci c word embeddings are used.
For Subtask B a multiclass classi er is proposed, taking as input a sequence of
features vectors of the tokens in the Shortest Dependency Path between pairs of
keyphrases.
      </p>
      <p>The rest of the paper is organized as follows. In section 2 is given a brief
overview of word embeddings, and the particular one used along the rest of the
paper. Sections 3 and 4 describe speci cally the approach to solve Subtasks A
and B respectively. Then, the results of the models proposed are presented in
section 5, and nally, brief conclusions and future work lines are presented in
section 6.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Word embeddings</title>
      <p>
        Word embeddings are a strategy to represent words as real numbers vectors on
a reduced-dimension space. It is desired for these vectors to have the property
of context similarity, this is, for words that appear commonly in the same
context, their respective vectors must be close in the embedding space, under some
distance measure. There are many methods to obtain such embeddings in
literature, most of them based on probabilistic models and/or neural networks. Among
most popular are found word2vec [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], fastText morphological representation [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
and GloVe (Global Vectors for Word Representations) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Regarding neural network-based word embeddings, the corpus used to train
them is crucial in its performance, precisely because the corpus determines the
words and contexts in which the words appear. Intuitively, domain-speci c
corpora should be better at showing contextual and semantic relations
regarding that speci c domain. Consequently, a corpus was built based on Spanish
Wikipedia 1, extracting medical content pages. The corpus size is of
approximately 27 million words, with essentially medical content. To capture
domainspeci c semantic and contextual information, a word embedding was trained
on this corpus. To do this, it was used the word2vec algorithm API o ered by
gensim [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] python library, using the architecture CBOW (Continuous Bag of
Words) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Embedding details are shown next:
. Embedding space dimensions: 300.
. Windows size: 5.
. Vocabulary size: approximately 500 thousand words.
      </p>
      <p>. Negative sample: 5</p>
      <sec id="sec-2-1">
        <title>1 es.wikipedia.org</title>
        <p>3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Subtask A</title>
      <p>The goal of Subtask A was to extract keyphrases from sentences and to classify
them as Concept, Action, Reference or Predicate. The proposed solution splits
this subtask into four more speci c ones, each of those to extract and classify
concepts, actions, references and predicates respectively. The de ned
architecture is the same in all the four cases, but each model is trained independently,
using as training examples only those of its corresponding task (e.g the model
that extracts and classi es keyphrases in Concept, only receives as input
annotations of Concept keyphrases). This is done in order to improve speci c weight
learning for each type of keyphrase since they could be under di erent
hypothesis functions, making di cult to the model learning 'good' weights for all of
them together. Moreover, to process them united could lead to more ambiguity
in the decoding process (which will be explained at 3.3), making more solutions
unfeasible. Finally, all the keyphrases detected by all the four models are put
together.
3.1</p>
      <sec id="sec-3-1">
        <title>Model Input</title>
        <p>The system receives as input a sentence string, thus it needs some preprocessing
to build an appropriated input to the models. The rst step is to tokenize the
sentences as all model inputs expect a sequence of tokens.</p>
        <p>For each token in which the sentence was split, the input for that token
consist of a list of three feature vectors:
. Character encodings: Concatenation of one-hot encoded vectors of the
characters contained in the word.
. PoS-tag vector: One hot encoded vector of Part of Speech (PoS)
information.
. Word indexes: One hot encoded index in the word embedding vocabulary.</p>
        <p>To obtain the rst standard ASCII alphabet was used. To extract PoS-tag
information the python library spacy2 was used. In the case of the third input,
some words are captured using regular expressions and substituted with special
tokens de ned in the word embedding vocabulary (e.g currencies, units of
measurement and other words with digits or non-latin characters). In the case of
words not appearing in the vocabulary, a special token 'unseen' was de ned.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Model Architecture</title>
        <p>Each of the four models used to solve the Subtask A receives a sequence of token
inputs as described in 3.1, and produces a same sized sequence with labels for
each token in the BMEWO-V tagging system which will be described in the
section below.</p>
        <p>The architecture is conformed by four main components:</p>
        <sec id="sec-3-2-1">
          <title>2 spacy.io</title>
          <p>
            . Word embedding matrix
. Char embedding BiLSTM [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ]
. Token-level BiLSTM
. CRF classi er [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ]
          </p>
          <p>It is pipelined as follows. For each token in the input sequence, the
pretrained word embedding layer produces an embedding vector using the word
index input. The character embedding layer receives the sequence of character
encodings contained in the word and produces a vector, capturing character level
information for each word. These two vectors are concatenated with the PoS-tag
vector information of the word, and all together serve as input to each time step
of the token-level BiLSTM layer. Finally, the outputs of the BiLSTM layer are
passed to a CRF layer.</p>
          <p>A summary of the model is shown in Figure 1.
The CRF layer produces a sequence of tags in the BMEWO-V tagging system.
This classi cation corresponds to B for begin of a keyphrase, M for medium, E
for end, W for tokens that are a keyphrase themselves and O for tokens that do
not represent anything. It also takes into account the possibility of keyphrases
overlapping, including the tag V in such cases. For the sentence: El cancer de
pulmon causa muerte prematura, the model detecting Concept keyphrases should
produce the output: O-V-M-E-O-B-E.</p>
          <p>Since the expected output in Subtask A is a sequence of keyphrases for each
sentence, a procedure is necessary to transform the BMEWO-V tag sequence
got from a given sentence, in a keyphrase sequence corresponding to the output
expected in Subtask A. This process was called decoding. There is an important
challenge in this process: tokens belonging to a keyphrase are not necessarily
continuous in the sentence. Taking this into account, the decoding process is
divided into two stages. First, discontinuous keyphrases are detected and then,
at a second moment, continuous keyphrases.</p>
          <p>In accordance to Spanish correct use, The set of tag sequences that must be
interpreted as a group of discontinuous keyphrases were reduced to those that
match the regular expressions (V+)((M*EO*)+)(M*E) and ((BO)+)(B)(V+).
The rst one corresponds to keyphrases that share their initial tokens, and the
second one to those that share their nal tokens. These two capture most of
the desired discontinuous keyphrases. Among the examples of the rst case it
is found the fragment cancer de pulmon y de mama, tagged as
V-M-E-O-ME, where keyphrases cancer de pulmon and cancer de mama are found. And,
as example of the latter, the fragment tejidos y organos humanos, tagged as
B-O-B-V, where keyphrases tejidos humanos and organos humanos are found.
When a match is detected and the keyphrases are extracted, all the tags in that
fragment are set to tag O.</p>
          <p>After the detection of possible discontinuous keyphrases, the second stage
starts assuming all the remaining keyphrases appear as continuous sequences of
tokens. To extract continuous keyphrases, an iterative process is carried on over
the tag sequence produced by the model. Due to limitations in the
BMEWOV system, the procedure also assumes that the maximum overlapping depth
is 2. Assuming otherwise only makes the process more ambiguous and does
not capture much more information since is not common in Spanish to nd
examples with deeper overlapping. Given this, along with the procedure, two
inconstruction keyphrases are maintained. In each iteration these two keyphrases
are created, extended or emitted in accordance to rules de ned considering only
the previous and the current tag. Tag B indicates to start a new keyphrase, M
the extension of an existent keyphrase and E its ending. Tag V introduces
overlapping, hence this is the one that causes that there could be two in-construction
keyphrases at a given moment. Tag W causes the current token to be reported
automatically as a keyphrase.
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Subtask B</title>
      <p>
        The goal of Subtask B was to detect semantic relationships between pairs of
keyphrases. The solution proposed consists of traversing every pair of keyphrases
and determine whether one of the de ned semantic relationships is established
between them or not, via a multiclass classi er. This is accomplished by building
a dependency tree for the tokens in the sentence and nding the shortest path
in-between the keyphrases along this tree. This is called Shortest Dependency
Path [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The model is agnostic to any restrictions de ned on the relations
domain (e.g it is not told in advance that for relation Subject, one of the keyphrases
should be an Action), needing to learn it by itself.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Model Input</title>
        <p>Similar to Subtask A models, this model expects a sequence of tokens. For each
token in that sequence, the input for that token consists of a list of four feature
vectors:
. Word indexes: One hot encoded index in the word embedding vocabulary.
. Syntactic dependency relation vector: One hot encoded vector of
syntactic dependency information.
. BMEWO-V tag encoding: One hot encoded BMEWO-V tag.
. Subtask A type of keyphrase encoding: One hot classi cation on
Concept, Action, Reference or Predicate of the keyphrase to which token belongs.</p>
        <p>The word indexes are obtained as described in 3.1. To extract syntactic
dependency information the python library spacy was used. The third and fourth
inputs are obtained from Subtask A if they were pipelined as in the case of
Scenario 1 in the challenge.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Model Architecture</title>
        <p>The architecture is conformed by three main components:
. Word embedding matrix
. Stacked BiLSTMs
. Two dense multiclass classi ers</p>
        <p>It is pipelined as follows. For each token in the input sequence, the pre-trained
word embedding layer produces an embedding vector using the word index input.
The embedding vector is then concatenated with the other three input vectors,
and all together serve as input to each time step of the stacked BiLSTM layers.
Finally, the last time step output of the stacked BiLSTM layers serves as input
of two Dense layers serving as multiclass classi ers, one for each direction in
which relationships could be established between the pair of keyphrases, since
those are not symmetric.</p>
        <p>
          A summary of the model is shown in Figure 2.
The evaluation in both subtasks was carried out using the annotated corpus
proposed in the challenge. The results were measured with precision, recall and
F1 in three scenarios as described in the details of IberLEF eHealth-KD 2019:
eHealth Knowledge Discovery [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>Tables 1, 2 and 3 show the results obtained by participants in Scenarios 1,2
and 3 respectively. Scenario 2 measures the results in Subtask A and Scenario 3
only in Subtask B, whereas Scenario 1 combines both Subtask A and B.</p>
        <p>As can be observed, the proposal for Subtask A had a competitive
performance, being only 0.0047 points lower than the rst place in F1 score. However,
results on Subtask B are not as promising. The rst place critically outperformed
the model proposed for Subtask B.</p>
        <p>In the case of Subtask A, the model showed faster convergence when training
on both Action and Reference labels. This is probably because of the syntactic
patterns they show, that are rapidly captured by the model.</p>
        <p>It is worth to mention the evaluations that were made on the BMEWO-V
decoder. It turned to be over 99% in both precision and recovery when
evaluated on perfectly annotated labels. It showed, however, a non-linear decline in
performance when evaluated on inaccurately-classi ed labels.</p>
        <p>The set of parameters and the hyper-parameters used to test the models are
the following:
. Words embeddings dimension: 300
. Characters embeddings dimension: 25
. BiLSTM dimension(Both char level and token level): 64
. BiLSTM dropout and recurrent dropout(Both char level and token level):
0.2
. Optimizer: adam
. Epochs: 30(Concept), 10(Action), 20(Predicate), 10(Reference)</p>
        <p>Subtask B model:
. Words embeddings dimension: 300
. First BiLSTM dimension: 64
. Recurrent dropout: 0.4
. Second BiLSTM dimension: 32
. Recurrent dropout: 0.2
. Optimizer: SGD(Nesterov, momentum = 0.9)
. Epochs: 20</p>
        <p>The number of epochs was selected empirically, based on the fast
convergence of the models, tending to quickly over t on training dataset, even though
validation data was used. The remaining parameters were selected as standard
for similar applications in literature.
6</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Work</title>
      <p>In this work were described the models presented by the UH-MAJA-KD team
for the IberLEF eHealth-KD 2019: eHealth Knowledge Discovery.</p>
      <p>In Subtask A a hybrid BiLSTM and CRF model with speci c domain
pretrained word embeddings was proposed. Our model obtained the third place in
the Scenario 2. In Subtask B a multiclass classi er using Shortest Dependency
Path with pre-trained word embeddings in a speci c domain was proposed. Our
model obtained the sixth place in the Scenario 3. Our team reached the sixth
position in the overall competition standing.</p>
      <p>The corpus in which the domain-speci c word embedding was trained is
relatively small. It is proposed as future work to build a more expressive and
abundant corpus to improve the word embedding performance. Also, could be
promising to try to concatenate both domain-speci c and general purpose word
embeddings, in order to gain one's speci city and the generalization capability
of the latter. To improve the capabilities of the system in the overall task, it
could be convenient to train the system (l.e both models) as a whole, providing
Subtask B with the output from Subtask A, needing the rst to deal with the
errors produced by the latter.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>We would like to acknowledge the joint project Tec-UH of Tecnomatica3
enterprise and the Arti cial Intelligence Group at the University of Havana, to allow
3 https://www.cupet.cu/footer/informatica-automatica-y-comunicaciones/
us to use high-performance computational equipment to develop and test our
ideas.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching word vectors with subword information</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>5</volume>
          ,
          <issue>135</issue>
          {
          <fpage>146</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Gulli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Deep Learning with Keras</article-title>
          .
          <source>Packt Publishing Ltd</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural computation 9(8)</source>
          ,
          <volume>1735</volume>
          {
          <fpage>1780</fpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>La</surname>
            <given-names>erty</given-names>
          </string-name>
          , J.,
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>C.N Pereira</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Conditional random elds: Probabilistic models for segmenting and labeling sequence data (</article-title>
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of sentences and documents</article-title>
          .
          <source>In: International conference on machine learning</source>
          . pp.
          <volume>1188</volume>
          {
          <issue>1196</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>A neural joint model for entity and relation extraction from biomedical text</article-title>
          .
          <source>BMC Bioinformatics</source>
          <volume>18</volume>
          (12
          <year>2017</year>
          ). https://doi.org/10.1186/s12859-017-1609-9
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <issue>1543</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Piad-Mor s</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gutierrez</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Consuegra-Ayala</surname>
            ,
            <given-names>J.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Estevez-Velarde</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almeida-Cruz</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , Mun~oz, R.,
          <string-name>
            <surname>Montoyo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of the ehealth knowledge discovery challenge at iberlef</article-title>
          <year>2019</year>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Rehurek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sojka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Software framework for topic modelling with large corpora</article-title>
          .
          <source>In: In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks. Citeseer</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Zavala</surname>
            ,
            <given-names>R.M.R.</given-names>
          </string-name>
          , Mart nez, P.,
          <string-name>
            <surname>Segura-Bedmar</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>A hybrid bi-lstm-crf model for knowledge recognition from ehealth documents</article-title>
          .
          <source>In: Proceedings of TASS 2018: Workshop on Semantic Analysis at SEPLN (TASS</source>
          <year>2018</year>
          )
          <article-title>(</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>