<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Generic Neural Exhaustive Approach for Entity Recognition and Sensitive Span Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mohammad Golam Sohrab</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pham Minh Thang</string-name>
          <email>pham.thangg@aist.go.jp</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Makoto Miwa</string-name>
          <email>makoto-miwa@toyota-ti.ac.jp</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Arti cial Intelligence Research Center, National Institute of Advanced Industrial Science and Technlogy</institution>
          ,
          <addr-line>2-4-7 Aomi, Koto-ku, Tokyo, 135-0064</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Toyota Technological Institute</institution>
          ,
          <addr-line>2-12-1 Hisakata Tempaku-ku Nagoya</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>735</fpage>
      <lpage>743</lpage>
      <abstract>
        <p>In this work, we present a deep exhaustive framework for the MEDDOCAN shared task. The framework employs a generic named entity recognition (NER) model that captures the underlying semantic information of texts. The key idea of our model is to enumerate all possible spans as potential entity mentions and classify them with deep neural networks. We introduce di erent sets of learning algorithms, including base representation(BR) average (BR-Avg), BR with attention mechanigm (BR-Attn), LSTM-Minus-based average (LM-Avg), LSTMMinus-based attention (LM-Attn), where with or without context is used after LSTM layer (Context or None) and an ensemble approach using maximum voting of all the approaches. We evaluate our exhaustive model on two sub-tasks in the MEDDOCAN shared task in medical domain using the o cial evaluation script. Among the ve submitted runs, the best run for each sub-task achieved the F-score of 93.12% on Sub-task 1 and the F-scores of 93.52% (strict) and 94.92% (merged) on Sub-task 2 without any external knowledge resources.</p>
      </abstract>
      <kwd-group>
        <kwd>Deep learning</kwd>
        <kwd>NER</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The MEDDOCAN shared task [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is an open challenge medical entity detection
task that allows participants to use any methodology and knowledge sources for
the clinical records with protected health information (PHI). The task allows
the comparison of the participating systems using the same benchmark
dataset and evaluation method. Named entity recognition has drawn considerable
attentions as the rst step towards many natural language processing (NLP)
applications including relation extraction [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], event extraction [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], co-reference
resolution [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and entity linking [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Recently, deep neural networks have shown
impressive performance on at named entity recognition in several domains [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
Such models achieved the state-of-the-art results without requiring any
handcrafted features or external knowledge resources.
      </p>
      <p>In this paper, we present a novel neural exhaustive model that detects at
and nested entities. The model reasons over all the regions within a speci ed
maximum size. The model rst represents each region as the combination of
the boundary and inside representations by using the outputs of bidirectional
long short-term memory (LSTM). The inside representation simply treats all the
tokens in a region equally by taking the average of LSTM outputs corresponding
to tokens inside the region. It then classi es the regions into their entity types or
non-entity. Unlike the existing models that rely on token-level labels, our model
directly employs an entity type as the label of a region.</p>
      <p>We evaluated our model on the MEDDOCAN task in clinical domain, which
aims at named entity recognition (NER), which is o cially called NER o set
and entity type classi cation, and sensitive span detection. The best run for
each sub-task achieved the F-score of 93.12% on sub-task 1 and the F-scores of
93.52% (strict) and 94.92% (merged) on sub-task 2.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Works</title>
      <p>
        Sohrab et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] detected the inner and outermost entities using exhaustive
approach and outperformed the state-of-the-art results by achieving 77.1% in
terms of F-score. Zhou et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] detected nested entities in a bottom-up way.
They detected the innermost at entities and then found other NEs containing
the at entities as sub-strings using rules derived from the detected entities. The
authors reported an improvement of around 3% in the F-score under certain
conditions on the GENIA corpus [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Recent studies show that the conditional
random elds (CRFs) can produce signi cantly higher tagging accuracy in at
or nested (stacking at NER to nested representation) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] NERs. Ju et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
proposed a novel neural model to address nested entities by dynamically stacking
at NER layers until no outer entities are extracted. A cascaded CRF layer is
used after the LSTM output in each at layer. The authors reported that the
model outperforms state-of-the-art results by achieving 74.5% in F-score.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Neural Exhaustive Approach</title>
      <p>We solve the NER and sensitive span detection (SSD) tasks using a neural
exhaustive approach that exhaustively consider all possible regions in a sentence
using a single neural network. The model detects nested entities by enumerating
all possible spans or regions. Our model is built upon a shared bidirectional
LSTM (Bi-LSTM) layer. Figure 1 shows the exhaustive model to solve the entity
recognition and SSD.
3.1</p>
      <sec id="sec-3-1">
        <title>Embedding Layer</title>
        <p>In the embedding layer, each word is represented by concatenating the pretrained
word embedding and character-based word representations where we encode the
character-level information of the word. The character-based word
representations are obtained by feeding the sequence of character embeddings comprising
a word to a Bi-LSTM layer and concatenate the forward and backward output
representations. The character embeddings in a word is randomly initialized.
Given an input sentence sequence X = fx1; x2; :::xng, where xi denotes the
i-th word and n denotes the number of words in the sentence sequence, the
distributed embeddings of words, which are introduced in the last section, are
fed into a bidirectional LSTM (Bi-LSTM) layer. The Bi-LSTM layer computes
the hidden vector sequence in forward !h = nh!1; h!2; : : : ; h!no and backward
h = nh1; h2; : : : ; hno manners. We concatenate the forward and backward
outputs as hi = h!hi; hii, where [; ] denotes concatenation.
3.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Exhaustive Layer</title>
        <p>The exhaustive layer enumerates all possible regions by exhaustive combination.
We generate all possible regions with the sizes less than or equal to the maximum
region size L, which is prede ned. We use (i; k) to represent the region from i
to k inclusive, where 1 i &lt; k n and k i &lt; L.</p>
        <p>We represent each region using the outputs of the shared underlying LSTM
layer. We represent the region with two separate representations: the boundary
representation for region detection and the inside representation for semantic
type classi cation. In the latter part of this section, we rst introduce the base
region representations and then explain two enhancements.</p>
        <p>Base Region Representations The boundary representation is prepared to
capture the both ends of the region. We rely on the outputs of the bidirectional
LSTM layer corresponding to the boundary words of a target region for this
purpose. We obtain the left- and right-boundary representations R(i; k)[L;R] of
the region (i; k) as follows:</p>
        <p>R(i; k)[L;R] = [hi; hk] :
(1)</p>
        <p>The inside representation is prepared to capture its semantic type by
encoding the whole semantic information of the region. In the base representation, we
average the outputs of the Bi-LSTM layer in the region to treat them equally.</p>
        <p>Using the boundary and inside representations, we obtain the left-, inside
with average representation, and right-boundary R(i; k)[L;A;R] of the region (i; k)
as follows:</p>
      </sec>
      <sec id="sec-3-3">
        <title>Region Representations using Attention Mechanism Instead of relying</title>
        <p>
          only on the average of the outputs of Bi-LSTM layer, we also try an attention
mechanism [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] over words in each region for the task of notion of headness.
Speci cally, we extend the inside representations using attention mechanism as
follows:
1
i + 1
k
X hj ; hk5 :
j=i
3
(2)
(3)
(4)
(5)
(6)
(7)
(8)
1
i + 1
k
X hj ; hi
j=i
        </p>
        <p>3
hk+15 :
t = w F F N N</p>
        <p>!xt ;
exp( t)
Pend(i)
k=start(i) exp( k)</p>
        <p>;
end(i)</p>
        <p>X
k=start(i)
i;t !xt;
where !xt is the concatenated output of the Bi-LSTM layer over a region. xiis
a weighted sum of word vectors in region (i; k). Instead of Eq. 2, we obtain left-,
inside with attention-based representation, and right-boundary R(i; k)[L;A;R] of
the region (i; k) as follows:</p>
        <p>R(i; k)[L;A;R] = [hi; xi; hk] :</p>
      </sec>
      <sec id="sec-3-4">
        <title>Region Representations using LSTM-Minus We also employ LSTM-Minus</title>
        <p>
          [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] for the boundary representation. The left-boundary computed as the
representation of the previous word of the region subtracted from the representation
of the last word of the current region. Similarly, the right-boundary computed
as the representation of the next word of the region subtracted from the
representation of the rst word of the current region. We obtain the representation
R(i; k)[L;R] of the region (i; k) by concatenating the left- and right-boundary
based on LSTM-Minus and it is computed as follows:
        </p>
        <p>R(i; k)[L;R] = [hk
hi 1; hi
The above region or span information is concatenated with average embeddings
of the region (i; k) to produce the LSTM-Minus-based representations as:
With the LSTM output hi, we introduce a context level representation from
bidirectional LSTM layer. The idea of this approach is to capture the surrounding
LSTM output of a target region (i; k) by concatenating vector output of
previous hi 1, current hi, and next index hk+1 of LSTM output. With contextual
region representations, we can further generate new representation from Eqs.
1-9. Figure 3 shows an architecture of contextual level integration. We then feed
the representation of each segmented region to a recti ed linear unit (ReLU) as
an activation function. Finally, the output of the activation layer is passed to a
softmax output layer to classify the region into a speci c entity type.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental Settings</title>
      <sec id="sec-4-1">
        <title>Evaluation Settings</title>
        <p>
          We evaluated our exhaustive model on MEDDOCAN3 dataset to provide
empirical evidence for the e ectiveness of the exhaustive model both in NER and SSD.
Our model is implemented in Chainer4 deep learning framework. We generated
task speci c word embeddings by merging the raw text of training, development,
and test (including background set) sets, which included 200-dimensional
embeddings of 77,559 vocabulary. We used Adam [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] for learning with a mini-batch
size of 10. We used the same hyper-parameters in all the experiments; we set
the dimension of word embedding to 200, the dimension of character embedding
to 25, the hidden layer size to 200, the gradient clipping to 5, and the Adam
hyper-parameters to its default values [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. We employed the o cial
MEDDOCAN evaluation script5 to evaluate our system performances for both tasks.
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Data Pre-processing</title>
        <p>We read text directly from input text les. We learn and detect spans using the
neural exhaustive approach from Bi-LSTM layer, creating all possible
combination from beginning to end of a given sequence. Unlike the traditional NER
models, our model is independent from traditional 'BIO' tagging scheme, where
'B', 'I', and 'O' are stands for 'Begin', 'Inside', and 'Outside' of named entities,
respectively. Thus, each text and annotation les are processed by several simple
rules only for tokenization. After tokenization, each text with mapping
annotation les are passed to deep neural approach for mention detection, classi cation,
and sensitive token detection. Note that the o sets are restored to the original
o sets in evaluation.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Results and Discussions</title>
      <p>In order to evaluate the performance of NER and sensitive token detection,
we conduct experiments on di erent sets of learning algorithms, including base
representation(BR) average (BR-Avg), BR attention (BR-Attn),
LSTM-Minusbased average (LM-Avg), LSTM-Minus-based attention (LM-Attn), where with
3 http://temu.bsc.es/meddocan/index.php/data/
4 https://chainer.org/
5
https://github.com/PlanTL-SANIDAD/MEDDOCAN-CODALAB-EvaluationScript
or without context is used after LSTM layer (Context or None). Table 1 shows
the ve submitted results on NER in terms of F-score on the test sets. In
strict matching, it is shown that ensemble approach using maximum voting of
all the approaches, including BR-avg-None, BR-Attn-None, BR-Avg-Context,
BR-Attn-Context, LM-avg-None, LM-Attn-None, LM-Avg-Context,
LM-AttnContext for NER and sensitive token detection is very e ective to improve the
system performance. In contrast, the BR-Avg-None shows the best performance
on NER in terms of F-score when using merged matching. Table 2 shows the
categorical performances on the MEDDOCAN dataset.</p>
      <p>We show the di erences in performance on the development data set to
compare the possible scenarios of the given solutions and to report the best system
submissions for NER and SSD. Table 3 shows the performances of di erent
approaches on the development set in Sub-task 1 and 2. Table 3 in Sub-task 1
shows that almost all the results in di erent approaches are close to each other
to solve the Sub-task 1. In contrast, Table 3 in Sub-task 2 shows that attention
and average with di erent boundary representations of a region are e ective
both in strict and merged evaluations to detect sensitive token.
This paper presented approaches of neural exhaustive and neural contextual
exhaustive models model that considers all possible regions exhaustively for named
entity recognition and sensitive token detection. The model obtains the
representation of each region using the outputs of the underlying shared LSTM layer,
and it represents the di erent regions by concatenating boundary and inside
representations of the region. Several enhancements, namely attention mechanism,
LSTM-Minus, context from base representations, and context from LSTM-Minus
are investigated for the representations. It then classi es the region into an entity
type or non-entity. The model does not depend on any external NLP tools. In
the experiment, we show that our model learns to detect at and nested entities
from the generated mention candidates of all possible regions. Among the ve
submitted runs, the best run for each subtask achieved the F-score of 93.12% on
Sub-task 1 and the F-scores of 93.52% (strict) and 94.92% (merged) on Sub-task
2 without any external knowledge resources.</p>
      <p>Acknowledgments We thank the anonymous reviewers for their valuable
comments. This paper is based on results obtained from a project commissioned by
the New Energy and Industrial Technology Development Organization (NEDO).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bahdanau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cho</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.:</given-names>
          </string-name>
          <article-title>Neural machine translation by jointly learning to align and translate</article-title>
          .
          <source>In: ICLR</source>
          <year>2015</year>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Collier</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Park</surname>
            ,
            <given-names>H.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ogata</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tateisi</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nobata</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ohta</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sekimizu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Imai</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ibushi</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsujii</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The GENIA Project: Corpus-based Knowledge Acquisition and Information Extraction from Genome Research Papers</article-title>
          .
          <source>In: Proceedings of EACL</source>
          . pp.
          <volume>171</volume>
          {
          <fpage>172</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qin</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>A Language-Independent Neural Network for Event Detection</article-title>
          .
          <source>In: Proceedings of the 54th Annual Meeting of the ACL (Volume 2: Short Papers)</source>
          . pp.
          <volume>66</volume>
          {
          <fpage>71</fpage>
          . Berlin, Germany (
          <year>2016</year>
          ), http://anthology.aclweb.org/P16-2011.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Fragkou</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Applying named entity recognition and co-reference resolution for segmenting english texts</article-title>
          .
          <source>Progress in Arti cial Intelligence</source>
          <volume>6</volume>
          (
          <issue>4</issue>
          ),
          <volume>325</volume>
          {
          <fpage>346</fpage>
          (
          <year>2017</year>
          ), https://doi.org/10.1007/s13748-017-0127-3.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roth</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Entity Linking via Joint Encoding of Types, Descriptions, and Context</article-title>
          .
          <source>In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>2671</volume>
          {
          <fpage>2680</fpage>
          . ACL, Copenhagen, Denmark (
          <year>2017</year>
          ), https://www.aclweb.org/anthology/D17-1284.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ju</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miwa</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ananiadou</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A Neural Layered Model for Nested Named Entity Recognition</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long Papers). pp.
          <volume>1446</volume>
          {
          <fpage>1459</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          , New Orleans,
          <string-name>
            <surname>Louisiana</surname>
          </string-name>
          (
          <year>2018</year>
          ), http://www.aclweb.org/anthology/P16-1105
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Ba., J.:
          <article-title>Adam: A Method for Stochastic Optimization</article-title>
          . In: ICLR (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lample</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ballesteros</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kawakami</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Neural Architectures for Named Entity Recognition</article-title>
          .
          <source>In: Proceedings of the 2016 Conference of the North American Chapter of the ACL: Human Language Technologies. ACL</source>
          . vol.
          <volume>1</volume>
          , pp.
          <volume>260</volume>
          {
          <fpage>270</fpage>
          . ACL, San Diego, California (
          <year>2016</year>
          ), http://www.aclweb.org/anthology/N16-1030.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Marimon</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalez-Agirre</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Intxaurrondo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rodrguez</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez</surname>
            <given-names>Martin</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Villegas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Krallinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            :
            <surname>Automatic</surname>
          </string-name>
          de
          <article-title>-identi cation of medical texts in spanish: the meddocan track, corpus, guidelines, methods and evaluation of results</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ). vol.
          <source>TBA</source>
          , p.
          <source>TBA. CEUR Workshop Proceedings (CEUR-WS.org)</source>
          , Bilbao,
          <source>Spain (Sep</source>
          <year>2019</year>
          ), TBA
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Miwa</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bansal</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures</article-title>
          .
          <source>In: Proceedings of the 54th Annual Meeting of the ACL</source>
          . pp.
          <volume>1105</volume>
          {
          <fpage>1116</fpage>
          .
          <string-name>
            <surname>ACL</surname>
          </string-name>
          , Berlin, Germany (
          <year>2016</year>
          ), http://aclweb.org/anthology/N18-1131
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Sohrab</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miwa</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Deep exhaustive model for nested named entity recognition</article-title>
          .
          <source>In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>2843</volume>
          {
          <fpage>2849</fpage>
          . Association for Computational Linguistics, Brussels, Belgium (Oct-Nov
          <year>2018</year>
          ), https://www.aclweb.org/anthology/D18-1309
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Son</surname>
            ,
            <given-names>N.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Minh</surname>
            ,
            <given-names>N.L.</given-names>
          </string-name>
          :
          <article-title>Nested Named Entity Recognition Using Multilayer Recurrent Neural Networks</article-title>
          .
          <source>In: Proceedings of PACLING 2017</source>
          . pp.
          <volume>16</volume>
          {
          <fpage>18</fpage>
          .
          <string-name>
            <surname>Sedona</surname>
            <given-names>Hotel</given-names>
          </string-name>
          , Yangon, Myanmar (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Graph-based dependency parsing with bidirectional LSTM. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</article-title>
          . pp.
          <volume>2306</volume>
          {
          <fpage>2315</fpage>
          . Association for Computational Linguistics, Berlin, Germany (Aug
          <year>2016</year>
          ). https://doi.org/10.18653/v1/
          <fpage>P16</fpage>
          - 1218, https://www.aclweb.org/anthology/P16-1218
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , Zhang, J.,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Recognizing Names in Biomedical Texts: a Machine Learning Approach</article-title>
          .
          <source>Bioinformatics</source>
          <volume>20</volume>
          (
          <issue>7</issue>
          ),
          <volume>1178</volume>
          {
          <fpage>1190</fpage>
          (
          <year>2004</year>
          ), https://doi.org/10.1093/bioinformatics/bth060.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>