<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LSI2 UNED at eHealth-KD Challenge 2019</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alicia Lara-Clares</string-name>
          <email>alara@lsi.uned.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ana Garcia-Serrano</string-name>
          <email>agarcia@lsi.uned.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Nacional de Educacion a Distancia (UNED)</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>60</fpage>
      <lpage>66</lpage>
      <abstract>
        <p>In this work, we describe a Few-Shot Learning approach for Named Entity Recognition (NER) in eHealth documents to identify and classify key phrases in a document (subtask A in the IberLEF eHealthKD 2019 competition [10]). The architecture is an hybrid Bi-LSTM and CNN model with four input layers that can recognize multi-word entities using the BIO encoding format for the labels. The system obtained a F-score of 73.15% (baseline is 54,66%), with a 78,17% of precision, according to the eHealth-KD evaluation procedure. This improvement is reached mainly because (a) the correct selection of the hybrid model for NER that obtains better results using a POS tagger and (2) the addition of Wikidata entities to extend the vocabulary that improves the precision by nearly 10%.</p>
      </abstract>
      <kwd-group>
        <kwd>NER</kwd>
        <kwd>Knowledge Discovery</kwd>
        <kwd>Bi-LSTM</kwd>
        <kwd>CNN</kwd>
        <kwd>wikipedia2vec</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Currently, the number of medical data is growing at an exponential rate.
Literature in the medical domain, moreover, is often found as unstructured or
semistructured data. In these cases, it is necessary to nd methods to automatically
extract and categorize the data contained in them, using di erent techniques
as, for example, Named Entity Recognition (NER). NER aim is to recognize,
identify and categorize pieces of information that refers to di erent entities of
interest, i.e. a disease, a treatment or a patient name. First NER systems relied
heavily on heuristic, hand-crafted features and language-speci c knowledge as
in the work presented by Rau[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] to extract and recognize company names.
      </p>
      <p>
        In any research domain approximations based on the integration of di erent
approaches or the integration of external resources are commonly used in order
to improve the outcome of the research goal ([
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]). This is the case of neural
networks that are especially successful in complex NLP tasks [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], as for
example, G. Fabregat et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] work that use a deep learning model for disabilities and
diseases recognition using Convolutional Neural Networks (CNN) and Recurrent
Neural Networks (RNN). Also research work with word embedding based
techniques is frequently used, for example to simplify drug package lea ets written
in Spanish [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] or to de ne reproducible experiments and replication datasets
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        The aim of Few-shot Learning is to extract complex statistics and learn high
level features using a very small set of training data. This problem has been
addressed in several domains, such as [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] with one-shot learning, or [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] using
zero-shot learning. M. Hofer et al.[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] demonstrate the e ect of ve sequential
improvements on the learning capabilities of a neural network when having very
few annotated examples, using as baseline the state-of-the-art NER architecture
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        In this paper, we propose a hybrid Bi-LSTM CNN model following the work
presented at [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Speci cally, we have extended the model by adding a
Part-ofspeech (POS) tagging layer and information about multi-word entities.
Moreover, in this work, we use wikipedia2vec [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], a pre-trained word embedding
model from Wikipedia, and we extend the vocabulary by adding wikidata
entities such diseases, health problems, etc. The results obtained in the eHealth-KD
evaluation, improves the baseline by 18,5%.
      </p>
      <p>The rest of the paper is organized as follows. In section 2, we describe the
architecture of the system. Section 3 describes the evaluation process and results
obtained. Finally, section 4 outlines the conclusions and future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>System description</title>
      <p>
        The system process is divided into two steps. First, it is necessary to pre-process
the data and prepare it to be the input of the neural network and secondly after
to process the data using the implemented neural network it is needed a
postproccess of the output to be evaluated in the tasks of the IberLEF eHealth-KD
2019 competition [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In the next sub-sections both descriptions are included.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Pre and Post processing of the data</title>
        <p>
          All documents are pre-processed following the next steps. First, sentences are
splitted and tokenized using the Stanford CoreNLP natural language processing
toolkit [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], ignoring all non-alphanumeric symbols. Then, each token is annotated
using the BIO scheme, to preserve the multi-word entities. After that, we get
the POS tag of each token (using the Stanford Core-NLP POS tagger). After
the proccessing of the input data, the output data has to be converted into the
BRAT format [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. The BRAT format allows to include some aspects of the
data original le, because it store all the information together with the labels of
each category and the positions of the tokens in the text. Given this di erence
between data formats the nal step is to process the documents as shown in the
Table 1: concept, POS tags and BIO-label.
        </p>
        <p>Word POS tag BIO-label
No ADV O
existe VERB B-Action
un NUM O
tratamiento NOUN B-Concept
que CONJ O
restablezca NOUN B-Action
la DET O
funcion NOUN B-Concept
ovarica ADJ I-Concept
normal ADJ B-Concept</p>
        <p>Table 1. Structure of processed data in this work
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Network architecture</title>
        <p>The network architecture used in this work is shown in Figure 1. It has four
input layers, named as character level, word level, casing input and POS tag
level, described in the following:
{ The rst input layer corresponds to the character level. It starts with a
character embedding that maps a vocabulary of 120 possible characters to
an embedding initialized randomly. The maximum number of character per
word is 52. It has a dropout layer (with drop rate 0.5) used to avoid the risk
of over tting. Finally, it has a convolutional layer to process the 1-dimension
character layer.
{ The second input layer uses the wikipedia2vec pretrained embeddings in
Spanish language of 300 dimensions 3, mapping the existing vocabulary from
the dataset.
{ The third layer maps a vocabulary of eight casing types: numeric, allLower,
allUpper, mainly numeric, initialUpper, contains digit, padding and other.
{ The fourth layer maps into a one-hot embedding the POS tags existing in
the vocabulary.</p>
        <p>
          The architecture starts processing these four inputs independently, to nally
merge them into the last process. The bidirectional LSTM layer Bi-LSTM [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]
transforms the input data into two vectors of 200 dimensions. In the last step,
the softmax function is used to obtain a prediction for locating and classifying
sequences of words in the input text.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <p>The evaluation of the proposed model is carried out using the annotated
corpus delivered in the 2019 competition that was extracted from the available
MedlinePlus resources 4.</p>
      <p>The IberLEF eHealth-KD 2019 corpus is divided in three sections: training,
development and test. The training set contains a total of 600 sentences manually
annotated in Brat and post-processed to match the input format. The
development set has 100 annotated sentences, and the test data has 8800 non-annotated
sentences for competition purposes.
3. Predicate, used to represent a function or lter of another set of elements,
which has a semantic label in the text
4. Reference, a textual element that refers to a concept of the same sentence
or of di erent one, which can be indicated by textual clues.</p>
      <p>In this work, tokens are annotated with the previous categories using the
di erent labels (see Table 2) following the BIO encoding format.</p>
      <p>Then the scores are computed (correct, partial, missing, incorrect and
spurious matches). The expected and actual output les do not need to agree on the
ID for each phrase, nor on their order. The detailed information of the evaluation
is in the eHealth KD competition website 5.
3.1</p>
      <sec id="sec-3-1">
        <title>Results</title>
        <p>In this work has been carried out a series of experiments on the development
corpus delivered by eHealth-KD 2019. The most interesting results are brie y
described below, and they can be seen in Table 3.</p>
        <p>The experiments have been focused on the embeddings model used, and in
the impact of the POS tagging in the neural network results. We used four
embedding models Fastext 6, FastText and GloVe embeddings from SBWC 7
and wikipedia2vec 8. The rst experimental conclusions achieved are:
5 https://knowledge-learning.github.io/ehealthkd-2019/evaluation
6
https://github.com/facebookresearch/fastText/blob/master/docs/pretrainedvectors.md
7 https://github.com/dccuchile/spanish-word-embeddings
8 https://wikipedia2vec.github.io/wikipedia2vec/pretrained/
1. The use of wikipedia2vec improves the performance and maintains the results
from FasText in Spanish language.
2. Adding Wikidata entities improve the precision by approximate 10%.
3. POS tags do not improve results signi cantly in this task.
4. Adding fastext embeddings decreases system e ciency and does not improve
results over wikipedia2vec.
5. Other embeddings in Spanish language are worse in terms of e ciency and
accuracy.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and Future Work</title>
      <p>In this work, we propose a hybrid Bi-LSTM and CNN model with four input
layers that can recognize multi-word entities using the BIO encoding format
for the labels. The vocabulary is improved using Wikidata entities such as
diseases, health problems, treatments, etc. This entities are labeled as BIO-concepts
and added in the corpus data as sentences. Our system can achieve satisfactory
performance without requiring hand-crafted features. Our results demonstrated
that in Spanish language, the wikipedia2vec pretrained embedding vectors has
better performance in this task than other embeddings such as Fastext or Glove.</p>
      <p>
        We plan to experiment with other BIO-based formats to detect discontinuous,
overlapped or nested entities, such as BMEWO-V [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Moreover, we will extend
the annotation using domain-speci c formats and using external sources (such
as Wikipedia with cui2vec format [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]).
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>Funding: This work was supported by the UNED predoctoral grant started in
April 2019 (BICI N7, November 19th, 2018).</p>
      <p>The authors want to thank PhD Juan J. Lastra-Diaz for his support.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Beam</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kompa</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fried</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palmer</surname>
            ,
            <given-names>N.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cai</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohane</surname>
            ,
            <given-names>I.S.:</given-names>
          </string-name>
          <article-title>Clinical concept embeddings learned from massive sources of multimodal medical data</article-title>
          .
          <source>arXiv preprint arXiv:1804</source>
          .
          <volume>01486</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Benavent</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benavent</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>de Ves</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>Granados</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Serrano</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Experiences at imageclef 2010 using cbir and tbir mixing information approaches</article-title>
          .
          <source>In: CEUR Proceedings. 2010 CLEF September</source>
          , Padua, Italy. vol.
          <volume>1176</volume>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Castellanos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cigarran</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Serrano</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Formal concept analysis for topic detection: A clustering quality experimental analysis</article-title>
          .
          <source>Information Systems (66)</source>
          ,
          <volume>24</volume>
          {
          <fpage>42</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Chiu</surname>
            ,
            <given-names>J.P.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nichols</surname>
          </string-name>
          , E.:
          <article-title>Named entity recognition with bidirectional lstm-cnns</article-title>
          .
          <source>Transactions of the Association for Computational Linguistics</source>
          <volume>4</volume>
          ,
          <issue>357</issue>
          {
          <fpage>370</fpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Fabregat</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Araujo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martinez-Romo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Deep neural models for extracting entities and relationships in the new rdd corpus relating disabilities and rare diseases</article-title>
          .
          <source>Computer Methods and Programs in Biomedicine 164</source>
          ,
          <issue>121</issue>
          {
          <fpage>129</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Fei-Fei</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fergus</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perona</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>One-shot learning of object categories</article-title>
          .
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>28</volume>
          (
          <issue>4</issue>
          ),
          <volume>594</volume>
          {
          <fpage>611</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Hofer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kormilitzin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nevado-Holgado</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Few-shot learning for named entity recognition in medical text</article-title>
          . arXiv preprint arXiv:
          <year>1811</year>
          .
          <volume>05468</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lastra-Diaz</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Serrano</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Batet</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fernandez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chirigati</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Hesml: a scalable ontology-based semantic similarity measures library with a set of reproducible experiments and a replication dataset</article-title>
          .
          <source>Information Systems</source>
          <volume>66</volume>
          ,
          <fpage>97</fpage>
          {
          <fpage>118</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Surdeanu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bauer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finkel</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McClosky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The stanford corenlp natural language processing toolkit</article-title>
          . In:
          <article-title>Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations</article-title>
          . pp.
          <volume>55</volume>
          {
          <issue>60</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Piad-Mor s</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gutierrez</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Consuegra-Ayala</surname>
            ,
            <given-names>J.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Estevez-Velarde</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almeida-Cruz</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , Mun~oz, R.,
          <string-name>
            <surname>Montoyo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of the ehealth knowledge discovery challenge at iberlef 2019</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          , CEUR-WS.org (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Rau</surname>
            ,
            <given-names>L.F.</given-names>
          </string-name>
          :
          <article-title>Extracting company names from text</article-title>
          .
          <source>In: [1991] Proceedings. The Seventh IEEE Conference on Arti cial Intelligence Application</source>
          . vol.
          <volume>1</volume>
          , pp.
          <volume>29</volume>
          {
          <fpage>32</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Schuster</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paliwal</surname>
            ,
            <given-names>K.K.:</given-names>
          </string-name>
          <article-title>Bidirectional recurrent neural networks</article-title>
          .
          <source>IEEE Transactions on Signal Processing</source>
          <volume>45</volume>
          (
          <issue>11</issue>
          ),
          <volume>2673</volume>
          {
          <fpage>2681</fpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Segura-Bedmar</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martinez</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Simplifying drug package lea ets written in spanish by using word embedding</article-title>
          .
          <source>Journal of Biomedical Semantics</source>
          <volume>8</volume>
          (
          <issue>45</issue>
          ) (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Stenetorp</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pyysalo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Topic</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ohta</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ananiadou</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsujii</surname>
          </string-name>
          , J.:
          <article-title>Brat: a web-based tool for nlp-assisted text annotation</article-title>
          .
          <source>In: Proceedings of the Demonstrations at the 13th Conference of the European Chapter of the Association for Computational Linguistics</source>
          . pp.
          <volume>102</volume>
          {
          <fpage>107</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Xian</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lampert</surname>
            ,
            <given-names>C.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schiele</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akata</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Zero-shot learning-a comprehensive evaluation of the good, the bad and the ugly</article-title>
          .
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Yamada</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Asai</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shindo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Takeda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Takefuji</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Wikipedia2vec: An optimized implementation for learning embeddings from wikipedia</article-title>
          . arXiv preprint arXiv:
          <year>1812</year>
          .
          <volume>06280</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Young</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hazarika</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poria</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cambria</surname>
          </string-name>
          , E.:
          <article-title>Recent trends in deep learning based natural language processing</article-title>
          .
          <source>ieee Computational intelligenCe magazine 13</source>
          (
          <issue>3</issue>
          ),
          <volume>55</volume>
          {
          <fpage>75</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Zavala</surname>
            ,
            <given-names>R.M.R.</given-names>
          </string-name>
          , Mart nez, P.,
          <string-name>
            <surname>Segura-Bedmar</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>A hybrid bi-lstm-crf model for knowledge recognition from ehealth documents</article-title>
          .
          <source>Proceedings of TASS 2172</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>