<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Applying existing Named Entity taggers at BARR IBEREVAL 2017 task</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rodrigo Agerri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>German Rigau rodrigo.agerri@ehu.eus</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IXA NLP Group, University of the Basque Country UPV/EHU</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <fpage>272</fpage>
      <lpage>277</lpage>
      <abstract>
        <p>We present our experiments applying, off-the-shelf, two existing Named Entity Recognition (NER) taggers for the Biomedical Abbreviation Recognition and Resolution (BARR) task at IberEval 2017. The first system is a Perceptron tagger based on sparse, shallow features whereas the second is a bidirectional Long-Short Term Memory neural network with a sequential conditional random layer above it (LSTM-CRF) and initialized with ngram word embeddings. Due to time constraints, we only managed to submit a run from the Perceptron tagger, although in this paper we will also report results with the LSTM-CRF tagger evaluated on the development data. Results show that both Perceptron and LSTMCRF perform reasonably well for SHORT abbreviations whereas the Perceptron model fails to generalize properly for the LONG entity class.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        The literature on biomedical Named Entity Recognition and Classification is rather
extensive, see for example the CHEMDNER tasks [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], but for this first edition of BARR
we were particularly interested in evaluating current, general purpose, existing Named
Entity Recognition (NER) taggers off-the-shelf. The idea was to establish how well can
we perform for free in this task, namely by training and applying existing tools without
feature or hyperparameter tuning. For this exercise we chose two taggers from different
conceptual frameworks, ixa-pipe-nerc [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and a LSTM-CRF NER tagger [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] in order
to annotate biomedical entities and their abbreviations, as illustrated by the following
example:
(1) “Se describe la relacio´ n entre diferentes factores de riesgo cardiovasculares
(FRCV) y la obesidad a partir de una muestra representativa de la poblacio´ n adulta
de Madrid”
      </p>
      <p>The BARR task consists of detecting and classifying both LONG and SHORT
mentions of biomedical entities and to establish that the SHORT mention is an abbreviation
of the LONG entity mention. Thus, in this example “factores de riesgo
cardiovasculares” corresponds to a mention of the LONG entity class whereas “FRCV” would be
a mention of the SHORT class. Furthermore, the SHORT mention is an abbreviation of
the LONG one. Our interest relies purely on entity recognition and, more specifically,
on evaluating current existing taggers and models on the BARR evaluation setting, so
we participated only in the entity recognition subtask, and not on the relation
recognition task.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Methodology</title>
      <p>
        For our experiments we trained on the train1 subset and evaluated on the train2 set.
Tokenization and pre-processing was performed using the IXA pipes tokenizer [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] without
any fine-tuning for the medical domain. The BARR-E background set distributed in the
task [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] is leveraged in order to induce clusters and word embeddings to train both
ixapipe-nerc and LSTM-CRF systems. In the following we briefly describe the architecture
of both systems.
      </p>
      <sec id="sec-2-1">
        <title>2.1 ixa-pipe-nerc</title>
        <p>
          The design of ixa-pipe-nerc aims to establish a simple and shallow feature set,
avoiding any linguistic motivated features, with the objective of removing any reliance on
costly extra gold annotations (POS tags, lemmas, syntax, semantics) and/or cascading
errors if automatic language processors are used. The underlying motivation is to
obtain robust models to facilitate the development of NER systems for other languages
and datasets/domains while obtaining state of the art results. Our system consists of:
(i) Local, shallow features based mostly on orthographic, word shape and n-gram
features plus their context; (ii) three types of simple clustering features, based on unigram
matching; (iii) publicly available gazetteers. ixa-pipe-nerc learns supervised models via
the Perceptron algorithm as described by [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. To avoid duplication of efforts,
ixa-pipenerc uses the Apache OpenNLP project implementation of the Perceptron algorithm1
customized with its own features. Specifically, ixa-pipe-nerc implements, on top of the
local features, a combination of word representation features: (i) Brown [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] clusters,
taking the 4th, 8th, 12th and 20th node in the path; (ii) Clark [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] clusters and, (iii)
Word2vec [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] clusters, based on K-means applied over the extracted word vectors
using the skip-gram algorithm. The implementation of the clustering features looks for the
cluster class of the incoming token in one or more of the clustering lexicons induced
following the three methods listed above. If found, then we add the class as a feature.
The Brown clusters only apply to the token related features, which are duplicated.
        </p>
        <p>
          The ixa-pipe-nerc tagger includes a simple method to combine and stack various
types of clustering features induced over different data sources or corpora, with state of
the art results in newswire Named Entity Recognition [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] and Opinion Target Extraction
[
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], both in out-of-domain and in-domain evaluations.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>LSTM-CRF</title>
        <p>Long Short-Term Memory Networks were design to address the bias of Recurrent
neural networks (RNNs) to learn their most recent input so that long-range dependencies
can be more effectively captured. LSTMs are a family of neural networks that operate
on sequential data. They take as input a sequence of vectors (x1; x2; :::; xn) and return
another sequence (h1; h2; :::; hn) that represents some information about the sequence
at every step in the input. LSTMs try to learn long-range dependencies by using
several gates that control the proportion of the input to give to the memory cell, and the
proportion from the previous state to forget.
1 http://opennlp.apache.org/</p>
        <p>
          Lample et al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] present a hybrid tagging architecture consisting of LSTMs whose
output is then modeled by Conditional Random Fields to tag decisions jointly instead
of independently [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. This architecture is similar to the one presented by Collobert [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
Furthermore, the LSTM-CRF lookup table is initialized using pretrained word
embeddings. They use a variation of word2vec [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] embeddings that accounts for word order
[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. Recurrent models such as RNNs and LSTMs are interesting for the BARR task
because they are theoretically capable of encoding very long sequences. This system
obtained similar results to ixa-pipe-nerc in newswire NER evaluated on CoNLL 2002
and 2003 data.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>
        We train both ixa-pipe-nerc and LSTM-CRF systems with the default parameters and
features as described in Agerri and Rigau [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and Lample et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], using both BILOU
and BIO annotation schemes. The BARR-E background is leveraged to train Brown,
Clark and Word2vec clusters for ixa-pipe-nerc, and ngram word embeddings for
LSTMCRF. Tables 1 and 2 summarize the results of training and evaluating both systems on
the train1 and test2 set, respectively. The reported results for both systems are evaluated
using the CoNLL script for Named Entity Recognition [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        We also submitted the runs in Table 1 to the Markyt evaluation platform [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
Although it seems that the micro F1 score is computed following the CoNLL
evaluation methodology, there were small disagreements between the results obtained using
the CoNLL script and those reported by the Markyt platform. For example, the
set02clark300 run scored 72.31 F1 using the CoNLL script whereas the Markyt scorer
computed 72.24 F1. By looking at the details of the Markyt evaluation, we found a couple
of possible reasons to explain this slight misalignment:
1. Our runs only contained LONG, SHORT and MULTIPLE classes, yet the Markyt
system reported that we annotated some NESTED mentions.
2. The Markyt system reported results for LONG, SHORT, NESTED and UNKNOWN.
      </p>
      <p>However, the shared task description asked for LONG and SHORT classes only.
Moreover, the UNKNOWN class we could not find in any of the annotated data
provided.</p>
      <p>Although we still do not know the official results of the other participants in the
task, it looks like the results for the SHORT class are reasonably good, especially if we
consider that we are applying the systems off-the-shelf. For both classes, ixa-pipe-nerc
Class LONG SHORT
Features Precision Recall F1 Precision Recall F1
OVERALL</p>
      <p>F1
BIO 65.85 69.77 67.75 88.08 85.88 86.97 76.43
BILOU 60.49 74.70 66.85 89.25 87.21 88.22 76.97</p>
      <p>Table 2. LSTM-CRF results training on train1 and testing on train2.
is particularly strong in terms of precision, failing the clustering features to increase
the recall in order to obtain more competitive F1 results. In particular, the results show
that ixa-pipe-nerc fails to generalize on the LONG class. In this sense, the LSTM-CRF
system models much better the LONG class, surpassing ixa-pipe-nerc results for this
entity type by around 7 points in F1 score.</p>
      <p>In terms of official results, we only managed to submit the set02-clark300 due to
the time and formatting constraints of the shared task. In this sense, it would have been
interesting to send the LSTM-CRF runs which seem to be more competitive in this
evaluation setting. The submitted run obtained 65.29 F1 (micro averaged) for entity
recognition (71.78 precision and 59.87 recall). At the time of writing, no further
information was available about the official run (e.g., false positives, evaluation per class,
etc.).
4</p>
    </sec>
    <sec id="sec-4">
      <title>Concluding Remarks</title>
      <p>In this paper we present the results of directly applying, off-the-shelf, two existing NER
taggers for the BARR entity recognition subtask. In this way, we could have an idea of
actual performance of existing systems when applied to other discourse genre and/or
domain. The results show that both systems perform reasonably well considering the
lack of feature and/or parameter tuning performed in this exercise. Both systems scores
are similar for the SHORT class whereas the LSTM-CRF system models better the
long-range dependencies required to perform well for the LONG entity class.</p>
      <p>
        We believe a number of issues made this task unnecessarily difficult. Firstly, it seems
customary now for every new task to create their own dataset formats and evaluation
scripts, even if the task is as old as Named Entity Recognition for which a long tradition
of shared tasks exists [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. In this case, both the documents and entities sets of the
corpus were formatted differently for training and testing, that is, four different formats
needed to be parsed [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>Secondly, the test set required to be processed for every run was unusually large
(26 MB of text distributed in 19K documents, and more than 20 times larger than the
training set) containing, once tokenized using IXA pipes, more than 170K sentences.
This made it quite cumbersome to obtain multiple test runs based on the results reported
in Tables 1 and 2. For example, ixa-pipe-nerc required around 45 minutes to tag the
test for each run whereas the LSTM-CRF tagger required around three hours to do so.
With respect to the training process, ixa-pipe-nerc trained a model in less than a minute
whereas the LSTM-CRF tagger needed around 20 hours to complete 100 epochs on the
950 documents of the training set.</p>
      <p>Finally, there were a number of annotation issues which, in our opinion, did not help
participant systems:
1. Wrong tokenization in entity annotation: expressions such as min, ml, i.e, cm and
EE.UU instead of min., ml., i.e., cm. and EE.UU. were labeled as entities by the
task annotators.
2. Fractions and dashes: hundreds of fraction expressions such as 23/mg and others
containing a forward slash or dash such as radio/tv, angio-TC and h/da were splitted
in the annotation (mistakenly in our opinion). For example, instead of 23/mg being
an entity mention, only mg was annotated. In the same way, instead of angio-TC
being annotated as an entity, only TC was tagged.
3. Entities inside tokens: many entities were annotated inside a token. The most
frequent were mmHg (splitted into mm and Hg entities instead of considering the
token mmHg an entity mention) and H1N1, from which the GLOBAL H and N
entities were annotated, instead of labeling H1N1 as an entity by itself. This decision
means that to be successful in this task entity taggers should work at character level,
because otherwise all these entities inside token words would not be modeled.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This work has been supported by Spanish Ministry of Economy and Competitiveness
(MINECO/FEDER, UE), under the projects TUNER (TIN2015-65308-C5-1-R) and
CROSSTEXT (TIN2015-72646-EXP).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Agerri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bermudez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rigau</surname>
          </string-name>
          , G.:
          <article-title>IXA pipeline: Efficient and ready to use multilingual NLP tools</article-title>
          .
          <source>In: Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)</source>
          . pp.
          <fpage>3823</fpage>
          -
          <lpage>3828</lpage>
          . Reykjavik, Iceland (May
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Agerri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rigau</surname>
          </string-name>
          , G.:
          <article-title>Robust multilingual named entity recognition with shallow semisupervised features</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>238</volume>
          ,
          <fpage>63</fpage>
          -
          <lpage>82</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Brown</surname>
            ,
            <given-names>P.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Desouza</surname>
            ,
            <given-names>P.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mercer</surname>
            ,
            <given-names>R.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pietra</surname>
            ,
            <given-names>V.J.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          :
          <article-title>Class-based n-gram models of natural language</article-title>
          .
          <source>Computational linguistics 18(4)</source>
          ,
          <fpage>467</fpage>
          -
          <lpage>479</lpage>
          (
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Combining distributional and morphological information for part of speech induction</article-title>
          .
          <source>In: Proceedings of the tenth conference on European chapter of the Association for Computational Linguistics-Volume</source>
          <volume>1</volume>
          . pp.
          <fpage>59</fpage>
          -
          <lpage>66</lpage>
          . Association for Computational Linguistics (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5. Collins,
          <string-name>
            <surname>M.</surname>
          </string-name>
          :
          <article-title>Discriminative training methods for hidden markov models: Theory and experiments with perceptron algorithms</article-title>
          .
          <source>In: Proceedings of the ACL-02 conference on Empirical methods in natural language processing-</source>
          Volume
          <volume>10</volume>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Collobert</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A unified architecture for natural language processing: Deep neural networks with multitask learning</article-title>
          .
          <source>In: Proceedings of the 25th international conference on Machine learning</source>
          . pp.
          <fpage>160</fpage>
          -
          <lpage>167</lpage>
          . ACM (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Krallinger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Intxaurrondo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez-Martin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , de la Pea,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Prez-Prez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>PrezRodrguez</surname>
          </string-name>
          , G.,
          <string-name>
            <surname>Santamara</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Villegas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akhondi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loureno</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valencia</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Resources for the extraction of abbreviations and terms in spanish from medical abstracts: the barr corpus, lexical resources and document collection</article-title>
          .
          <source>In: SEPLN</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Krallinger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leitner</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rabal</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vazquez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oyarzabal</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valencia</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of the chemical compound and drug name recognition (chemdner) task</article-title>
          .
          <source>In: BioCreative challenge evaluation workshop</source>
          . vol.
          <volume>2</volume>
          , p.
          <volume>2</volume>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Lafferty</surname>
          </string-name>
          , J.:
          <article-title>Conditional random fields : Probabilistic models for segmenting and labeling sequence data</article-title>
          .
          <source>Computer</source>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Lample</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ballesteros</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kawakami</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Neural architectures for named entity recognition</article-title>
          .
          <source>In: Proceedings of the</source>
          <year>2016</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-</article-title>
          <year>2016</year>
          ). pp.
          <fpage>260</fpage>
          -
          <lpage>270</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ling</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          , Chu-Cheng, L.,
          <string-name>
            <surname>Tsvetkov</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amir</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Astudillo</surname>
            ,
            <given-names>R.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Black</surname>
            ,
            <given-names>A.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Trancoso</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Not all contexts are created equal: Better word representations with variable attention</article-title>
          .
          <source>In: Proceedings of Empirical Methods in Natural Language Processing (EMNLP)</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          . pp.
          <fpage>3111</fpage>
          -
          <lpage>3119</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. Pe´rez-Pe´rez,
          <string-name>
            <given-names>M.</given-names>
            , Pe´rez-Rodr´ıguez, G.,
            <surname>Rabal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            ,
            <surname>Vazquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Oyarzabal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Fdez-Riverola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Valencia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Krallinger</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          , Lourenc¸o,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>The markyt visualisation, prediction and benchmark platform for chemical and gene entity recognition at biocreative/chemdner challenge</article-title>
          .
          <source>Database</source>
          <year>2016</year>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14. San Vicente, I.n.,
          <string-name>
            <surname>Saralegi</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agerri</surname>
          </string-name>
          , R.:
          <article-title>Elixa: A modular and flexible absa platform</article-title>
          .
          <source>In: Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2015</year>
          ). pp.
          <fpage>748</fpage>
          -
          <lpage>752</lpage>
          . Association for Computational Linguistics, Denver, Colorado (
          <year>June 2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Tjong Kim Sang</surname>
            ,
            <given-names>E.F.</given-names>
          </string-name>
          :
          <article-title>Introduction to the CoNLL-2002 shared task: Language-independent named entity recognition</article-title>
          .
          <source>In: Proceedings of CoNLL-2002</source>
          . pp.
          <fpage>155</fpage>
          -
          <lpage>158</lpage>
          . Taipei,
          <string-name>
            <surname>Taiwan</surname>
          </string-name>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>