<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A pipelined approach to Anaphora Resolution in Chemical Patents</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ritam Dutt</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sopan Khosla</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Carolyn Rosé</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Carnegie Mellon University</institution>
          ,
          <addr-line>5000 Forbes Avenue, Pittsburgh, PA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We present our pipelined approach for the sub-task of anaphora resolution in chemical patents as part of the ChEMU shared task at CLEF, 2021. Our approach consists of independently trained mention extraction and relation classification modules. For the former, we set up a BERT-CRF and leverage the BIO scheme to represent the mentions. We include a post-processing step after mention extraction to correct boundary errors and handle nested mentions. For relation classification, we develop a BERT-based model that captures the context between the two candidate mentions to predict the relation between the two. Our final submission ensembles BERT models pretrained on diferent types of clinical data and achieves a Strict F1 of 0.785 on the oficial test set.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Information Extraction</kwd>
        <kwd>Anaphora Resolution</kwd>
        <kwd>Chemical Patents</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Chemical patents play a crucial role in disseminating information about the synthesis, properties,
and applications of new chemical compounds[
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. The rapid publication pace over the past
decade necessitates the need of automated techniques to extract semi-structured knowledge
from the patent text [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ], such as components and process conditions corresponding to chemical
reactions.
      </p>
      <p>
        A key step in understanding how chemical reactions (in patent text) involves identifying
anaphoric dependencies between entities mentioned in the reaction[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These dependencies
involve co-reference relations where diferent surface mentions refer to the same chemical entity,
or bridging relations where diferent entities interact amongst themselves in a particular manner.
The first instance in Table 1 highlights a co-referent relation between N-methylpyrrolidone and
NMP. Likewise, the second instance talks about how the stirring event transforms the mixture.
We describe them in detail in §2.
[N-methylpyrrolidone]1 [(NMP)]2 was stirred for 1 day over CaH2 and finally dis- CR(1, 2)
tilled of.
[The mixture]1 was stirred at room temperature for 1 day. A 2 molL aqueous solu- TR(1, 2)
tion of hydrochloric acid was added to [the mixture]2.
[Acetic acid (9.8 ml)]1 and [water (4.9 ml)]2 were
[the solution of Compound (4) (0.815 g, 1.30 mmol)in THF
[The mixture]4 was stirred for 3 hrs at 50 and then cooled to 0.
added to
(4.9 ml)]3.
      </p>
      <p>RA(1,4),
RA(2,4),
RA(3, 4)
[The mixture]1 was extracted with [ethyl acetate]2 for 3 times. WU(1,3),
[The combined organic layer]3 was washed with water and saturated aque- WU(2,3)
ous sodium chloride.
[Pyrazinecarboxylic acid (152.8 mg, 1.23 mmol, 1 eq)]1and
[H-Phe-OtBu-HCl(253.8 mg, 0.98 mmol, 0.8 eq)]2were charged
[eggplant flask]3
into
an</p>
      <p>CN(1,3),
CN(2,3)
Relation</p>
      <p>We present a pipe-lined approach to solve anaphoric resolution in chemical patents,
comprising of two key phases of Mention Extraction and Relation Classification. We perform ensembling
after each of these two phase to reduce spurious corrleations and improve prediction. We also
incorporate a post-processing module after extracting mentions to handle boundary issues,
discontinuous, and nested spans. We describe our methodology in detail in §3 and Figure 1
illustrates a pictorial representation of our approach.</p>
      <p>
        We provide the experimental details in §4 and present our results in §5. Our proposed
approach achieves a performance of 0.804 F1 score and 0.785 F1 score on the validation and test
set respectively for the strict matching paradigm, successfully beating the proposed baseline
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. For relaxed or inexact match, our scores are substantially higher by almost a margin of 0.07
F1 scores. We conclude and present future ideas in §6.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Description</title>
      <p>
        We focus on the sub-task of anaphora resolution in chemical patents as part of the ChEMU
shared task at CLEF, 2021 1. The task of anaphora resolution seeks to identify the nature of
anaphoric dependencies between mentions/expressions in chemical patents. Prior work [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] has
investigated the following 5 anaphoric dependencies in chemical patents. We present instances
of those in Table 1.
      </p>
      <p>1. Coreference (CR): The relationship between expressions or mentions wherein they refer
to the same chemical Mention.
2. Reaction Associated (RA): The relationship between a chemical compound and its
immediate sources via a mixing/chemical process.
1http://chemu.eng.unimelb.edu.au/chemu/overview
3. Transformed (TR) : The relationship between expressions or mentions, which have
undergone physical changes (e.g., pH and temperature) but have the same chemical
composition.
4. Work Up (WU): The relationship between chemical compounds that were used for isolating
or purifying mentions, and their corresponding outputs.
5. Contained (CN): The association between chemical compounds and the equipment in
which they are placed.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>We outline the details of a pipelined architecture to anaphora resolution in this section. Our
approach consists of two major steps of Mention Extraction and Relation Classification. In the
Mention Extraction phase, we identify all possible mentions from the patent text, whereas in
the Relation Classification phase, we infer whether a given pair of mentions have any anaphoric
dependency between them. We describe the neural architecture we have employed for these
two phases.</p>
      <sec id="sec-3-1">
        <title>3.1. Mention Extraction</title>
        <p>
          Prior work has demonstrated the success of neural architectures in extracting chemical and
bio-medical mentions [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], spans of chemical reaction [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and the specific roles of mentions in a
reaction [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
        <p>After cooling, the solid was ... washed with cold dichloromethane to give
N-(4-(2-oxo-1,2,3,4tetrahydroquinolin-6-yl)thiazol-2-yl)oxazole-5-carboxamide (0.121 g, 87%) as a beige solid</p>
        <p>Linear Conditional Random Field</p>
        <p>Self attention + Feed-forward x 12
After cooling, the solid was ... washed with cold dichloromethane to give
N-(4-(2-oxo-1,2,3,4tetrahydroquinolin-6-yl)thiazol-2-yl)oxazole-5-carboxamide (0.121 g, 87%) as a beige solid</p>
        <p>
          In this task, we consider any span of text snippet which was annotated either as an antecedent
or an anaphora as a mention. Based on the annotation corpus of [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], mentions include quantified
chemical compounds (0.51 g of methanol, K2CO3 (300 mg, 2.2 mmol)), proper nouns (DMF,
(2,6-dichloro-4-fluorophenyl)hydrazine hydrochloride ), identifiers (5i, 4a), pronouns (it, they,)
and noun phrases (the solvent, an autoclave). We note that approximately 3% of mentions in
the dataset have dis-continuous spans, and we leverage post-processing techniques to deal with
such spans.
        </p>
        <p>We thus, model the task of mention extraction as a sequence labeling task. For this phase, we
encode the longest contiguous span of text that includes these individual, discontinuous spans
as the span of the given mention. Based on the recent success of transformer-based modules
like [6] in information extraction [7, 8, 9], we employ a similar approach in our case. We use a
transformer-based encoder to encode the mentions and then pass the encoding through a Linear
Conditional Random Field (CRF) [10]. An overview of the Mention Extraction architecture is
shown in Figure 2.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Relation Classification</title>
        <p>WORK UP</p>
        <p>Linear</p>
        <p>WORK UP</p>
        <p>Linear</p>
        <p>COREFERENCE</p>
        <p>Linear
Mean</p>
        <p>Mean</p>
        <p>Mean</p>
        <p>Mean</p>
        <p>Self attention + Feed-forward x 12
After cooling, the solid was ... washed with cold dichloromethane to give
N-(4-(2-oxo-1,2,3,4tetrahydroquinolin-6-yl)thiazol-2-yl)oxazole-5-carboxamide (0.121 g, 87%) as a beige solid</p>
        <p>We present an overview of the Relation Classification architecture in Figure 3. For a given
pair of mentions, we define the context for a pair of mentions as the sequence of sentences
that has the mention pair. We pass the context through a transformer Based encoder, and use
mean-pooling over the individual mention tokens to obtain a representation corresponding
to that mention. We concatenate the representations of the two mention spans and project it
through a linear layer of 6 classes. These correspond to the 5 anaphoric dependencies and the
No-Relation class if there is no dependency between the pair of mentions.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>4.1. Mention Extraction</title>
        <p>
          For the task of mention extraction, we experiment with several transformer-Based encoder
modules, such as BERT [6], Clinical BERT (trained on clinical notes) [11] and Pubmed-BERT
(trained on Pubmed abstracts) [12]. Moreover, since chemical compounds are often several
characters long, a single compound could be decomposed into several tokens. To circumvent
this tokenization issue, we include the “LONG TOKEN” similar to [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], as a special token, to
subsume the remaining tokens of a compounds beyond a certain length. For our experiments,
the length is kept to 25. We use the BIO (Beginning Inside Outside) scheme to represent the
mentions. For example “the residue is heated” will be converted to “B-ENT I-ENT O O”.
        </p>
        <p>
          We evaluate mention extraction in terms of precision, recall, and F1 score, for both exact
(strict) and inexact (relaxed) match similar to [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. We use the BRAT evaluation script provided
by the organizers to compute the scores. We ran our models using the huggingface transformers
library in PyTorch, with a batch size of 8, a learning rate of 1e-5, a dropout of 0.1, Adam optimizer,
and patience of 5.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Post-Processing</title>
        <p>To correct boundary errors and extract nested spans, we further post-process the output from
the neural mention extractor using several rule-based sieves. The sieves were chosen after close
inspection on the validation data and are described in detail in the subsequent section.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Relation Classification</title>
        <p>
          For the task of relation classification, we experiment with several transformer-Based encoder
modules, namely BERT-Base and BERT-large [6], Clinical BERT [11], Pubmed-BERT [12] and
BioBERT [13]. Moreover, since we have to check anaphoric dependency between all possible
pairs of entities during validation and testing, it is imperative to incorporate negative instances
during training. Thus, all pairs of entities which do not have an anaphoric dependency
between them are taken as negative instances and assigned the “NO RELATION” label. We also
experiment by varying the proportion of negative instances during training. Similar to mention
extraction, we use the BRAT evaluation script provided by the organizes. We find the precision,
recall, and F1 score for the relation classification task, similar to [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
        </p>
        <p>We ran our relation classification models using the huggingface transformers library in
PyTorch, with a batch size of 16, a learning rate of 3e-5, a dropout of 0.1, Adam optimizer, and
patience of 5. We curated negative samples by pairing mentions that were not more than 10
mentions apart from each other in the patent document.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Ensembling</title>
        <p>We perform ensembling twice, once after the mention extraction phase and once after the
relation classification phase. We carry out majority voting over the outputs of five models and
consider only those outputs which have been predicted by at least three models. The outputs
correspond to the extracted mention-span for mention extraction, and over pairs of extracted
span and their corresponding relation label for relation classification. Ensembling has been
proven to reduce spurious correlations and improve performance [14], and has been employed
for several tasks [15, 16, 9].</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Analysis</title>
      <sec id="sec-5-1">
        <title>5.1. Mention Extraction</title>
        <p>We note the results for mention extraction in Table 2. At the outset, we observe that the models
achieve almost 0.98 F1 score for the inexact (relaxed) match. However, they sufer almost a
0.09 drop in F1 score under the exact (strict) match evaluation. A majority of the errors occurs
due to nested spans or discontinuous spans (see Example 4 in the post-processing sub-section).
Another common issue is the omission or inclusion of tokens at the beginning/end of the
span like (see Examples 1,2, and 3 in the post-processing sub-section), which we refer to as
“boundary issues”. Since the entities are multi-faceted and can range from simple noun-phrases
and identifiers to complex chemical quantifiers, it makes the task exceedingly challenging.</p>
        <p>We also observe that incorporating tokenizers corresponding to bio-medical and clinical
domains perform slightly better than the uncased BERT model. The best performance is observed
for PubMed-BERT-Long with F1 score of 0.890, as opposed to 0.865 for the uncased BERT-Long
model. Inclusion of the “LONG TOKEN” is shown to benefit PubMed-BERT and BERT-uncased
but fares worse for the Clinical-BERT model.</p>
        <p>Unsurprisingly, ensembling over these 5 models achieves the highest score in the exact match
setting across all three metrics. We use the entities extracted from the ensemble setting as the
ifnal ones and post-process them before the relation classification phase.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Post-Processing</title>
        <p>For post-processing, we pass the outputs of the mention extraction module through several
sequential rule-based sieves:
1. If the extracted mention ends with strings like {’ and’, ’ under’, ’ or’, ’ over’}, remove
them from the mention. E.g. [Alcohol and]1 → [Alcohol]1 and
2. If a mention is preceded by an article like {’a’, ’the’}, include that article in the mention.
3. If the extracted mention ends with {’ with’, ’ of’, ’ in’} and there is an adjoining mention
after it, combine the two. E.g.</p>
        <p>
          [ethanol in]1 [the reaction mixture]2 → [ethanol in the reaction mixture]1
4. We observed that often the patent documents refer to compounds with an ID. This ID is
annotated as a coreferent mention to the actual compound. E.g.
[7-fluorobenzofuran3(2H)-one]1 [84c]2 [(340 mg, 2.2 mmol)]1. But since our neural model can only extract
contiguous mentions, it outputs [7-fluorobenzofuran-3(2H)-one ]1 [84c (340 mg, 2.2
mmol)]2. In order to extract these coreferents from outputs with such patterns we identify
instances that follow the below template. If predicted m2 starts with a word (w1) which
follows regex ([
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">0-9</xref>
          ]+[a-z]+), and contains a second word (w2) which starts with a {’(’},
then combine m2 with m1 excluding w1 which is separated out as the coreferent.
We find that this post-processing substantially improves the performance (from 0.895 F1 to 0.922
F1) on the oficial mention extraction Strict metrics (Table 2). Each of our sieves work towards
increasing the number of exact matches between gold and system mentions. Furthermore, Sieve
3 and 4 also uncover new spans simultaneously impacting the Relaxed Match scores (from 0.967
F1 to 0.970 F1).
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Relation Classification</title>
        <p>We present the performance of models for the relation classification task for both the gold and
predicted mentions in Table 3. We observe that pre-trained Pubmed-BERT and BioBERT fares
slightly better than the uncased BERT-Base and BERT-Large model, highlighting again the
benefits of pre-training on clinical data. Moreover, our pair-wise relation classification approach
achieves nearly a 0.91 F1 score for anaphoric relations using the gold mentions. We also observe
empirically that including 100% of all negative examples during training achieves the highest
performance. Models that are trained on 5% and 10% of the total negative samples achieve
an F1-Score (gold entities) of 0.762 and 0.794 respectively, around 0.15 F1 points below their
100% counterpart. A majority of the misclassification errors takes place when an anaphoric
dependency between a pair of mentions is predicted as “NO RELATION” and vice versa, since the
negative classes account for around 87% of all labels. The only other instance of misclassification
takes place where the RA relation is predicted as WU, since both describe associations between
chemical compounds.</p>
        <p>
          Unlike [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], our pair-wise approach can circumvent the problems of discontinuous and nested
spans and hence we include those mentions. Nevertheless, we note how the errors in the mention
extraction phase propagate downstream and downgrade the relation classification performance
on the predicted entities. This results in an average score of 0.79 F1 for predicted entities (a
drop in approximately 0.12 F1 points). While it would be prudent ideally to carry out the two
phases in a joint fashion like [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] (to prevent cascading errors), the crucial post-processing step
of fixing boundary issues and extracting additional nested mentions, necessitates the pipelined
approach.
        </p>
        <p>In fact, our architecture beats the transformer baseline that performs joint co-reference and
bridging on both the validation and test set by 0.03 F1 score on the exact match metric. The
boost for relaxed match is substantially higher, with our model outperforming the baseline
by approximately 0.07 F1 score on both validation and test. Moreover, ensembling over the
diferent models boosts the performance further by 0.01 F1 for both the gold and predicted
entities.</p>
        <p>
          We report the performance for the 5 individual anaphoric relations in Table 4. Coreference (CR)
relations with their nuanced rules and long-term dependencies have the poorest performance
[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], whereas the bridging relations being more local and specific in nature, fared considerably
better. We acknowledge there is immense scope for improvement and posit how incorporating
additional information like events or entity types can help bolster performance. We defer this
exploration for future work.
        </p>
        <p>Our final performance on the validation set was 0.804 F1 and 0.887 F1 for the strict and
relaxed match respectively. Likewise, our performance on the test set was 0.785 and 0.872 F1
for strict and relaxed match. We are currently ranked the first in the shared task.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>Resolving anaphora dependencies in chemical patents plays a key role in understanding the
nuances of how chemical reactions are described, and the interactions between participating
entities. We describe a pipelined approach to address this challenge using independently trained
mention extraction and relation classification modules. Such a design choice facilitates the
inclusion of a rule-based post-processing module to handle boundary errors, and
discontinuous/nested spans. We achieve a Strict F1 score of 0.785 and a relaxed F1 score of 0.872 on the
oficial test set, significantly outperforming the baseline.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgement</title>
      <p>We thank the anonymous reviewers for their insightful comments. This was funded in part by
NSF grants (IIS 1917668 and IIS 1822831, IIS 1949110, and IIS 1546393) and funding from Dow
Chemical.
[6] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, in: Proceedings of the 2019 Conference of
the North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, Volume 1 (Long and Short Papers), 2019, pp. 4171–4186.
[7] S. Vashishth, D. Newman-Grifis, R. Joshi, R. Dutt, C. Rose, Improving broad-coverage
medical entity linking with semantic type prediction and large-scale datasets, arXiv
preprint arXiv:2005.00460 (2020).
[8] A. Thillaisundaram, T. Togia, Biomedical relation extraction with pre-trained language
representations and minimal task-specific architecture, in: Proceedings of The 5th
Workshop on BioNLP Open Shared Tasks, Association for Computational Linguistics,
Hong Kong, China, 2019, pp. 84–89. URL: https://www.aclweb.org/anthology/D19-5713.
doi:10.18653/v1/D19-5713.
[9] J. W. Y. R. Z. Zhang, Y. Zhang, Melaxtech: A report for clef 2020–chemu task of chemical
reaction extraction from patent, in: Working Notes of CLEF 2020–Conference and Labs of
the Evaluation Forum, Vol. 2696. CEUR Workshop Proceedings, 2020.
[10] J. Laferty, A. McCallum, F. Pereira, Conditional random fields: Probabilistic models for
segmenting and labeling sequence data: proceedings of the 18th international conf. on
machine learning, 2001, San Francisco, CA, USA (2001).
[11] E. Alsentzer, J. Murphy, W. Boag, W.-H. Weng, D. Jindi, T. Naumann, M. McDermott,
Publicly available clinical BERT embeddings, in: Proceedings of the 2nd Clinical Natural
Language Processing Workshop, Association for Computational Linguistics, Minneapolis,
Minnesota, USA, 2019, pp. 72–78. URL: https://www.aclweb.org/anthology/W19-1909.
doi:10.18653/v1/W19-1909.
[12] Y. Peng, S. Yan, Z. Lu, Transfer learning in biomedical natural language processing: An
evaluation of BERT and ELMo on ten benchmarking datasets, in: Proceedings of the 18th
BioNLP Workshop and Shared Task, Association for Computational Linguistics, Florence,
Italy, 2019, pp. 58–65. URL: https://www.aclweb.org/anthology/W19-5006. doi:10.18653/
v1/W19-5006.
[13] J. Lee, W. Yoon, S. Kim, D. Kim, S. Kim, C. H. So, J. Kang, Biobert: A pre-trained
biomedical language representation model for biomedical text mining, Bioinformatics (Oxford,
England) 36 (2020) 1234–1240.
[14] Z. Allen-Zhu, Y. Li, Towards understanding ensemble, knowledge distillation and
selfdistillation in deep learning, arXiv preprint arXiv:2012.09816 (2020).
[15] S. Khosla, Emotionx-ar: Cnn-dcnn autoencoder based emotion classifier, in: Proceedings
of the sixth international workshop on natural language processing for social media, 2018,
pp. 37–44.
[16] S. Khosla, R. Joshi, R. Dutt, A. W. Black, Y. Tsvetkov, Ltiatcmu at semeval-2020 task 11:
Incorporating multi-level features for multi-granular propaganda span identification, in:
Proceedings of the Fourteenth Workshop on Semantic Evaluation, 2020, pp. 1756–1763.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Akhondi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Rey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schwörer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Toomey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nau</surname>
          </string-name>
          , G. Ilchmann,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sheehan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Irmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bobach</surname>
          </string-name>
          , et al.,
          <article-title>Automatic identification of relevant chemical compounds from patents</article-title>
          ,
          <year>Database 2019</year>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Akhondi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Druckenbrodt</surname>
          </string-name>
          , T. Cohn,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gregory</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Verspoor</surname>
          </string-name>
          ,
          <article-title>Improving chemical named entity recognition in patents with contextualized word embeddings</article-title>
          ,
          <source>in: Proceedings of the 18th BioNLP Workshop</source>
          and Shared Task, Association for Computational Linguistics, Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>328</fpage>
          -
          <lpage>338</lpage>
          . URL: https: //www.aclweb.org/anthology/W19-5035. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>W19</fpage>
          -5035.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Druckenbrodt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Akhondi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Baldwin</surname>
          </string-name>
          , K. Verspoor,
          <article-title>ChEMU-ref: A corpus for modeling anaphora resolution in the chemical domain</article-title>
          ,
          <source>in: Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics:</source>
          Main Volume,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>1362</fpage>
          -
          <lpage>1375</lpage>
          . URL: https://www.aclweb.org/anthology/2021.eacl-main.
          <volume>116</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yoshikawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Druckenbrodt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hoessel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Akhondi</surname>
          </string-name>
          , T. Cohn,
          <string-name>
            <given-names>T.</given-names>
            <surname>Baldwin</surname>
          </string-name>
          , et al.,
          <article-title>Chemu: named entity recognition and event extraction of chemical reactions from patents</article-title>
          ,
          <source>in: European Conference on Information Retrieval</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>572</fpage>
          -
          <lpage>579</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Yoshikawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Druckenbrodt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Akhondi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Baldwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Verspoor</surname>
          </string-name>
          ,
          <article-title>Detecting chemical reactions in patents</article-title>
          ,
          <source>in: Proceedings of the The 17th Annual Workshop of the Australasian Language Technology Association, Australasian Language Technology Association</source>
          , Sydney, Australia,
          <year>2019</year>
          , pp.
          <fpage>100</fpage>
          -
          <lpage>110</lpage>
          . URL: https://www.aclweb.org/anthology/U19-1014.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>