<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cross-lingual transfer-learning approach to negation scope resolution</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anastassia Shaitarova</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lenz Furrer</string-name>
          <email>lenz.furrerg@uzh.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabio Rinaldi</string-name>
          <email>fabio.rinaldi@idsia.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dalle Molle Institute for Artificial Intelligence Research</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computational Linguistics, University of Zurich</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Detecting instances of negation in text is crucially important for several applications, yet it is often neglected. Several decades of research in automated negation detection have not yet provided a reliable solution, especially in a multilingual context. Negation scope resolution poses particular challenges since identifying the scope of influence of a negation cue in a sentence requires a deeper level of natural language understanding. Little work has been done on negation scope resolution in languages other than English. Meanwhile, transfer learning is in wide use and large multilingual models are available to the public. This paper explores the feasibility of a cross-lingual transfer-learning approach to negation scope resolution. Preliminary experiments with the Multilingual BERT model and data in English, French, and Spanish show solid results with the highest F1-score 84.73 on zeroshot transfer between English and French.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>Negation detection has been a kind of stumbling
block in NLP research for many years. Linguists
have not yet presented a proper generalization that
can explain this highly complex linguistic
phenomenon. The ability to negate is what makes us
human, claims Horn (2001), which is not a
comforting message to the NLP community, where
human is practically synonymous with ambiguous.</p>
      <p>Indeed, negation expresses itself through high
syntactic and morphological variability which
complicates automated detection significantly.
However, there is a certain degree of logical and
semantic uniformity in negation, which is
exhibited cross-linguistically. This uniformity can be
potentially exploited by the new state-of-the-art
NLP models. The work on negation detection is
further complicated by the sparsity of annotated
data, particularly in languages other than English.
Therefore the search for annotation-independent
approaches must continue.</p>
      <p>The task of negation detection classically
consists of two steps: 1) negation trigger or cue
detection, and 2) negation scope resolution. Negation
cues are words (no, not) or parts of a word
(unin unhealthy) that signal negation, while negation
scope includes the part of a sentence that is
semantically influenced by this signal. Identifying
scope is more computationally challenging since
the sphere of influence of each negation cue
depends on a number of factors.</p>
      <p>In this research we explore the ability of a
multilingual BERT model (here: mBERT) released
by Devlin (2018) to solve a fine-grained
linguistic task of negation scope resolution across
languages. We focus on surface form pertinent
negations. The scope tokens are selected based on a
binary classification negated/affirmed. Our research
is guided by two main objectives:</p>
      <p>explore the feasibility of a zero-shot model
transferring approach. In other words, can a model
that is fine-tuned on labeled data in one language
resolve negation scope in another language?
test the possibility of using very few labeled
examples as training data.</p>
      <p>In Section 2 we highlight studies and
approaches relevant for the question of cross-lingual
negation scope resolution. Section 3 describes the
datasets involved in this study. The preliminary
experiments and their results are described in
Section 4.</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Computational endeavours in negation detection
began in the medical domain since it has a direct
impact on the reliability of diagnosis. NegFinder
        <xref ref-type="bibr" rid="ref28">(Mutalik et al., 2001)</xref>
        and NegEx
        <xref ref-type="bibr" rid="ref5">(Chapman et al.,
2001)</xref>
        were the first two algorithms specifically
designed to handle negation. These and other early
systems regarded negation detection from a
medical rather than a linguistic perspective. Evaluation
accounted only for correctly identified negated
medical terms. The growing demand for reliable
negation detection in other computational fields
such as, inter alia, sentiment analysis
        <xref ref-type="bibr" rid="ref9">(Councill
et al., 2010)</xref>
        , textual entailment
        <xref ref-type="bibr" rid="ref24">(de Marneffe et al.,
2006)</xref>
        , and machine translation
        <xref ref-type="bibr" rid="ref3">(Baker et al., 2010)</xref>
        promoted the development of negation detection
as an NLP task in its own right.
      </p>
      <p>
        Early lexical rule-based approaches like NegEx
and its expanded versions
        <xref ref-type="bibr" rid="ref16">(e.g. ConText, Harkema
et al., 2009)</xref>
        used a fixed scope length that depends
on a predefined number of tokens and the
presence of delimiter words such as but and however.
This is often sufficient for clinical texts that
commonly contain short and/or ungrammatical
sentences. Researchers also observed that negation
scope is trully syntactic in nature
        <xref ref-type="bibr" rid="ref25 ref26 ref34 ref35">(Szarvas et al.,
2008; Morante and Blanco, 2012)</xref>
        . Thus, most
other rule-based, hybrid, and machine learning
systems used syntax and dependency parsing to
identify negation scope patterns. Context-free
grammars were often used to employ these
patterns as rules.
      </p>
      <p>The entire first decade of research, however,
was dedicated to solving negations in English.
Only later, and gradually, the work on negation
detection in other languages started taking place.
2.1</p>
      <sec id="sec-2-1">
        <title>NegEx and its cross-lingual adaptations</title>
        <p>
          The NegEx algorithm has been adopted for several
European languages including Swedish
          <xref ref-type="bibr" rid="ref33">(Skeppstedt, 2010)</xref>
          , French
          <xref ref-type="bibr" rid="ref25 ref26">(Dele´ger and Grouin, 2012)</xref>
          ,
German
          <xref ref-type="bibr" rid="ref6">(Chapman et al., 2013)</xref>
          , and Spanish
          <xref ref-type="bibr" rid="ref7">(Costumero et al., 2014)</xref>
          . A comparative study
conducted by Chapman et al. revealed several
significant challenges associated with the adaptation
process, the main one being the collection of
triggers.
        </p>
        <p>Negation trigger lists are inherently incomplete
despite the fact that, following Zipf’s law, a
handful of triggers is responsible for most of the
negations in the text. A set of negation triggers depends
greatly on the type of text in which it is found.
Even if texts belong to the same domain but are of
different types (e.g. radiology reports, discharge
summaries, and surgical notes), they contain
different negation triggers.</p>
        <p>NegEx rules can be considered language
agnostic but the compilation of negation cues has to be
language specific. The same triggers have
various levels of ambiguity and usage frequency
depending on the language. Whether a cue negates
to its left or right varies from language to
language. Translating cues from English into
languages that have a richer morphosyntactic
variability (like French) or exhibit different or more
flexible word order (like Swedish) introduces
additional problems.</p>
        <p>The restrictions described above make it
advisable that a native speaker, and preferably a
domain specialist, is involved in the compilation of a
NegEx trigger list in the target language. Abdaoui
et al. (2017) involved a bilingual text mining
specialist to validate French cues automatically
translated from English. A specialist, however, is not
always available.</p>
        <p>
          Despite all these difficulties NegEx is still in
wide use because it is simple, reliable and needs
no labelled data. Some researchers show that
NegEx is enough for biomedical texts
          <xref ref-type="bibr" rid="ref13 ref8">(Cotik et al.,
2016; Elazhary, 2017)</xref>
          and assert that no other
complex approach is necessary. Others disagree,
referring to the inherent inability of rule-based
systems to generalize
          <xref ref-type="bibr" rid="ref32 ref37">(Wu et al., 2014; Sergeeva
et al., 2019)</xref>
          .
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Machine learning approaches</title>
        <p>The first study focusing on negation scope
resolution was spearheaded by Morante et al. (2008).
They framed it as a chunking problem and proved
that it can be handled as a classification task at a
token level. They used a k-nearest neighbor
algorithm to assess each token in relation to the
negation trigger.</p>
        <p>
          The work was conducted on the BioScope
corpus
          <xref ref-type="bibr" rid="ref34 ref34 ref35 ref35">(Szarvas et al., 2008; Vincze et al., 2008)</xref>
          .
BioScope was the first publicly available sizeable
corpus annotated for negation cues and scopes. It
contains biomedical texts in English. The results
showed an F1-score of 88.40 for negation scope
resolution with the use of gold-standard cues.
        </p>
        <p>Morante and Blanco (2012) went further and
organized the *SEM2012 Shared Task that was
dedicated entirely to resolving the scope and focus of
negation. Additionally, in order to compensate for
the lack of negation training data in the general
domain, they annotated several Sherlock Holmes
stories by Sir Arthur Conan Doyle. The best
F1score for scope detection on the Conan Doyle
corpus (here: Sherlock) was 85.26, which was later
outperformed with a score of 88.2 by Packard et al.
(2014).</p>
        <p>
          Fancellu et al. (2016) noted that previous
systems for negation scope resolution were highly
engineered, parser-dependent, and specific to
English. They outperformed all previous results on
the Sherlcok corpus with an F1-score of 88.72
using BiLSTMs, pre-trained embeddings, and
universal POS tags. The team then worked with a
parallel English-Chinese corpus, NegPar
          <xref ref-type="bibr" rid="ref23">(Liu et al.,
2018)</xref>
          and asked the question that inspired this
paper’s research:
        </p>
        <p>Can we learn a model that detects
negation scope in English and use it in a
language where annotations are not
available? (Fancellu et al., 2018, p. 1)</p>
        <p>They used universal dependencies to abstract
away from word order that differs between
languages. The best performing cross-lingual model
reached an F1-score of 72.46 and set the precedent
for zero-shot cross-lingual negation scope
detection.</p>
        <p>
          When Google AI open-sourced the
Bidirectional Encoder Representation for Transformers
          <xref ref-type="bibr" rid="ref12">(BERT: Devlin et al., 2019)</xref>
          , it “marked the
beginning of a new era in NLP”
          <xref ref-type="bibr" rid="ref2">(Alammar, 2018)</xref>
          .
Negation detection research followed suit,
exploiting the transformer’s architectural features as well
as linguistic knowledge gained from pre-training
on massive amounts of data
          <xref ref-type="bibr" rid="ref17 ref21 ref30 ref31 ref32 ref4">(Khandelwal and
Sawant, 2019; Britto and Khandelwal, 2020)</xref>
          . The
results showed solid improvement on scope
resolution across domains in English. The developed
architecture, NegBERT, was customized for the
experiments in this paper.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Data</title>
      <p>
        There are several negation annotated corpora
available to the public, but most of them are in
English. Moreover, they all suffer from a lack
of standardization in terms of annotation
guidelines
        <xref ref-type="bibr" rid="ref20">(Jime´nez-Zafra et al., 2020)</xref>
        . For example,
Sherlock
        <xref ref-type="bibr" rid="ref25 ref26">(Morante and Daelemans, 2012)</xref>
        is the
only corpus in English that annotates
morphological negation cues such as affixes. It also allows
discontinuous scopes which is not the case in other
corpora. Sherlock and BioScope differ on whether
they include the negation cue and the clause’s
subject into the scope. Both corpora are freely
available for download.
      </p>
      <p>
        The SFU Review-NEG corpus
        <xref ref-type="bibr" rid="ref22">(Konstantinova
et al., 2012)</xref>
        , a large multi-domain corpus of
product reviews, mostly follows BioScope’s guidelines
and does not include cues into the scope of
negation. It is available on request from Simon Fraser
University. We use all three English corpora in our
experiments. We also combine them together into
one data set that provides 7044 negation sentences.
      </p>
      <p>
        The French data that we use here is described
in Dalloux et al. (2017) and is publicly available
on request1. It has 3790 sentences total and is
loosely modeled on the Sherlock corpus. The data
in Spanish comes from the SFU ReviewSP-NEG
corpus
        <xref ref-type="bibr" rid="ref19">(Jime´nez-Zafra et al., 2018)</xref>
        that can be
requested via the same link as The SFU
ReviewNEG corpus above. It has 9445 sentences and
its annotations follow the guidelines of the three
aforementioned English corpora.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>We used NegBERT for our experiments but
instead of the BERT-base-uncased model we
imported bert-base-multilingual-cased. mBERT has
been pre-trained on non-parallel data in 104
languages with the same training objectives as BERT:
masked language modeling and next sentence
prediction. The switch between the models is
simple since they share a tokenizer and, for
our task, can be initialized with a built-in class
call BertForTokenClassification from
HuggingFace2 .</p>
      <p>The preprocessing of the three aforementioned
English corpora is completed by NegBERT which
duplicates sentences with multiple negations into
multiple copies containing a single negation. The
French and Spanish corpora are stored in formats
that differ from the English corpora. We
preprocessed these corpora ourselves and, for the
sake of simplicity and consistency in the
experiments, considered only sentences with one
negation scope. Here we refer to these subcorpora
1http://people.irisa.fr/Clement.</p>
      <p>Dalloux/</p>
      <p>2https://huggingface.co/transformers/
_modules/transformers/modeling_bert.html
SHERLOCK
BIOSCOPE</p>
      <p>SFU
EN
FR</p>
      <p>SP
as OneScopeFR for French and OneScopeSP for
Spanish. OneScopeFR consists of 717 sentences
while OneScopeSP has 2197.</p>
      <p>Following our objectives, we trained models
that can be used for zero-shot transfer and
minimal training data experiments. Thus we fine-tuned
mBERT on the English corpora, OneScopeSP,
OneScopeFR, mini subsets of OneScopeFR and
OneScopeSp as well as various combinations of
corpora and subsets. All models were fine-tuned
with MAX LEN 250, batch size 8, and learning
rate 1e-5. The reported F-scores are calculated on
the per-token basis for label 1 (token in scope).
4.1</p>
      <sec id="sec-4-1">
        <title>Zero-shot model transfer vs. rule-based approach for biomedical French texts.</title>
        <p>Since the French corpus is a collection of
medical reports, we decided to assess the ease and
efficacy of NegEx and use the results as a
baseline. We used the publicly available NegEx
adaptation for Python3 which comes with cues in
English. In order to compile a list of triggers in
French we initially contacted Dele´ger and Grouin.
They readily provided the list they had produced
for their own adaptation of NegEx. Processing
OneScopeFR with these triggers resulted in an
F1score 45.55%.</p>
        <p>In order to improve the result and to have a
fair comparison, we collected gold negation cues
from OneScopeFR. Consecutive multiword cues
such as absence de (en: lack of) were collected
as multiword units. Non-consecutive multiword
cues such as ne ... pas ... ni ... ni, sans ... ne ...
aucun were collected as single token cues ne, pas, ni,
sans, aucun. The final list consisted of 27 triggers.</p>
        <p>Manual examination of the data helped decide
on the directionality of each trigger. For example,
our observations revealed that absent (en: absent)
3https://code.google.com/archive/p/
negex/downloads?page=2
and ne´gatif (en: negative) and their
morphological variations generally affect context to the left
while all the other cues negate to the right. As
a result, 22 right- and five left-negating cues were
added. Scope delimiters such as mais (en: but) and
pseudo-triggers such as ne cause pas (en: does not
cause) were copied from the original French list.
Additionally, NegEx was modified not to include
the full stop in the scope.</p>
        <p>These adjustments boosted the results up to
an F1-score of 63.89%. OneScopeFR was also
tested with mBERT fine-tuned on the combined
English data which produced the best zero-shot
result across the board with an F1-score of 84.73%.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Zero-shot and minimal training data for</title>
      </sec>
      <sec id="sec-4-3">
        <title>French and Spanish texts.</title>
        <p>In this set of experiments we test zero-shot
approach using both OneScopeFR and OneScopeSP.
Additionally, we create minimal training data sets
for both languages by randomly selecting 125
sentences from each subcorpus. In the process of
finetuning mBERT on these miniature sets, 20% of the
data (25 sentences) is used for validation and
prevention of overfitting through early stopping. We
call the resulting models FR100 and SP100.</p>
        <p>The motivation behind this setup is a potential
situation where annotated material for a particular
language might not exist. It is, however,
conceivable for a hypothetical researcher or a developer
to be able to annotate about a hundred sentences
without unreasonable effort.</p>
        <p>Table 1 shows the results of our experiments.
All the models are tested on sentences that are
left in OneScopeFR and OneScopeSP after the
extraction of the mini-sets. Thus, frXX contains 592
sentences in French and spXX has 2072 sentences
in Spanish.</p>
        <p>The results suggest that fine-tuning mBERT
even on one hundred sentences produces results
that are worth considering. FR100 ! frXX shows
an F1-score of 83.68, SP100 ! spXX is at 82.72.</p>
        <p>These results are further improved by adding
available training material in other languages to
the mini-set. The best score for French (87.82)
is achieved by the addition of all English data
EN FR100 ! frXX, while Spanish benefits most
from the addition of French FR SP100 ! spXX:
83.87. The highest score for zero-shot transfer is
achieved by EN ! frXX which is consistent with
the previous experiment.
4.3</p>
      </sec>
      <sec id="sec-4-4">
        <title>Discussion</title>
        <p>The results produced by our experiments are
positive. With low effort and little to no training data
we can obtain results comparable to former
stateof-the-art models. It is difficult to say, however,
how much this outcome was to be expected.</p>
        <p>
          Pires et al. (2019) showed that mBERT
performs best for languages that share the same word
order features. English, French, and Spanish
are all SVO languages, meaning that their
sentences mostly follow the subject-verb-object
pattern. Thus, it was reasonable to expect good
results from our experiments. When it comes to
negation, however, the languages differ
typ
          <xref ref-type="bibr" rid="ref10">ologically. Dahl (1979</xref>
          ) examined negation patterns in
240 different languages in 40 different language
families and concluded that English, French, and
Spanish belong to three different categories. In
English the negation marker immediately follows
the verb, but in Spanish the marker precedes it,
while French shows both patterns.
        </p>
        <p>
          A look into neumerous studies dedicated to
understanding BERT did not provide clear ideas
of what to expect from our experiments. On
one hand, BERT’s contextualized representations
seem to possess robust knowledge over syntactic
and dependency parse trees
          <xref ref-type="bibr" rid="ref17 ref21 ref30 ref32">(Hewitt and Manning,
2019)</xref>
          . On the other hand, Warstadt et al. (2019)
claim that the syntactic knowledge in BERT
appears to be partial and inconclusive. Rogers et al.
(2020) analyzed over forty studies on BERT and
concluded that BERT does not understand
negation.
        </p>
        <p>An examination of the output provides our own
insights. An example of a French sentence with
its translation into English and gold standard
annotation is shown below. The negation cue is
marked in bold, while scope is enclosed between
the [square brackets].</p>
        <p>French: Il n’ [y avait] pas [de syndrome
inflammatoire biologique] (SIB) et les bilans
phosphocalciques sanguin et urinaire e´taient normaux .</p>
        <p>English: [There was] no [biological
inflammatory syndrome] (BIS) and blood and urine
calcium phosphate levels were normal.</p>
        <p>Fig. 1 shows the same sentence as a constituent
tree4 which provides syntactic visualization of
scope resolution by different agents. The scope of
negation marked by BERT in French is generally
very consistent. It usually ends before a
punctuation mark, another cue, a finite verb, or the
conjunction et. It tends to trace the constituent
boundaries. In fact, it is unclear why, in this example,
human annotation did not include the abbreviation
of the syndrome inside the scope.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>Considering the linguistic complexity of negation
scope resolution, the F1-score of 84.73 on a
zeroshot transfer test is substantial. Fine-tuning on
minimal training data also provided decent results.
A deeper examination of the output could
provide us with further improvements. Potentially, a
standardized annotation scheme for the English
corpora could improve all outcomes. Aside from
negation, this study could be turned into a
behavioral probing task that further explores BERT’s
linguistic and cross-linguistic abilities. It would
be interesting to test typologically different
languages as well.</p>
      <p>4Created with Berkeley Neural Parser</p>
      <p>Louise Dele´ger and Cyril Grouin. 2012. Detecting
Negation of Medical Problems in French Clinical
Notes. In Proc of Int Health Inform, Miami Beach,
FL.</p>
      <p>Jacob Devlin. 2018. Multilingual BERT Readme
document. Library Catalog: github.com.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Amine</given-names>
            <surname>Abdaoui</surname>
          </string-name>
          , Andon Tchechmedjiev,
          <string-name>
            <given-names>William</given-names>
            <surname>Digan</surname>
          </string-name>
          , Sandra Bringay, and
          <string-name>
            <given-names>Clement</given-names>
            <surname>Jonquet</surname>
          </string-name>
          .
          <year>2017</year>
          . French ConText: De´
          <article-title>tecter la ne´gation, la temporalite´ et le sujet dans les textes cliniques Franc¸ais.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Jay</given-names>
            <surname>Alammar</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Kathryn</given-names>
            <surname>Baker</surname>
          </string-name>
          , Michael Bloodgood, Bonnie Dorr, Nathaniel W. Filardo, Lori Levin, and
          <string-name>
            <given-names>Christine</given-names>
            <surname>Piatko</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>A Modality Lexicon and its use in Automatic Tagging</article-title>
          .
          <source>In Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)</source>
          , Valletta,
          <string-name>
            <given-names>Malta. European</given-names>
            <surname>Language Resources Association (ELRA).</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Benita</given-names>
            <surname>Kathleen</surname>
          </string-name>
          Britto and
          <string-name>
            <given-names>Aditya</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Resolving the Scope of Speculation and Negation using Transformer-Based Architectures</article-title>
          . ArXiv, abs/
          <year>2001</year>
          .02885.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Wendy W. Chapman</surname>
            , Will Bridewell, Paul Hanbury, Gregory
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Cooper</surname>
          </string-name>
          , and
          <string-name>
            <surname>Bruce</surname>
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Buchanan</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>A Simple Algorithm for Identifying Negated Findings and Diseases in Discharge Summaries</article-title>
          .
          <source>Journal of Biomedical Informatics</source>
          ,
          <volume>34</volume>
          (
          <issue>5</issue>
          ):
          <fpage>301</fpage>
          -
          <lpage>310</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Wendy W. Chapman</surname>
          </string-name>
          , Dieter Hillert, Sumithra Velupillai, Maria Kvist, Maria Skeppstedt, Brian E. Chapman, Mike Conway, Melissa Tharp, Danielle L.
          <string-name>
            <surname>Mowery</surname>
          </string-name>
          , and Louise Dele´ger.
          <year>2013</year>
          .
          <article-title>Extending the NegEx Lexicon for Multiple Languages</article-title>
          .
          <article-title>Studies in health technology and informatics</article-title>
          ,
          <volume>192</volume>
          :
          <fpage>677</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Costumero</surname>
          </string-name>
          , Federico Lopez, Consuelo Gonzalo-Mart´ın, Marta Millan, and
          <string-name>
            <given-names>Ernestina</given-names>
            <surname>Menasalvas</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>An Approach to Detect Negation on Medical Documents in Spanish</article-title>
          .
          <source>In Brain Informatics and Health, Lecture Notes in Computer Science</source>
          , pages
          <fpage>366</fpage>
          -
          <lpage>375</lpage>
          , Cham. Springer International Publishing.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Viviana</given-names>
            <surname>Cotik</surname>
          </string-name>
          , Roland Roller, Feiyu Xu, Hans Uszkoreit, Klemens Budde, and
          <string-name>
            <given-names>Danilo</given-names>
            <surname>Schmidt</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Negation Detection in Clinical Reports Written in German</article-title>
          .
          <source>In Proceedings of the Fifth Workshop on Building and Evaluating Resources for Biomedical Text Mining (BioTxtM2016)</source>
          , pages
          <fpage>115</fpage>
          -
          <lpage>124</lpage>
          , Osaka, Japan.
          <source>The COLING 2016 Organizing Committee.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Isaac</given-names>
            <surname>Councill</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ryan McDonald</surname>
            ,
            <given-names>and Leonid</given-names>
          </string-name>
          <string-name>
            <surname>Velikovich</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>What's great and what's not: learning to classify the scope of negation for improved sentiment analysis</article-title>
          .
          <source>In Proceedings of the Workshop on Negation and Speculation in Natural Language Processing</source>
          , pages
          <fpage>51</fpage>
          -
          <lpage>59</lpage>
          , Uppsala, Sweden. University of Antwerp.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>O</given-names>
            <surname>¨ sten Dahl</surname>
          </string-name>
          .
          <year>1979</year>
          .
          <article-title>Typology of sentence negation</article-title>
          .
          <source>Linguistics</source>
          ,
          <volume>17</volume>
          (
          <issue>1-2</issue>
          ):
          <fpage>79</fpage>
          -
          <lpage>106</lpage>
          . Publisher: De Gruyter Mouton Section: Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <article-title>Cle´ment Dalloux, Vincent Claveau</article-title>
          , and
          <string-name>
            <given-names>Natalia</given-names>
            <surname>Grabar</surname>
          </string-name>
          .
          <year>2017</year>
          . De´tection de la ne´
          <article-title>gation : corpus franc¸ais et apprentissage supervise´</article-title>
          .
          <source>In SIIM 2017 - Symposium</source>
          sur l'Inge´nierie de l'Information Me´dicale, pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          , Toulouse, France.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Jacob</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ming-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Kenton</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Kristina</given-names>
            <surname>Toutanova</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          .
          <source>In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers), pages
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          , Minneapolis, Minnesota. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Hanan</given-names>
            <surname>Elazhary</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>NegMiner: An Automated Tool for Mining Negations from Electronic Narrative Medical Documents</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Federico</given-names>
            <surname>Fancellu</surname>
          </string-name>
          , Adam Lopez, and
          <string-name>
            <given-names>Bonnie</given-names>
            <surname>Webber</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Neural Networks For Negation Scope Detection</article-title>
          .
          <source>In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source>
          , pages
          <fpage>495</fpage>
          -
          <lpage>504</lpage>
          , Berlin, Germany. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Federico</given-names>
            <surname>Fancellu</surname>
          </string-name>
          , Adam Lopez,
          <string-name>
            <given-names>and Bonnie L.</given-names>
            <surname>Webber</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Neural Networks for Cross-lingual Negation Scope Detection</article-title>
          . ArXiv, abs/
          <year>1810</year>
          .02156.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Henk</given-names>
            <surname>Harkema</surname>
          </string-name>
          , John N. Dowling, Tyler Thornblade, and
          <string-name>
            <surname>Wendy</surname>
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Chapman</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Context: An Algorithm for Determining Negation, Experiencer, and Temporal Status from Clinical Reports</article-title>
          .
          <source>Journal of biomedical informatics</source>
          ,
          <volume>42</volume>
          (
          <issue>5</issue>
          ):
          <fpage>839</fpage>
          -
          <lpage>851</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>John</given-names>
            <surname>Hewitt</surname>
          </string-name>
          and
          <string-name>
            <given-names>Christopher D.</given-names>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>A Structural Probe for Finding Syntax in Word Representations</article-title>
          .
          <source>In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers), pages
          <fpage>4129</fpage>
          -
          <lpage>4138</lpage>
          , Minneapolis, Minnesota. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Laurence R.</given-names>
            <surname>Horn</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>A natural history of negation. The David Hume series</article-title>
          . CSLI, Stanford, Calif.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Salud</given-names>
            <surname>Mar</surname>
          </string-name>
          <article-title>´ıa Jime´nez-</article-title>
          <string-name>
            <surname>Zafra</surname>
            ,
            <given-names>M.</given-names>
            Teresa Mart´ınValdivia, M.
          </string-name>
          <string-name>
            <surname>Dolores</surname>
            Molina-Gonza´lez, and
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Alfonso</surname>
          </string-name>
          <article-title>Uren˜a-Lo´pez</article-title>
          .
          <year>2018</year>
          .
          <article-title>Relevance of the SFU ReviewSP-NEG corpus annotated with the scope of negation for supervised polarity classification in Spanish</article-title>
          .
          <source>Information Processing &amp; Management</source>
          ,
          <volume>54</volume>
          (
          <issue>2</issue>
          ):
          <fpage>240</fpage>
          -
          <lpage>251</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Salud</given-names>
            <surname>Mar</surname>
          </string-name>
          <article-title>´ıa Jime´nez-</article-title>
          <string-name>
            <surname>Zafra</surname>
          </string-name>
          ,
          <article-title>Roser Morante, Mar´ıa Teresa Mart´ın-</article-title>
          <string-name>
            <surname>Valdivia</surname>
            , and
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Alfonso</surname>
          </string-name>
          <article-title>Uren˜a-Lo´pez</article-title>
          .
          <year>2020</year>
          .
          <article-title>Corpora Annotated with Negation: An Overview</article-title>
          .
          <source>In Computational Linguistics</source>
          , volume
          <volume>0</volume>
          , pages
          <fpage>1</fpage>
          -
          <lpage>87</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>Aditya</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          and
          <string-name>
            <given-names>Suraj</given-names>
            <surname>Sawant</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>NegBERT: A Transfer Learning Approach for Negation Detection and Scope Resolution</article-title>
          . ArXiv, abs/
          <year>1911</year>
          .04211.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>Natalia</given-names>
            <surname>Konstantinova</surname>
          </string-name>
          ,
          <string-name>
            <surname>Sheila C.M. de Sousa</surname>
            , Noa P. Cruz,
            <given-names>Manuel J</given-names>
          </string-name>
          .
          <source>Man˜a</source>
          ,
          <string-name>
            <surname>Maite Taboada</surname>
            , and
            <given-names>Ruslan</given-names>
          </string-name>
          <string-name>
            <surname>Mitkov</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>A review corpus annotated for negation, speculation and their scope</article-title>
          .
          <source>In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC'12)</source>
          , pages
          <fpage>3190</fpage>
          -
          <lpage>3195</lpage>
          , Istanbul, Turkey.
          <source>European Language Resources Association (ELRA).</source>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>Qianchu</given-names>
            <surname>Liu</surname>
          </string-name>
          , Federico Fancellu, and
          <string-name>
            <given-names>Bonnie</given-names>
            <surname>Webber</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>NegPar: A parallel corpus annotated for negation</article-title>
          .
          <source>In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC</source>
          <year>2018</year>
          ), Miyazaki,
          <string-name>
            <given-names>Japan. European</given-names>
            <surname>Language Resources Association (ELRA).</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Marie-Catherine de Marneffe</surname>
            , Bill MacCartney, Trond Grenager, Daniel Cer, Anna Rafferty, and
            <given-names>Christopher D</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Learning to distinguish valid textual entailments</article-title>
          .
          <source>page 6.</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <given-names>Roser</given-names>
            <surname>Morante</surname>
          </string-name>
          and
          <string-name>
            <given-names>Eduardo</given-names>
            <surname>Blanco</surname>
          </string-name>
          .
          <year>2012</year>
          . *
          <article-title>SEM 2012 Shared Task: Resolving the Scope and Focus of Negation</article-title>
          .
          <source>In *SEM 2012: The First Joint Conference on Lexical and Computational Semantics - Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval</source>
          <year>2012</year>
          ), pages
          <fpage>265</fpage>
          -
          <lpage>274</lpage>
          , Montre´al, Canada. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Roser</given-names>
            <surname>Morante</surname>
          </string-name>
          and
          <string-name>
            <given-names>Walter</given-names>
            <surname>Daelemans</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>ConanDoyle-neg: Annotation of negation in Conan Doyle stories</article-title>
          .
          <source>page 6.</source>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <given-names>Roser</given-names>
            <surname>Morante</surname>
          </string-name>
          , Anthony Liekens, and
          <string-name>
            <given-names>Walter</given-names>
            <surname>Daelemans</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Learning the Scope of Negation in Biomedical Texts</article-title>
          .
          <source>In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing</source>
          , pages
          <fpage>715</fpage>
          -
          <lpage>724</lpage>
          , Honolulu, Hawaii. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <given-names>Pradeep</given-names>
            <surname>Mutalik</surname>
          </string-name>
          ,
          <string-name>
            <surname>Aniruddha M. Deshpande</surname>
          </string-name>
          , and
          <string-name>
            <surname>Prakash</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Nadkarni</surname>
          </string-name>
          .
          <year>2001</year>
          . Research Paper:
          <article-title>Use of General-purpose Negation Detection to Augment Concept Indexing of Medical Documents: A Quantitative Study Using the UMLS</article-title>
          . JAMIA.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>Woodley</given-names>
            <surname>Packard</surname>
          </string-name>
          ,
          <string-name>
            <surname>Emily M. Bender</surname>
            , Jonathon Read, Stephan Oepen, and
            <given-names>Rebecca</given-names>
          </string-name>
          <string-name>
            <surname>Dridan</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Simple Negation Scope Resolution through Deep Parsing: A Semantic Solution to a Semantic Problem</article-title>
          .
          <source>In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source>
          , pages
          <fpage>69</fpage>
          -
          <lpage>78</lpage>
          , Baltimore, Maryland.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <given-names>Telmo</given-names>
            <surname>Pires</surname>
          </string-name>
          , Eva Schlinger, and
          <string-name>
            <given-names>Dan</given-names>
            <surname>Garrette</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>How multilingual is Multilingual BERT? In ACL</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <given-names>Anna</given-names>
            <surname>Rogers</surname>
          </string-name>
          , Olga Kovaleva, and
          <string-name>
            <given-names>Anna</given-names>
            <surname>Rumshisky</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>A Primer in BERTology: What we know about how BERT works</article-title>
          . arXiv:
          <year>2002</year>
          .12327 [cs].
          <source>ArXiv</source>
          :
          <year>2002</year>
          .12327.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <given-names>Elena</given-names>
            <surname>Sergeeva</surname>
          </string-name>
          , Henghui Zhu,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Prinsen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Amir</given-names>
            <surname>Tahmasebi</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Negation Scope Detection in Clinical Notes and Scientific Abstracts: A Featureenriched LSTM-based Approach</article-title>
          .
          <source>AMIA Summits on Translational Science Proceedings</source>
          ,
          <year>2019</year>
          :
          <fpage>212</fpage>
          -
          <lpage>221</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <given-names>Maria</given-names>
            <surname>Skeppstedt</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Negation Detection in Swedish Clinical Text</article-title>
          .
          <source>In Proceedings of the NAACL HLT 2010 Second Louhi Workshop on Text and Data Mining of Health Documents, Louhi '10</source>
          , pages
          <fpage>15</fpage>
          -
          <lpage>21</lpage>
          , USA.
          <article-title>Association for Computational Linguistics</article-title>
          . Event-place: Los Angeles, California.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <article-title>Gyo¨rgy Szarvas, Veronika Vincze</article-title>
          , Richa´rd Farkas, and Ja´nos Csirik.
          <year>2008</year>
          .
          <article-title>The BioScope corpus: annotation for negation, uncertainty and their scope in biomedical texts</article-title>
          .
          <source>In Proceedings of the Workshop on Current Trends in Biomedical Natural Language Processing</source>
          , pages
          <fpage>38</fpage>
          -
          <lpage>45</lpage>
          , Columbus, Ohio. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <given-names>Veronika</given-names>
            <surname>Vincze</surname>
          </string-name>
          , Gyo¨rgy Szarvas, Richa´rd Farkas,
          <source>Gyo¨rgy Mo´ra, and Ja´nos Csirik</source>
          .
          <year>2008</year>
          .
          <article-title>The BioScope corpus: biomedical texts annotated for uncertainty, negation and their scopes</article-title>
          .
          <source>BMC Bioinformatics</source>
          ,
          <volume>9</volume>
          (
          <issue>S11</issue>
          ):
          <fpage>S9</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <given-names>Alex</given-names>
            <surname>Warstadt</surname>
          </string-name>
          , Yu Cao, Ioana Grosu, Wei Peng, Hagen Blix, Yining Nie, Anna Alsop, Shikha Bordia, Haokun Liu, Alicia Parrish,
          <string-name>
            <surname>Sheng-Fu</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Jason Phang, Anhad Mohananey, Phu Mon Htut, Paloma Jeretic, and
          <string-name>
            <surname>Samuel</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Bowman</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Investigating BERT's Knowledge of Language: Five Analysis Methods with NPIs</article-title>
          .
          <source>In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          , pages
          <fpage>2877</fpage>
          -
          <lpage>2887</lpage>
          ,
          <string-name>
            <surname>Hong</surname>
            <given-names>Kong</given-names>
          </string-name>
          , China. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <given-names>Stephen</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Timothy</given-names>
            <surname>Miller</surname>
          </string-name>
          , James Masanz,
          <string-name>
            <given-names>Matt</given-names>
            <surname>Coarr</surname>
          </string-name>
          , Scott Halgrim, David Carrell,
          <string-name>
            <given-names>and Cheryl</given-names>
            <surname>Clark</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Negation's Not Solved: Generalizability Versus Optimizability in Clinical Natural Language Processing</article-title>
          .
          <source>PLOS ONE</source>
          ,
          <volume>9</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>