<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Negation Detection for Robust Adverse Drug Event Extraction From Social Media Texts*</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Simone Scaboro</string-name>
          <email>scaboro.simone@spes.uniud.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Beatrice Portelli</string-name>
          <email>portelli.beatrice@spes.uniud.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emmanuele Chersoni</string-name>
          <email>emmanuele.chersoni@polyu.edu.hk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Santus</string-name>
          <email>esantus@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giuseppe Serra</string-name>
          <email>giuseppe.serra@uniud.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CSAIL MIT</institution>
          ,
          <addr-line>Cambridge (MA)</addr-line>
          ,
          <country country="US">United States of America</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The Hong Kong Polytechnic University</institution>
          ,
          <country country="HK">Hong Kong</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Udine</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Adverse Drug Event (ADE) extraction from user-generated content has gained popularity as a tool to aid researchers and pharmaceutical companies to monitor side efect of drugs in the wild. Automatic models can rapidly examine large collections of social media texts. However it is currently unknown if such models are robust in face of linguistic phenomena such as negation and speculation, which are pervasive across language varieties. We evaluate three state-of-the-art systems, showing their fragility against negation, and then we introduce two possible strategies to increase the robustness of these models: (i) a pipeline approach, using a specific component for negation detection; (ii) an augmentation of the dataset with artificially negated samples to further train the models. We show that both strategies bring significant increases in performance.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Bio-medical data</kwd>
        <kwd>Social media</kwd>
        <kwd>Annotated corpora creation</kwd>
        <kwd>Negation detection</kwd>
        <kwd>Adverse drug events</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        As more users keep reporting their personal experience with drugs on social media, blogs and
health forums, automatic Adverse Drug Event (ADE) detection in social media texts is becoming
a fundamental tool in the field of pharmacovigilance [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. It is common for Internet users
to report their personal experiences with drugs on forums and microblogging platforms, but
also messaging pharmaceutical companies directly on social media, via chatbots or emails.
This is why both researchers and the industry are looking for ways to make use of this great
amount of unprocessed and potentially informative data. User-generated texts, and social media
texts in particular, are inherently noisy (containing colloquial language, slang and metaphors,
non-standard syntactic constructions etc.) and require specialized data cleaning and handling
techniques. The task becomes even more complicated if the final objective is to map them to a
      </p>
      <p>Tweet collection</p>
      <p>Collection and extraction modules</p>
      <p>Classification module</p>
      <p>Contains ADE?</p>
      <p>No
Discard</p>
      <p>Yes</p>
      <p>Extraction module
Which are the ADE?
headache
arm pain
clot
formal medical dictionary or ontology.</p>
      <p>
        In the last decade, the Natural Language Processing (NLP) community dedicated a consistent
efort in developing robust methods for mining biomedical information from user-generated
texts, also leading to the creation of several dedicated shared tasks series on ADE detection
(SMM4H – Social Media Mining for Health) [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7 ref8">4, 5, 6, 7, 8</xref>
        ]. Although these models have seen great
advancements in the last years, it is still unknown how robust they are in face of some pervasive
linguistic phenomena such as negation and speculation. However, general investigations on
machine comprehension and question answering tasks confirmed that such phenomena often
pose a serious challenge [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The distinction between certain, hypothesized and negated and
speculated events is of key importance in biomedical NLP tasks [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. In the same way, it
is essential to know whether the causal link between a drug and an ADE is being stated or
negated in pharmacovigilance.
      </p>
      <p>
        Detecting the scope of negation and speculation has been object of NLP research for at least
one decade, via both rule-based and machine learning approaches. An early, popular system
was introduced by Chapman et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], whose NegEx algorithm exploited regular expressions
to identify negations in clinical documents in English. The latest advancements are represented
by BERT-based models [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ], also with the aid of multitask learning architectures [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>As of today, the research in biomedical NLP mostly focused on scope detection of negations
and speculations per se and on more formal types of texts (e.g. clinical notes, articles). Given
the growing demand to process and analyze large collections of user-generated content from
social media, we choose to focus on ADE detection on Twitter posts. They are characterized by
a noisier and more informal writing style. The goal is to enable to systems to be more successful
at distinguishing between factual and non-factual information.</p>
      <p>In this paper, introduce an extended dataset to analyze the performance of ADE extraction
in presence of asserted and negated Adverse Events. We show that the latest state-of-the-art
ADE detection systems cannot recognize and handle negations correctly and introduce two
strategies to increase the robustness of existing systems: (i) adding a negation detection module
in a pipeline fashion to exclude the negated ADEs predicted by the models; (ii) augmenting the
training set with artificially negated samples.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Proposed Strategies</title>
      <p>
        Following the latest advancements in the SMM4H Shared Tasks, we choose three
Transformerbased models that showed high performance on the ADE extraction dataset of SMM4H [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ],
... arm pain ... ... no clot ...
      </p>
      <p>... no headache ...</p>
      <p>Collection and
extraction modules
headache
arm pain clot</p>
      <p>Negation detection
module</p>
      <p>Artificial
samples
Remove intersections</p>
      <p>no headache no clot
arm pain
... arm pain ... ... no clot ...</p>
      <p>... no headache ...</p>
      <p>Extraction module</p>
      <p>
        Collection and
extraction modules
arm pain
and are currently at the top of the corresponding leaderboard: BERT [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], SpanBERT [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and
PubMedBERT [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. The models are fine-tuned for token classification, predicting an IOB label
for each token in the sentence to detect the boundaries of ADE mentions.
      </p>
      <p>We analyze two possible strategies to increase the robustness of the baseline models: (i) adding
a negation (or speculation) detection module in a pipeline fashion to exclude some incorrect
adverse events predicted by the models; (ii) augmenting the training set with artificially created
samples. Figure 2 illustrates the two approaches.</p>
      <sec id="sec-2-1">
        <title>2.1. Specialized negation detection modules</title>
        <p>We propose a simple pipeline to enhance the robustness of the base models against negation by
combining them with a negation detection module. Let us consider a text , a ADE extraction
base model ℬ and a negation detection module  . Given , ℬ outputs a set of substrings of  that
are labeled as ADE mentions: ℬ() = {1, . . . , }. Similarly,  takes a text and outputs a set
of substrings, which are considered to be entities within a negation scope:  () = {1, . . . , }.</p>
        <p>
          A combined pipeline model is obtained by discarding all ADE spans  ∈ ℬ() that overlap
one of the negation spans  ∈  (): ℬ () = { ∈ ℬ() | ∀( ∈  () ∧  ∩  = ∅)}
Modules used We introduce two negation detection modules: NegEx, a Python implementation
[
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] of the NegEx algorithm, based on simple regular expressions, which evaluates whether
named entities are negated; BERTneg, a BERT model (bert-base-uncased) that we finetuned
for token classification. We trained BERTneg on BioScope [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ], which contains medical texts
annotated for the presence of negation and speculation cues and their related scopes. We selected
3190 sentences (2801 with a negation scope) and finetuned the model for scope detection (10
epochs, learning rate 1 − 4).
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Data Augmentation</title>
        <p>
          While there are several datasets for ADE detection on social media texts [
          <xref ref-type="bibr" rid="ref23 ref24">23, 24</xref>
          ], the largest
collection is the one released yearly for the SMM4H Workshop and Shared Task.
        </p>
        <p>However, most datasets are made of samples that either do or do not contain an ADE (useful
to train the Classification module in Figure 1). Because of this, they include a small number of
negated ADEs by construction: no particular attention is given to these samples when curating
the data and, even when they are present, they are labelled as noADE samples. This makes it
harder to study this phenomenon.</p>
        <p>
          We augment the SMM4H19 dataset (the training set for the ADE extraction Task of SMM4H19
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]) in two ways: (i) recovery of real samples; (ii) generating negated versions of real samples.
Both activities were carried out by four volunteer annotators with a high level of proficiency in
English.
        </p>
        <p>
          Recovery of real samples We look for real samples that negate the presence of an ADE using
SMM4H19 and SMM4H20 , the datasets for the binary classification tasks in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] and [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. These
are meant to be used as test samples, to check the robustness of the model.
        </p>
        <p>Generation of negated samples We manually create negated versions for the ADE tweets
in the test split of SMM4H19. These are meant to be used as additional training samples, to
teach the model how to distinguish asserted and negated adverse events. The result of this
procedure is a new set of tweets denying the presence of an ADE. As an example, the original
tweet “fluoxetine, got me going crazy” was transformed into “fluoxetine, didn’t get me going
crazy”.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Data Partitioning</title>
      <p>We split the available data in a train and a test set, both containing the three categories of tweets:
ADE, noADE and negADE. Given the small amount of real negADE tweets, we use all of them in
the test set to evaluate the performance only on real tweets. Conversely, the training set only
contains the manually generated negADE samples.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>
        All the reported results are the average over 5 runs. For the Transformer models we used the
same hyperparameters reported by Portelli et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. As metrics, we consider the number of
false positive predictions (FP) and the relaxed precision (P), recall (R) and F1 score as defined
in the SMM4H shared tasks [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]: the scores take into account “partial” matches, in which it is
suficient for a prediction to partially overlap with the gold annotation. We report the number
of FP both on the whole test set and on individual partitions (ADE, noADE and negADE samples).
For brevity, here we report the results for just one of the baseline models (PubMedBERT, Table
1). Results for the other baseline models behave similarly and can be found in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>As a preliminary step for all experiments, the two negation detection models are trained and
used to predict the negation scopes for all the test samples once. This allows us to compute the
predictions of any pipeline model.</p>
      <p>Exp 0 (row 1) To provide a measure of the initial robustness of the base models and their
general performance, we train them on the ADE and noADE samples only. The base models have
a high number of FP, especially in the negADE category. This strongly suggests that they are
not robust against this phenomenon.</p>
      <p>Exp 1 (rows 2–3) We test the eficacy of the pipeline negation detection method, applying
NegEx and BERTneg to the base models. When combined with NegEx (row 2), the FP decreases
by almost 50 points, showing that the regular expression module removes a great number of
unwanted predictions. BERTneg decreases the number of FP too, but only by 38 points, being
less aggressive than NegEx. However, if we look at P and R in the first three rows, we see
that the negation detection modules increase P at the cost of large drops in R: some correct
predictions of the base models get discarded (i.e., ADEs that contain a negation such as “After
taking this drug I cannot sleep anymore”).</p>
      <p>Exp 2 (row 4) We add to the training set all negADE generated samples and train the base
models on them to test the efect of augmenting the dataset. This lowers the number of FP
predictions for all models as much as using NegEx (compare row 2 and 4), especially on the
negADE set. We still observe a drop in R, but less severe than in Exp 1 (less true positives are
being discarded). The increase in P is also more noticeable, leading to an overall increase in F1.
Exp 3 (rows 5–6) To investigate whether the two methods are complementary in their action,
we combine the two strategies, applying the pipeline architecture to the models trained on the
augmented dataset. They are in some way complementary, as shown by the further decrease in
FP in all categories. However, combining the two approaches might not be the best strategy, as
it leads to a further decrease in R.</p>
      <p>Observations The results show that introducing a small number of new samples (even if
artificial) is the best way to directly increase the model knowledge about the phenomenon.
However, this solution could be expensive in absence of annotated data. For this reason, the
pipeline models might be a viable alternative, as they maintain the F1 score while still decreasing
the number of FP.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this paper, we evaluate the impact of negations on state-of-the-art ADE detection models.
We introduce and compare two strategies to tackle the problem: using a negation detection
module and adding negSamp samples in the training set. Both of them bring significant increases
in performance. Future work should focus on more refined techniques to accurately model
the semantic properties of the samples, also by jointly handling negation and speculation
phenomena. This might be an essential requirement for dealing with the noisiness and variety
of social media texts.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Scaboro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Portelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Chersoni</surname>
          </string-name>
          , E. Santus,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Serra, NADE: A benchmark for robust adverse drug events extraction in face of negations</article-title>
          , in: Proceedings of the Seventh Workshop on Noisy User-generated
          <string-name>
            <surname>Text</surname>
          </string-name>
          (
          <article-title>W-NUT</article-title>
          <year>2021</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>230</fpage>
          -
          <lpage>237</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .wnut-
          <volume>1</volume>
          .
          <fpage>26</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Metke-Jimenez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gaire</surname>
          </string-name>
          , C. Paris, Text and
          <article-title>Data Mining Techniques in Adverse Drug Reaction Detection, ACM Computing Surveys (CSUR) 47 (</article-title>
          <year>2015</year>
          )
          <fpage>1</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          , G. Gonzalez,
          <article-title>Portable Automatic Text Classification for Adverse Drug Reaction Detection via Multi-corpus Training</article-title>
          ,
          <source>Journal of Biomedical Informatics</source>
          <volume>53</volume>
          (
          <year>2015</year>
          )
          <fpage>196</fpage>
          -
          <lpage>207</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Paul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Brownstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikfarjam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Scotch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Gonzalez, Social Media Mining for Public Health Monitoring and Surveillance</article-title>
          ,
          <source>in: Biocomputing</source>
          <year>2016</year>
          ,
          <year>2016</year>
          , pp.
          <fpage>468</fpage>
          -
          <lpage>479</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Gonzalez-Hernandez, Overview of the Second Social Media Mining for Health (SMM4H) Shared Tasks at AMIA 2017, Training 1 (</article-title>
          <year>2017</year>
          )
          <fpage>1239</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Weissenbacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Paul</surname>
          </string-name>
          , G. Gonzalez,
          <article-title>Overview of the Social Media Mining for Health (SMM4H) Shared Tasks at EMNLP 2018</article-title>
          ,
          <source>in: Proceedings of the EMNLP Workshop on Social Media Mining for Health Applications</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Weissenbacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Magge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Daughton</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. O'Connor</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Paul</surname>
          </string-name>
          , G. Gonzalez,
          <article-title>Overview of the Fourth Social Media Mining for Health (SMM4H) Shared Tasks at ACL 2019, in: Proceedings of the ACL Social Media Mining for Health Applications Workshop</article-title>
          &amp; Shared Task,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Alimova</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Flores</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Magge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Miftahutdinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.-L.</given-names>
            <surname>Minard</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>O'connor,</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tutubalina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Weissenbacher</surname>
          </string-name>
          , et al.,
          <source>Overview of the Fifth Social Media Mining for Health Applications Shared Tasks at Coling</source>
          <year>2020</year>
          ,
          <source>in: Proceedings of the COLING Workshop on Social Media Mining for Health Applications</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          , Beyond Accuracy:
          <article-title>Behavioral Testing of NLP Models with CheckList</article-title>
          ,
          <source>in: Proceedings of ACL</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E.</given-names>
            <surname>Velldal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Øvrelid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Read</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Oepen</surname>
          </string-name>
          , Speculation and Negation: Rules, Rankers, and the Role of Syntax,
          <source>Computational Linguistics</source>
          <volume>38</volume>
          (
          <year>2012</year>
          )
          <fpage>369</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N. P.</given-names>
            <surname>Cruz</surname>
          </string-name>
          <string-name>
            <surname>Díaz</surname>
          </string-name>
          ,
          <source>Detecting Negated and Uncertain Information in Biomedical and Review Texts, in: Proceedings of the RANLP Student Research Workshop</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>W. W.</given-names>
            <surname>Chapman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Bridewell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hanbury</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. F.</given-names>
            <surname>Cooper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. G.</given-names>
            <surname>Buchanan</surname>
          </string-name>
          ,
          <article-title>A Simple Algorithm for Identifying Negated Findings and Diseases in Discharge Summaries</article-title>
          ,
          <source>Journal of Biomedical Informatics</source>
          <volume>34</volume>
          (
          <year>2001</year>
          )
          <fpage>301</fpage>
          -
          <lpage>310</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sawant</surname>
          </string-name>
          ,
          <article-title>Negbert: A Transfer Learning Approach for Negation Detection and Scope Resolution</article-title>
          ,
          <source>in: Proceedings of LREC</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B. K.</given-names>
            <surname>Britto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <article-title>Resolving the Scope of Speculation and Negation using Transformer-Based Architectures</article-title>
          , arXiv preprint arXiv:
          <year>2001</year>
          .
          <volume>02885</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. K.</given-names>
            <surname>Britto</surname>
          </string-name>
          ,
          <article-title>Multitask Learning of Negation and Speculation using Transformers</article-title>
          ,
          <source>in: Proceedings of the EMNLP International Workshop on Health Text Mining and Information Analysis</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>B.</given-names>
            <surname>Portelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lenzi</surname>
          </string-name>
          , E. Chersoni,
          <string-name>
            <given-names>G.</given-names>
            <surname>Serra</surname>
          </string-name>
          , E. Santus,
          <string-name>
            <surname>BERT</surname>
          </string-name>
          <article-title>Prescriptions to Avoid Unwanted Headaches: A Comparison of Transformer Architectures for Adverse Drug Event Detection</article-title>
          ,
          <source>in: Proceedings of EACL</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>B.</given-names>
            <surname>Portelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Passabì</surname>
          </string-name>
          , E. Lenzi,
          <string-name>
            <given-names>G.</given-names>
            <surname>Serra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Santus</surname>
          </string-name>
          , E. Chersoni,
          <article-title>Improving Adverse Drug Event Extraction with SpanBERT on Diferent Text Typologies</article-title>
          ,
          <source>arXiv preprint arXiv:2105.08882</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          ,
          <source>in: Proceedings of NAACL</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Weld</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <surname>O. Levy,</surname>
          </string-name>
          <article-title>SpanBERT: Improving Pre-training by Representing and Predicting Spans, Transactions of the Association for Computational Linguistics 8 (</article-title>
          <year>2019</year>
          )
          <fpage>64</fpage>
          -
          <lpage>77</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tinn</surname>
          </string-name>
          , H. Cheng, M. Lucas,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usuyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Poon</surname>
          </string-name>
          ,
          <article-title>Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing</article-title>
          , arXiv preprint arXiv:
          <year>2007</year>
          .
          <volume>15779</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pizarro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Reteig</surname>
          </string-name>
          , L. Murray, jenojp/negspacy: Minor Bug Fix,
          <source>Improve Chunk Prefix Functionality (Version v0.1.9)</source>
          .,
          <year>2020</year>
          . URL: https://doi.org/10.5281/zenodo.3702544. doi:
          <volume>10</volume>
          . 5281/zenodo.3702544.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vincze</surname>
          </string-name>
          , G. Szarvas,
          <string-name>
            <given-names>R.</given-names>
            <surname>Farkas</surname>
          </string-name>
          , G. Móra,
          <string-name>
            <given-names>J.</given-names>
            <surname>Csirik</surname>
          </string-name>
          , The BioScope Corpus:
          <article-title>Biomedical Texts Annotated for Uncertainty, Negation and their Scopes, BMC Bioinformatics 9 (</article-title>
          <year>2008</year>
          )
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Metke-Jimenez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kemp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Cadec: A Corpus of Adverse Drug Event Annotations</article-title>
          ,
          <source>Journal of Biomedical Informatics</source>
          <volume>55</volume>
          (
          <year>2015</year>
          )
          <fpage>73</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>N.</given-names>
            <surname>Alvaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Miyao</surname>
          </string-name>
          , N. Collier, TwiMed: Twitter and PubMed Comparable Corpus of Drugs, Diseases, Symptoms, and Their Relations,
          <source>JMIR Public Health Surveillance</source>
          <volume>3</volume>
          (
          <year>2017</year>
          )
          <article-title>e24</article-title>
          . URL: https://doi.org/10.2196/publichealth.6396. doi:
          <volume>10</volume>
          .2196/publichealth.6396.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>