<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Valentino Giudice[</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Aspie96 at FACT (IberLEF 2019): Factuality Classi cation in Spanish Texts with Character-Level Convolutional RNN and Tokenization</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Turin</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>0000</year>
      </pub-date>
      <volume>0002</volume>
      <fpage>119</fpage>
      <lpage>125</lpage>
      <abstract>
        <p>Being able to determine the factual status of events described in a text is crucial to analyze them. This report describes the system used by the Aspie96 team in the FACT shared task (part of IberLEF 2019) for factuality analysis and classi cation in texts in Spanish.</p>
      </abstract>
      <kwd-group>
        <kwd>fact</kwd>
        <kwd>factuality</kwd>
        <kwd>neural network</kwd>
        <kwd>natural language processing</kwd>
        <kwd>Spanish</kwd>
        <kwd>FACT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        The factuality of an event, as described in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], expresses its factual status: it
conveys whether it is characterized as corresponding to a fact, a possibility, or
a situation that doesn't actually hold.
      </p>
      <p>FACT (Factuality Analysis and Classi cation Task), a shared task organized
within IberLEF 2019 (Iberian Languages Evaluation Forum), aimed at the
creation of systems able to automatically label events in a text according to their
factuality.</p>
      <p>In each text, words representing events were marked and given one of three
labels:</p>
      <sec id="sec-1-1">
        <title>F (Fact) Situations presented as real by the author. CF (Counterfact) Situations presented as non real by the author. U (Unde ned) Situations presented as uncertain by the author as they had not yet happened or the author was unaware of their truth value.</title>
        <p>Thus, facts were not veri ed in accordance to the real word, just assessed
accordingly to how they had been presented by the author.</p>
        <p>In the training datasets, for each text, words representing events were
highlighted and classi ed according to their factuality.</p>
        <p>In the testing dataset, the factuality labels of events were not provided, but
words representing events were already highlighted: thus, the task was only to
label them correctly and identifying them was not necessary.</p>
        <p>The results of the task were measured using the macro-average F1-score.</p>
        <p>The competition was run using the CodaLab platform 1. Each team was
allowed a total maximum of 10 submissions. Each team could decide, at any
moment, which one submission to include in the leaderboard, which was always
visible to all participants and updated in real time.</p>
        <p>The Aspie96 team took part in the task, using a neural network based on
character-level features, adapted to classify words within the text.</p>
        <p>The structure of the model and its results are described in the following
sections.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Description of the System</title>
      <p>The system used by the Aspie96 team is a neural network that strictly uses
only the data provided for the task, without any additional information (such
as pretrained word embeddings).</p>
      <p>
        It is based on the system presented in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and on its adaptation presented in
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        The system presented in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] had, as a purpose, the (binary) classi cation of
tweets, making use of a character-level representation of them. It was the system
presented by the Aspie96 team at the IronITA 2018 task, described in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], for
irony detection in tweets in Italian.
      </p>
      <p>
        The system presented in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] by the Aspie96 team at the HAHA task described
in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] for humor detection in tweets in Spanish slightly modi es it, mainly to
adapt it to the language. It constitutes the basis of the system used in the
FACT task, thus it is crucial to understand it rst.
      </p>
      <p>The tweet classi cation system is represented in Figure 1. It begins with a
series of unidimensional convolutional layers followed by a bidirectional
recurrent layer. The output of the bidirectional layer, which is an individual dense
vector representing information about the whole tweet, is the input of a simple
fully connected layer, with one output, whose activation function is the logistic
function.</p>
      <p>
        The input is represented as a list with xed length (leading to padding or
truncation, where needed) of sparse vectors. Each vector of the list represents
an individual character of the tweet and contains ags whose values are either
0 or 1. A more in-depth description of the input representation can be found in
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        By removing the last layer from the tweet classifying neural network shown
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], the resulting neural network would return, for any text given as
input, a xed-length vector representation. For the sake of simplicity, from this
point, the neural network obtained in this way will be referred to as networkA
and is the backbone of the system used in the FACT task.
1 https://competitions.codalab.org/
      </p>
      <p>The FACT task was quite di erent from tweets classi cation: the given texts
were much longer and they were not to be classi ed as a whole: instead,
individual words within the text were to be classi ed.</p>
      <p>This required the creation of a neural network capable of processing the
individual words within the text as separate tokens. Because the whole premise
of the presented work is using strictly the data provided for the task, word
embeddings are not a solution.</p>
      <p>networkA, given a character-level representation of a text, as input, outputs
an individual vector representing the given text. It constitutes the basis for the
system used by the Aspie96 team in the FACT task.</p>
      <p>In each text, all words are detected: they are considered to be tokens. Each
word is represented as a xed-size list of vectors, each of which representing an
individual character. The word being represented is centered in its representation
and, because of the length of the representation of each word being xed, the
neighbouring characters, on the left and on the right, whether belonging to
other words or not (such as in the case of spaces or other special characters) are
used as padding. In the vector representation of each character one more ag is
added, the token ag : its value is 1 if the character belongs to the word being
represented, speci cally, and 0 otherwise.</p>
      <sec id="sec-2-1">
        <title>As an example, let us consider the following sentence:</title>
        <p>Lorem ipsum dolor sit amet, consectetur
adipisci elit, sed do eiusmod tempor
incidunt ut labore et dolore magna
aliqua.</p>
      </sec>
      <sec id="sec-2-2">
        <title>And let's assume the length of the representation of each word to be 14.</title>
        <p>The 5th word (amet) has a lenght of 4 and will
therefore need 14 4 = 10 characters of padding: 10=2 = 5
on the left and 10=2 = 5 on the right.</p>
        <p>It will, therefore, be represented as:</p>
        <p>sit amet, co</p>
      </sec>
      <sec id="sec-2-3">
        <title>The underlining means that the token ag for the</title>
        <p>marked characters has value 1.</p>
        <p>The whole text has 19 words and will thus be
represented as 19 lists of 14 vectors: each list representing
a word (and its neighbouring characters) and each
vector representing an individual character.</p>
        <p>Because the representations of each token already encodes the neighboring
characters as well there is no need to consider anything other than a word as a
token.</p>
        <p>This representation is still a character based representation, but the described
tokenization allows a neural network using this representation to recognize
individual words within the text and process individual words separately.</p>
        <p>The neural network used for the FACT tasks uses networkA to convert the
representation of each word (which is a sparse matrix of xed size) into an
individual xed-size vector (networkA convolves trough the representation of
the text, considering each word one by one).</p>
        <p>Thus, the representation of each individual word (which includes the
neighbouring characters) is fed trough networkA separately, obtaining a vector
representation of each word: this produces a representation of the text in which
each word is encoded into an individual vector. Then, to each of such
wordrepresenting vectors one entry is added: the event ag, indicating whether the
word is an event or not (such information is included in both the training dataset
and the testing dataset for every word).</p>
        <p>The representation obtained in this way is the input to the following layer of
the neural network: a recurrent layer. All outputs of the recurrent layer, each of
which being a vector, are considered (as many as the words in the text), not just
the last one. A dense layer is then applied to get, for each word, its classi cation
in one of the three classes. The classi cation is ignored for words that do not
represent events.</p>
        <p>
          The full network presented by the Aspie96 team in the FACT task is shown
in Figure 2. The purpose of the recurrent layer is to read the text, in a
humanlike fashion, encoding, in each word, its meaning, according to the previous ones.
Thanks to the usage of the neighbouring characters of each word as padding,
there is no need to use additional data to represent punctuation. Also, the same
word will be represented in di erent ways depending on its surroundings: this
ensures a more unambiguous representation of the meaning of each word (a word
may have di erent meanings depending on its surroundings, also it is easier to
infer the meaning of a word if its context is provided, given the fact that the
meaning of each word must be inferred for character-level features only), also
including information about the following words (the padding is big enough to
allow this. Note that humans don't usually need to read much ahead to fully
understand the meaning of a word). It must be noted that the structure of
networkA has been slightly adapted from task to task and is not identical to the
neural network used in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
        <p>The Aspie96 team ranked 2nd, with a macro-average F1-score of 0.554, just
below team premjithb, with a macro-average F1-score of 0.561. The 3rd ranking
team was jimblair, with a macro-average F1-score of 0.489. The accuracy of the
system presented by the Aspie96 team was 0.635.</p>
        <p>A baseline system was provided by the task organizers. The baseline system
(aiala) assigned labels randomly with a 0.7 probability for the F class, a 0.1
probability for the CF class and a 0.2 probability for the U class.</p>
        <p>The results of all teams are showed in Table 1.
This paper presented the system used by the Aspie96 team in the FACT task
for factuality classi cation of events within a text. The system produced good
results, with a macro-average F1-score very close to that of the rst ranking
team.</p>
        <p>
          The presented system, based upon adaptations of the one originally presented
in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], is a character-level convolutional recurrent neural network which makes no
use of pretrained features (such as word embeddings), nor of additional
knowledge or intuition about the task, but takes advantage of tokenization to classify
individual words within the text.
        </p>
        <p>The features of the neural network are meant to make it as general as possible:
it should be possible to use the system to, in general, classify words within a text,
regardless of the speci c high-level task (thus, regardless of it being factuality
analysis or not).</p>
        <p>This result has not been reached yet: despite the system proving itself to be
able to achieve good results in factual analysis and classi cation, much work is
still ahead to make it more general, as much worse results have been obtained
for other tasks.</p>
        <p>As for the structure of networkA, it had been slightly tweaked between
different tasks (not always out of necessity, but resulting in several slightly di erent
versions).</p>
        <p>Thus, further research is needed to create an individual, more stable,
tweetclassifying neural network, usable for di erent tasks (by changing only the
number of outputs according to the number of classes) and, based on that, an
individual system for classi cation of individual words within a text, like in FACT.</p>
        <p>Time will reveal where the limits of such an approach lay. The results
obtained in FACT, considering the structure of the network having nothing to do
with event classi cation speci cally, are quite promising in this direction: more
work is needed to allow convergence towards an individual system.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etcheverry</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prada</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : Overview of HAHA at IberLEF 2019:
          <article-title>Humor Analysis based on Human Annotation</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          , CEUR-WS, Bilbao,
          <source>Spain (9</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Cignarella</surname>
            ,
            <given-names>A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frenda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Basile</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bosco</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patti</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , et al.:
          <article-title>Overview of the evalita 2018 task on irony detection in italian tweets (ironita)</article-title>
          .
          <source>In: Proceedings of the 6th evaluation campaign of Natural Language Processing and Speech tools for Italian (EVALITA'18)</source>
          . pp.
          <volume>26</volume>
          {
          <fpage>34</fpage>
          . CEUR Workshop Proceedings, CEUR-WS (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2263</volume>
          /paper005.pdf
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Giudice</surname>
          </string-name>
          , V.:
          <article-title>Aspie96 at IronITA (EVALITA</article-title>
          <year>2018</year>
          )
          <article-title>: Irony Detection in Italian Tweets with Character-Level Convolutional RNN</article-title>
          .
          <source>In: Proceedings of the 6th evaluation campaign of Natural Language Processing and Speech tools for Italian (EVALITA'18)</source>
          . pp.
          <volume>160</volume>
          {
          <fpage>165</fpage>
          . CEUR Workshop Proceedings, CEUR-WS (
          <year>2018</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2263</volume>
          /paper026.pdf
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Giudice</surname>
          </string-name>
          , V.:
          <article-title>Aspie96 at HAHA</article-title>
          (IberLEF
          <year>2019</year>
          )
          <article-title>: Humor Detection in Spanish Tweets with Character-Level Convolutional RNN</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          , CEUR-WS, Bilbao, Spain (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Saur</surname>
          </string-name>
          , R.:
          <article-title>A Factuality Pro ler for Eventualities in Text</article-title>
          .
          <source>Ph.D. thesis</source>
          , Brandeis University (
          <year>2008</year>
          ), https://www.cs.brandeis.edu/ roser/pubs/sauriDiss 1.5.pdf
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>