<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Kernel-based Approach for Irony and Sarcasm Detection in Italian</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Santilli</string-name>
          <email>andrea.santilli@live.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Danilo Croce</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberto Basili</string-name>
          <email>basilig@info.uniroma2.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universita ́ degli Studi di Roma “Tor Vergata” Via del Politecnico 1</institution>
          ,
          <addr-line>Rome, 00133</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>English. This paper describes the UNITOR system that participated to the Irony Detection in Italian Tweets task (IronITA) within the context of EvalIta 2018. The system corresponds to a cascade of Support Vector Machine classifiers. Specific features and kernel functions have been proposed to tackle the different subtasks: Irony Classification and Sarcasm Classification. The proposed system ranked first in the Sarcasm Detection subtask (out of 7 submissions), while it ranked sixth (out of 17 submissions) in the Irony Detection task.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Modern social networks allow users to express
themselves, writing their opinions about facts,
things and events. In social posting, people
often adopt figurative languages, e.g. Irony and
Sarcasm. These communication mechanism must be
carefully considered in the automatic processing
of texts in social media: as an example, they may
be used to convey the opposite of literal meaning
and thus just intentionally sketching a secondary
or extended meaning
        <xref ref-type="bibr" rid="ref9">(Grice, 1975)</xref>
        . On Twitter,
users can express themselves with very short
messages. Given the short length, the information
useful to detect figurative uses of natural language is
very limited or missing. Irony and sarcasm
detection represents challenging tasks within Sentiment
Analysis and Opinion Mining often undermining
the overall system accuracy. There is not a clear
separation between irony and sarcasm, but the
former is often considered to include the latter. In
particular sarcasm is defined as sharp or cutting
ironic expressions towards a particular target with
the intention to offend
        <xref ref-type="bibr" rid="ref10">(Joshi et al., 2016)</xref>
        .
      </p>
      <p>
        This paper presents and describes the UNITOR
system participating in the Irony Detection in
Italian Tweets (IronITA) task
        <xref ref-type="bibr" rid="ref7">(Cignarella et al., 2018)</xref>
        within the EvalIta 2018 evaluation campaign. The
system faces both the proposed subtasks within
IronITA: Irony Classification and Sarcasm
Classification. In a nutshell, the former subtask aims
at evaluating the performance of a system in
capturing whether a message is ironic or not. The
second subtask is intended to verify if, given an ironic
tweet, a system is able to detect sarcasm within the
message.
      </p>
      <p>
        The classification of each tweet is carried out by
applying a cascade of kernel-based Support
Vector Machines
        <xref ref-type="bibr" rid="ref16">(Vapnik, 1998)</xref>
        . In particular, two
binary SVM classifiers (one per subtask) are
designed to adopt specific combinations of
different kernel functions, each operating over a
taskspecific tweet representation. This work extends
the modeling proposed in
        <xref ref-type="bibr" rid="ref3">(Castellucci et al., 2014)</xref>
        that was proved to be beneficial within the Irony
Detection subtask within SENTIPOLC 2014. The
UNITOR system here presented ranked 1st and
2nd in the Sarcasm Detection subtask, while it
ranked 6th and 7th within the Irony Detection
subtask.
      </p>
      <p>In Section 2 the SVM classifiers, their features
and the underlying kernels are described and the
adopted workflow is presented. In Section 3 the
performance measures of the system are reported,
while Section 4 derives the conclusions.
2</p>
    </sec>
    <sec id="sec-2">
      <title>System Description</title>
      <p>
        The UNITOR system adopts a supervised learning
setting where a multiple kernel-based approach
is adopted to acquire two binary Support Vector
Machine classifiers
        <xref ref-type="bibr" rid="ref14">(Shawe-Taylor and
Cristianini, 2004)</xref>
        : a first classifier discriminates between
ironic and non ironic tweets, while a second one
decides whether an ironic tweet is sarcastic or not.
In the rest of this section, we first summarize the
pre-processing stage as well as the adopted
linguistic resources (e.g. word embeddings or
lexicons). Then, the feature modeling designed for
the two steps is discussed.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Tweet processing and resources</title>
        <p>
          Each tweet is linguistically processed through an
adapted version of the Chaos parser
          <xref ref-type="bibr" rid="ref1">(Basili and
Zanzotto, 2002)</xref>
          in order to extract the
information required for feature modeling, e.g. the
Partof-speech tags and lemmas of individual words.
A normalization step is applied before the
standard Natural Language Processing activity is
carried out. A number of actions is performed: fully
capitalized words are converted into their
lowercase counterparts; hyperlinks are replaced by a
special token, i.e. LINK; characters repeated more
than three times are cleaned, as they increase
lexical data sparseness (e.g. “nooo!!!!!” is converted
into “noo!!”); all emoticons are replaced by
special tokens1.
        </p>
        <p>In the feature modeling activities, we relied on
several linguistic resources, hereafter reported.</p>
        <p>
          First, we used a Word Space model (or Word
Embedding) to generalize the lexical information
of the (quite small) training material: this
semantic space is obtained starting from a corpus of
Italian tweets downloaded in July 2016 of about 10
millions of tweets (same used in Castellucci et
al. (2016a)) and it is a 250-dimensional
embedding generated according to a Skip-gram model
          <xref ref-type="bibr" rid="ref12">(Mikolov et al., 2013)</xref>
          2.
        </p>
        <p>
          Moreover, we adopted a large scale sentiment
1We normalized 113 well-known emoticons in 13 classes.
2The following settings were adopted: window 5 and
mincount 10 with hierarchical softmax.
specific lexicon, i.e., the Distributional Polarity
Lexicons (DPL)
          <xref ref-type="bibr" rid="ref5 ref6">(Castellucci et al., 2016b)</xref>
          3.
Distributional Polarity Lexicon (DPL) is introduced
to inject sentiment information of words in the
learning process through a large-scale polarity
lexicon that is automatically acquired according to
the methodology proposed in
          <xref ref-type="bibr" rid="ref4">(Castellucci et al.,
2015)</xref>
          . This method leverages on word
embeddings to model lexical polarity by transferring it
from entire sentences whose polarity is known.
The process is based on the capability of word
embeddings to represent both sentences and single
words in the same space
          <xref ref-type="bibr" rid="ref11">(Landauer and Dumais,
1997)</xref>
          . First, sentences (here tweets) are labeled
with some polarity classes: in
          <xref ref-type="bibr" rid="ref4">(Castellucci et al.,
2015)</xref>
          this labeling is achieved by applying
simple heuristics, e.g. Distant Supervision
          <xref ref-type="bibr" rid="ref8">(Go et al.,
2009)</xref>
          . The labeled dataset is projected in the
embedding space by applying a simple but effective
linear combination of the word vectors composing
each sentence. Then, a polarity classifier is trained
over these sentences in order to emphasize
dimensions of the space that are more related to the
polarity classes. The DPL is generated by classifying
each word (represented in the embedding through
a vector) with respect to each targeted class,
using the confidence level of the classification to
derive a word polarity signature. For example, in a
DPL the word ottimo is 0:89 positive, 0:04
negative and 0:07 neutral. For more details, please
refer to
          <xref ref-type="bibr" rid="ref4">(Castellucci et al., 2015)</xref>
          .
        </p>
        <p>
          Finally, we also adopted an Irony specific
Corpus to capture terms and patterns that are often
used to express irony (e.g., “non lo
riconosceresti neanche se ti cascasse” or “: : : allora piove
”): it is a corpus composed by a set of Italian
tweets automatically extracted using Distance
Supervision
          <xref ref-type="bibr" rid="ref8">(Go et al., 2009)</xref>
          . In particular the Irony
specific Corpus is composed by a set of 6,000
random tweets in Italian, freely available, assumed to
be ironic, as they contain hashtags such as #irony
or #ironia.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Modeling irony and sarcasm in kernel-based learning</title>
        <p>UNITOR is based on kernel functions
operating on vector representations of tweets, described
hereafter. After the language processing stage,
each tweet allows generating one of the
follow3The adopted lexicon has been downloaded from
http://sag.art.uniroma2.it/demo-software/
distributional-polarity-lexicon/
ing representations4 , later exploited by the
kernelbased SVM in the training/classification steps.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.2.1 Irony-specific Features</title>
        <p>
          The aim of this set of features is to capture irony
by defining a set of irony-specific features inspired
by the work of
          <xref ref-type="bibr" rid="ref3">(Castellucci et al., 2014)</xref>
          .
Word Space Vector (WS) is a 250-dimensional
vector representation of the average semantic
meaning of a tweet according to a Word space
model. It is used to generalize the lexical
information of tweets. We can summarize it as
P W e(t)=jT j, where T is the set of nouns,
t2T
verbs, adjectives, adverb and hashtag in a tweet
t and W e(t) is a function that returns the
250dimensional word embedding of the word t. Other
words, such as articles and preposition are
discarded as they do not convey useful information
within a word space.
        </p>
        <p>
          Irony Specific BOW (ISBOW) is a BoW vector
representing the lexical information expressed in
a message. The main difference with respect to
a conventional BOW representation is the adopted
weighting scheme. In fact, in this case we leverage
on the Word Space previously described. For each
dimension representing a lemma/part-of-speech
pair, its weight is computed as the cosine
similarity between the word embedding vector of the
considered word and the vector obtained from the
linear combination of all the other words in the
message (WS)5. This vector aims at capturing how
much odd is the occurrence of a given word in
a sentence aiming at capturing its unconventional
uses: it should be an indicator of potential ironic
mechanisms, as suggested in
          <xref ref-type="bibr" rid="ref3">(Castellucci et al.,
2014)</xref>
          .
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Irony Specific BOW(Adjective, Noun, Verb)</title>
        <p>(ISBOW-A), (ISBOW-S), (ISBOW-V) are three
BoW vectors that use the same weighting scheme
specified in ISBOW. Each vector represents one
individual part of speech (i.e. adjective, noun and
verb), as words belonging to different POS-tag
categories may be characterized by quite different
distributions.</p>
        <p>Irony Specific Mean and Variance (ISMV) is a
4-dimensional vector representation that
summa4The code for the feature vector generation is available at:
https://github.com/andry9454/ironySarcasmDetection
5If a word was not found in the word embedding, a
smoothing weight, representing the mean cosine similarity
between word and WS in the training set, is applied as cosine
similarity measure.
rized the information captured by the previous
representations. It contains mean and variance of the
cosine similarity, calculated between the words in
a tweet in the ISBOW representation, and the
maximum and minimum of the cosine similarity per
tweet. This vector aims at summarizing the
distribution and potential ”spikes” of unusual patterns
of use for words in a sentence.</p>
      </sec>
      <sec id="sec-2-5">
        <title>Irony Specific Mean and Variance (Adjective,</title>
        <p>Noun, Verbs) (ISMV-A), (ISMV-S), (ISMV-V)
are three distinct 4-dimensional vectors that are
the same specified in ISMV, with the only
difference that each representation works on one
specific part of speech, respectively adjectives, nouns
and verbs.</p>
        <p>Char n-gram BOWs (n-CHARS) is a
representation expressing the char n-grams
contained in a message. We used 4 n-CHARS
representations: 2-CHARS BoW vector
representing 2-char-ngrams contained in a
message, 3-CHARS BoW vector representing
3-charngrams, 4-CHARS BoW vector representing
4char-ngrams, 5-CHARS BoW vector representing
5-char-ngrams. The aim of this representation is to
capture the usage of specific textual patterns, e.g.,
hihihihi often used to express irony.</p>
        <p>
          Synthetic Features (SF) is a 7-dimensional vector
containing the following synthetic features,
traditionally used in Sentiment Analysis: percentage
of the number of uppercase letters in the tweet,
number of exclamation marks, number of question
marks, number of colons, number of semicolons,
number of dots, number of commas. It has been
inspired by works on irony detection of
          <xref ref-type="bibr" rid="ref13 ref2">(Carvalho
et al., 2009; Reyes et al., 2012)</xref>
          .
2.2.2
        </p>
      </sec>
      <sec id="sec-2-6">
        <title>Features based on Distribution Polarity</title>
      </sec>
      <sec id="sec-2-7">
        <title>Lexicons</title>
        <p>
          The aim of this group of features is to exploit the
negative evaluation towards a target typical of
sarcasm mechanism
          <xref ref-type="bibr" rid="ref10">(Joshi et al., 2016)</xref>
          using a
polarity lexicon, here a Distribution Polarity Lexicon
(DPL).
        </p>
        <p>Distributional Polarity Lexicon Sum (DSUM)
is a 15-dimensional vector representation made
by the concatenation of 5 different
representations, i.e. 1 P wp, 1 P wp,
jNT j w2NT jVT j w2VT
jAd1jT j w2PAdjT wp, jAd1vT j w2PAdvT wp, jT1j wP2T wp,
where NT , VT , AdjT , Adv are the nouns, verbs,
adjectives and adverbs occurring in the tweet,
T = NT [ VT [ AdjT [ AdvT and wp expresses
the 3-dimensional polarity lexicon entry6 for the
word w. This feature summarize the a-priori
sentiment of words according to the different
morphological categories. We speculate that the
regularities or contrasts between these distributions may
suggest the presence of irony or sarcasm.</p>
      </sec>
      <sec id="sec-2-8">
        <title>Distributional Polarity Lexicon BoW (DBOW) is</title>
        <p>a BoW vector representing, for each word in a
message, its polarity (positive, negative and
neutral) as a three dimensional score derived from the
DPL.</p>
      </sec>
      <sec id="sec-2-9">
        <title>2.2.3 Irony Corpus Features</title>
        <p>Generalizing linguistic information useful for
Irony or Sarcasm detection is a very challenging
tasks, as the adoption of these figurative languages
mainly concern extra-linguistic phenomena. The
idea underlying the following features is to
define a tweet representation that is not directly
connected to their (possibly limited) linguistic
material, but that is connected with respect to a larger
set of information derived from a Irony specific
Corpus, i.e., a large scale collection of a Ironic
tweets. This is used to extract an Irony specific
Lexicon: a set of words and patterns occurring in
such corpus with a high frequency.</p>
        <p>Irony Corpus BOW (ICBOW) is a BoW vector
representing lemmas of Nouns, Verbs, and
Adjective in a message. Again, the main difference with
respect to a conventional BoW representation is the
adopted weighting scheme: a word is weighted
1:0 if that particular word was in the Irony specific
Corpus, otherwise is weighted 0.</p>
      </sec>
      <sec id="sec-2-10">
        <title>Irony Corpus weighted BOW (ICwBOW) is</title>
        <p>a BoW vector representing lemmas of Nouns,
Verbs, and Adjective in a message. A word is
weighted log(f + 1) where f is the frequency of
that particular word in the Irony Corpus.
Irony Corpus weighted Mean (ICM) is a
2dimensional vector representation that summarize
the mean words weight observed in a ICBOW
representation and the mean over the ICwBOW. These
scores indicate how a words or patterns in a tweet
occur also in the Irony specific corpus. This
information is very interesting as it is not tied to the
lexical information from a tweet, so allowing a more
robust generalization.</p>
        <p>Irony Corpus BOW (bi-grams, three-grams)
(IC2BOW), (IC3BOW) are two distinct BoW
vec6If a word w is not present in the distributional polarity
lexicon, wp is set to the default [0:33; 0:33; 0:33].
tor respectively representing bi-grams and
threegrams of surface words in a message. The
weighting scheme is the same explained in ICBOW.</p>
      </sec>
      <sec id="sec-2-11">
        <title>Irony Corpus weighted BOW (bi-grams, three</title>
        <p>grams) (IC2wBOW), (IC3wBOW) are two
distinct BoW vectors respectively representing
bigrams and three-grams of terms in a message.
The weighting scheme is the same explained in
ICwBOW.</p>
        <p>Irony Corpus weighted Mean (bi-grams,
threegrams) (IC2M), (IC3M) are two distinct
2dimensional vector representations that contain
means that are the same specified in ICM, with the
only difference that the first representation works
on bi-grams (IC2BOW, IC2wBOW), while the
second works on three-grams (IC3BOW, IC3wBOW).</p>
        <p>irony
classifier
yes
sarcasm
classifier
yes
Ironic and
sarcastic
1 1
no
no</p>
        <p>Ironic and
not sarcastic
1
0</p>
        <p>Not ironic nor
sarcastic
0 0</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experimental evaluation and results</title>
      <p>The cascade of SVM classifiers implemented in
UNITOR is summarized in Figure 1. After the
linguistic processing stage and the feature extraction
stage, each tweet is classified by a binary
classifier, the so-called irony classifier. If a message is
judged as not ironic, we assume that it is also not
sarcastic (according to the task guidelines) and a
label 0 0 is assigned to it. Otherwise, if the tweet
is judged as ironic, the second binary classifier,
devoted to Sarcasm Detection, is invoked. If
positive, the tweet is sarcastic and the message is
labeled with 1 1, otherwise, 1 0.</p>
      <p>
        Separated representations are considered in the
constrained and unconstrained settings,
according to the guidelines in
        <xref ref-type="bibr" rid="ref7">(Cignarella et al., 2018)</xref>
        .
In the constrained setting only feature vectors
using tweet information or public available lexicons
are considered (Irony-specific Features and
Features derived from a DPL). In the unconstrained
setting, feature vectors are derived also using the
Irony specific Corpus.
      </p>
      <p>In our experiments, we train the SVM
classifiers using the same kernel combination for Irony
Detection and Sarcasm Detection. Even if this is
not a general solution (different tasks may require
different representations) we adopted this greedy
strategy, leaving the SVM to select the most
discriminative information.</p>
      <p>
        A normalized linear combination of specific
kernel functions is used in both subtasks. In the
linear combination, a specific linear kernel is
applied to the following sparse representations:
ISBOW, ISBOW-A, ISBOW-S, ISBOW-V,
DBOW, 2BOW, 3BOW, 4BOW, 5BOW, ICBOW,
IC2BOW, IC3BOW, ICwBOW, IC2wBOW,
IC3wBOW; in the same combination a RBF kernel
        <xref ref-type="bibr" rid="ref14">(Shawe-Taylor and Cristianini, 2004)</xref>
        is applied to
the following dense representations WS, SF, ICM,
IC2M, IC3M, DSUM, ISMV, ISMV-A, ISMV-S,
ISMV-V7.
      </p>
      <p>Each SVM classifier is built by using the KeLP
framework8 (Filice et al., 2018).</p>
      <p>Figure 1 reflects also the learning strategy that
has been set up during the training phase: the
Irony Classifier was trained on the complete
training dataset composed by the entire training set
(made of 3; 977 tweets) while the Sarcasm
Classifier is trained only on the ironic tweets in the
training dataset (made of 2; 023 tweets). A
10fold cross validation strategy was applied to
optimize the SVM parameters, while the linear
combination of the kernel assigns the same weights to
each kernel function.</p>
      <p>
        In Table 1 the performances of the Irony
Classification task are reported: in the constrained run
the UNITOR system ranks 7th, while in 6st
position in the unconstrained one. For this task the
adopted representations were able to correctly
determine whether a message is ironic with good
precision. However, the winning system (about 3
points ahead) results more effective in the
detection of non-ironic messages. In fact, according to
the F1-score on the Ironic class, the system would
have been ranked 2nd. We also evaluated a slightly
different modeling with two additional features
vector, i.e., a classic BoW composed of lemmas
derived from the input tweet, and a BoW of
bigrams. These features have been excluded from
7A with = 1 was used in each RBF kernel
8http://www.kelp-ml.org/
our official submission to keep the model simple.
However, these simple features would have been
beneficial and the system would have ranked 2nd.
Performances on the Sarcasm Classification are in
Table 2: UNITOR here ranks in 1st or in 2nd
position, in the constrained and unconstrained run,
respectively. Differences between the two results
are not significant. Nevertheless the further
features derived from the Irony specific corpus
allow improving results (especially in terms of
recall) in the Sarcasm Detection task. For this
latter task, results achieved by UNITOR suggest that
the proposed modeling, in particular the
contribution of Polarity Features, seem to be beneficial. To
prove it, we decided to evaluate a run with the
same winning features, except Polarity Features.
In this case the UNITOR system would have been
ranked 4th. These Polarity Features seem to
exploit the negative bias typical of sarcasm
        <xref ref-type="bibr" rid="ref10">(Joshi et
al., 2016)</xref>
        .
      </p>
      <p>
        1st
2nd*
6th(u)
7th(c)
BL
In this paper we described the UNITOR system
participating to the IronITA task at EvalIta 2018.
The system won 1 of the 2 evaluations carried
out in the task, and in the worst case it ranked
in the 6th position. The good results in
constrained and unconstrained settings suggest that
the proposed irony and sarcasm specific features
were beneficial to detect irony and sarcasm also in
short messages. However, further work is needed
to improve the non ironic F1 scores. The
nature of the task seems to be non trivial also for
a human reader, as some tweets extracted from
the test set suggest: “@beppe grillo Beppe..tu sei
un grande..questi si stanno finendo di mangiare
l’Italia..”, “scusa hai ancora posti liberi nella app
di braccialetti rossi?”; here the interpretation of
irony goes beyond the textual information and it is
very difficult to state if these messages are ironic
or not. Since tweets are very short, useful
information for detecting irony is often out of the
message, like this ironic tweet extracted from the test
set may suggest: “immagine perfetta ed esplicita
che descrive la realt a´ della ”buona scuola” a
renzopoli”; in this case the system may fail without a
proper representation for the meaning of the
neologism “renzopoli”. So we think that the contextual
approach suggested in
        <xref ref-type="bibr" rid="ref15">(Vanzo et al., 2014)</xref>
        will be
explored in future research.
Speech tools for Italian (EVALITA’18), Turin, Italy.
      </p>
      <p>CEUR.org.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          and Fabio Massimo Zanzotto.
          <year>2002</year>
          .
          <article-title>Parsing engineering and empirical robustness</article-title>
          .
          <source>Nat. Lang</source>
          . Eng.,
          <volume>8</volume>
          (
          <issue>3</issue>
          ):
          <fpage>97</fpage>
          -
          <lpage>120</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Paula</given-names>
            <surname>Carvalho</surname>
          </string-name>
          , Lu´ıs Sarmento, Ma´rio
          <string-name>
            <given-names>J.</given-names>
            <surname>Silva</surname>
          </string-name>
          , and Euge´nio de Oliveira.
          <year>2009</year>
          .
          <article-title>Clues for detecting irony in user-generated contents: Oh.</article-title>
          ..!! it's ”so easy” ;-).
          <source>In 1st CIKM WS on Topic-sentiment Analysis for Mass Opinion</source>
          , pages
          <fpage>53</fpage>
          -
          <lpage>56</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Giuseppe</given-names>
            <surname>Castellucci</surname>
          </string-name>
          , Danilo Croce, Diego De Cao, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>A multiple kernel approach for twitter sentiment analysis in italian</article-title>
          .
          <source>In Fourth International Workshop EVALITA</source>
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Giuseppe</given-names>
            <surname>Castellucci</surname>
          </string-name>
          , Danilo Croce, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Acquiring a large scale polarity lexicon through unsupervised distributional methods</article-title>
          .
          <source>In Proc. of 20th NLDB</source>
          , volume
          <volume>9103</volume>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Giuseppe</given-names>
            <surname>Castellucci</surname>
          </string-name>
          , Danilo Croce, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          . 2016a.
          <article-title>Context-aware convolutional neural networks for twitter sentiment analysis in italian</article-title>
          .
          <source>In Proceedings of 3rd Italian Conference on Computational Linguistics</source>
          (CLiC-it
          <year>2016</year>
          ) &amp;
          <source>Fifth EVALITA Workshop</source>
          <year>2016</year>
          ), Napoli, Italy, December 5-
          <issue>7</issue>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Giuseppe</given-names>
            <surname>Castellucci</surname>
          </string-name>
          , Danilo Croce, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          .
          <year>2016b</year>
          .
          <article-title>A language independent method for generating large scale polarity lexicons</article-title>
          .
          <source>In Proceedings of the 10th LREC Conference (LREC'16)</source>
          , Portoroz,
          <string-name>
            <given-names>Slovenia. European</given-names>
            <surname>Language Resources Association (ELRA).</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Alessandra</given-names>
            <surname>Teresa</surname>
          </string-name>
          <string-name>
            <surname>Cignarella</surname>
          </string-name>
          , Simona Frenda, Valerio Basile, Cristina Bosco, Viviana Patti, and
          <string-name>
            <given-names>Paolo</given-names>
            <surname>Rosso</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Overview of the evalita 2018 task on irony detection in italian tweets (ironita)</article-title>
          .
          <source>In Tommaso Caselli</source>
          , Nicole Novielli, Viviana Patti, and Paolo Rosso, editors,
          <source>Proceedings of the 6th evaluation campaign of Natural Language Processing and Simone Filice</source>
          , Giuseppe Castellucci, Giovanni Da San Martino, Alessandro Moschitti, Danilo Croce, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Kelp: a kernel-based learning platform</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>18</volume>
          (
          <issue>191</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Alec</given-names>
            <surname>Go</surname>
          </string-name>
          , Richa Bhayani, and
          <string-name>
            <given-names>Lei</given-names>
            <surname>Huang</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Twitter sentiment classification using distant supervision</article-title>
          .
          <source>Processing</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>H Paul</given-names>
            <surname>Grice</surname>
          </string-name>
          .
          <year>1975</year>
          .
          <article-title>Logic and conversation</article-title>
          .
          <source>1975</source>
          , pages
          <fpage>41</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Aditya</given-names>
            <surname>Joshi</surname>
          </string-name>
          , Pushpak Bhattacharyya, and Mark James Carman.
          <year>2016</year>
          .
          <article-title>Automatic sarcasm detection: A survey</article-title>
          .
          <source>CoRR, abs/1602</source>
          .03426.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Tom</given-names>
            <surname>Landauer</surname>
          </string-name>
          and
          <string-name>
            <given-names>Sue</given-names>
            <surname>Dumais</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>A solution to plato's problem: The latent semantic analysis theory of acquisition, induction and representation of knowledge</article-title>
          .
          <source>Psychological Review</source>
          ,
          <volume>104</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , Kai Chen, Greg Corrado, and
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Efficient estimation of word representations in vector space</article-title>
          .
          <source>CoRR, abs/1301</source>
          .3781.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Antonio</given-names>
            <surname>Reyes</surname>
          </string-name>
          , Paolo Rosso, and
          <string-name>
            <given-names>Davide</given-names>
            <surname>Buscaldi</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>From humor recognition to irony detection: The figurative language of social media</article-title>
          .
          <source>Data and Knowledge Engineering</source>
          ,
          <volume>74</volume>
          (
          <issue>0</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>John</surname>
            Shawe-Taylor and
            <given-names>Nello</given-names>
          </string-name>
          <string-name>
            <surname>Cristianini</surname>
          </string-name>
          .
          <year>2004</year>
          .
          <article-title>Kernel Methods for Pattern Analysis</article-title>
          . Cambridge University Press.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Andrea</given-names>
            <surname>Vanzo</surname>
          </string-name>
          , Danilo Croce, and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Basili</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>A context-based model for sentiment analysis in twitter</article-title>
          .
          <source>In Proceedings of COLING</source>
          , pages
          <fpage>2345</fpage>
          -
          <lpage>2354</lpage>
          . ACL and Dublin City University.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Vladimir N.</given-names>
            <surname>Vapnik</surname>
          </string-name>
          .
          <year>1998</year>
          .
          <article-title>Statistical Learning Theory</article-title>
          . Wiley-Interscience.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>