<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Study in Practical Solutions to Sarcasm Detection with Machine Learning and Knowledge Engineering Techniques</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chia Zheng Lin</string-name>
          <email>fchiazhengling@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michal Ptaszynski</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Masui Fumito</string-name>
          <email>f-masuig@cs.kitami-it.ac.jp</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gniewosz Leliwa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michal Wroczynski</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Graduate School of Computer Science, Kitami Institute of Technology</institution>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <fpage>23</fpage>
      <lpage>25</lpage>
      <abstract>
        <p>In this paper we tackle the problem of sarcasm detection with the use of machine learning and knowledge engineering techniques. Sarcasm detection is considered a complex and challenging task in Natural Language Processing and has been studied by various researchers in the past decade. To get a grasp on the present state of the art in sarcasm detection, we review the important previous research in this field, with a focus on text-based sarcasm detection in English texts. In the proposed method, we compare various dataset preprocessing techniques on the proposed Deep Convolutional Neural Network model. As a result, the most specific, or least preprocessed dataset ranked as the one with the highest performance. However, we observed that some level of data preprocessing could become useful in the task of sarcasm detection.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Sarcasm, often used together or interchangeably with irony,
is considered an important component of human
communication recognized as some of the most prominent and
pervasive devices of figurative and creative language widely used
from dating back to ancient religious texts to modern times
        <xref ref-type="bibr" rid="ref10 ref24">(Ghosh and Veale 2017)</xref>
        .
      </p>
      <p>
        Van
        <xref ref-type="bibr" rid="ref13">Hee (2017)</xref>
        suggested the important implications of
irony and sarcasm for Natural Language Processing (NLP)
tasks, which aim to explain construct of human language,
and the large potential in the domain of text mining. In the
recent years, there has been an increasing interest in,
especially, automatic sarcasm detection and classification, which
have been widely studied as a type of sentiment analysis task
(detecting whether a sentence conveys a positive or negative
connotation, or in this case: sarcastic or non-sarcastic).
Especially,
        <xref ref-type="bibr" rid="ref16">Kumar et al. (2017)</xref>
        surveyed some representative
work in the related area and categorized most of the popular
approaches into three types, namely, rule-based, statistical,
and deep learning-based approaches. We analyse some of
that research in the next section.
      </p>
      <p>
        Researchers’ interest in analysing this profound type of
figurative and creative use of language grew along with the
dramatic increase in the everyday use of social media over
the past decade. Especially, Twitter has become one of the
most popular venues for people to express their opinions,
share their thoughts and report real-time events, etc.
Moreover, the huge amount of data has drawn interest of
companies for the purpose of studying the opinion of people
towards different products, facilities and events. It has been
suggested that the nature of tweets makes them the most
suitable for studying sarcasm detection approaches
        <xref ref-type="bibr" rid="ref10 ref7">(Bouazizi and Otsuki 2016)</xref>
        .
      </p>
      <p>However, the lack of empirical investigations into
optimal approaches for sarcasm detection is a serious oversight
in many related studies carried out throughout the years.
Importantly, there have been no studies comparing the
differences in the preprocessing and manipulation of the dataset
to improve the results of detection.</p>
      <p>To contribute to dealing with the above-mentioned
problems, in this paper we investigate the variations in
sarcasm detection results caused by differences in applied
preprocessing techniques typically used in NLP research but
not applied before in works focusing on sarcasm detection.
To do that most effectively, we firstly review previous
related research on text-based sarcasm detection from
English tweets, describe the implemented dataset
preprocessing techniques, and discuss the results of an experiment
performed to compare preprocessing techniques implemented
on the dataset. As a result, we managed to observe the
impact contributed by hashtags and labels related to sarcasm.</p>
      <p>
        Finally,
        <xref ref-type="bibr" rid="ref21">Ptaszynski et al. (2010)</xref>
        in their research on
developing an expert system for Internet Patrol pointed out that,
especially with regard to the increased popularity of SNS,
sarcasm has been often used in personal attacks, such as
cyberbullying and concluded that sarcasm detection is one of
the important problems in cyberbullying detection.
Therefore, as one of the practical applications, in this research we
will verify how effective is sarcasm detection in the
detection of cyberbullying.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Research Background</title>
      <p>
        The word sarcasm originates from an Ancient Greek word
sarkasmo´s and means ”to tear flesh, bite the lip in rage,
sneer.” According to
        <xref ref-type="bibr" rid="ref18">Oxford dictionary (2019</xref>
        ), sarcasm is a
way of using words that are the opposite of what one means
in order to be unpleasant to somebody or to make fun of
them. They also described irony to be the use of words that
say the opposite of what you really mean, often as a joke.
      </p>
      <p>The relationship between irony and sarcasm has been
confused in many studies. In the literature, two types of irony are
widely considered: verbal irony and situational irony. While
situational irony involves an incongruence between two
situations, verbal irony, although applying verbal, or semantic
incongruence, is a statement in which the meaning that a
speaker employs is sharply different from the meaning that
is ostensibly expressed. Hence, verbal irony is considered
different from situational irony in that it is produced
intentionally by the speakers.</p>
      <p>
        When it comes to sarcasm, Van
        <xref ref-type="bibr" rid="ref13">Hee (2017)</xref>
        defines it to
be a form of verbal irony with an aggressive tone, is directed
at someone or something, and is used intentionally. Hence
the term “irony” and “sarcasm” are used interchangeably in
many related studies. In this study, we decided to not focus
on distinguishing between sarcasm and irony, and instead
implement the general term “sarcasm” throughout the paper.
      </p>
      <sec id="sec-2-1">
        <title>Previous Research</title>
        <p>
          <xref ref-type="bibr" rid="ref27">Tepperman (2006)</xref>
          ’s spoken dialogue system used feature
extraction approach for sarcasm detection as a subtask in
their system, by which they introduced sarcasm detection
into the scene of Nature Language Processing. One study
by
          <xref ref-type="bibr" rid="ref9">Davidov (2010)</xref>
          utilized tweets and Amazon reviews for
text-based sarcasm detection, and
          <xref ref-type="bibr" rid="ref28">Tsur (2010)</xref>
          proposed one
of the first attempts to use feature engineering and statistical
classifiers to detect sarcasm.
        </p>
        <p>
          A number of studies have sought to detail the recent
trend in sarcasm detection approaches, which can roughly be
classified into three parts: rule-based, statistical, and
deeplearning approaches
          <xref ref-type="bibr" rid="ref10 ref16 ref2 ref24">(Kumar, Somani, and Bhattacharyya
2017; Barbieri 2017)</xref>
          . Rule-based approaches attempt to
identify irony through specific evidence which could be
captured in terms of rules that rely on indicators of sarcasm.
Barberi (2017) argued that rule-based approaches which
require no training mostly rely on lexical information and do
not perform as well as statistical approaches.
          <xref ref-type="bibr" rid="ref26">Riloff (2013)</xref>
          aimed to recognize positive words in negative sentences
while presenting a bootstrapping algorithm that
automatically learns the rules from certain situations.
        </p>
        <p>
          Most of the early works on sarcasm detection applied
statistical approaches which varied in terms of features
and learning algorithms, basically composed of two phases
where data were converted into feature vectors before
being classified using machine learning algorithm. Some of
the most often used algorithms include Support Vector
Machines (SVM), and Na¨ıve Bayes. One of the first attempts
in this approach by
          <xref ref-type="bibr" rid="ref28">Tsur (2010)</xref>
          compiled a set of
sarcastic patterns composed of common combinations of words
extracted from sarcastic examples.
          <xref ref-type="bibr" rid="ref11">Gonzalez-Ibanez (2011)</xref>
          composed a model with three pragmatic features which were
positive emoticons, negative emoticons, and users’ tagging.
          <xref ref-type="bibr" rid="ref25">Reyes (2013)</xref>
          proposed another model based on four
features, signatures, unexpectedness, style and polarity, and
emotional scenarios.
        </p>
        <p>
          Deep Learning approaches have been successfully
brought into the scene of sarcasm detection when
          <xref ref-type="bibr" rid="ref1">Amir
(2016)</xref>
          used a standard binary classification with
Convolutional Neural Network (CNN) while
          <xref ref-type="bibr" rid="ref19">Poria (2016)</xref>
          implemented a combination of CNNs trained on different tasks.
Popular Deep Learning algorithms include CNN
          <xref ref-type="bibr" rid="ref17">(LeCun et
al. 1998)</xref>
          and Long Short Term Memory (LSTM)
          <xref ref-type="bibr" rid="ref14">(Hochreiter and Schmidhuber 1997)</xref>
          .
          <xref ref-type="bibr" rid="ref10">Ghosh and Veale (2017)</xref>
          proposed a network model composed of CNN followed by an
LSTM network which outperformed many other models at
that time. They utilized CNN to reduce frequency
variation through convolutional filters and extract discriminating
word sequences as a composite feature map for the LSTM
layer. Then the output of the LSTM layer was passed to a
fully connected Deep Neural Network (DNN) layer,
producing a higher order feature set based on the LSTM output.
        </p>
        <p>
          Follow
          <xref ref-type="bibr" rid="ref5">ing the Semantic Evaluation 2018</xref>
          international
workshop Task 3: Irony Detection
          <xref ref-type="bibr" rid="ref5">in English Tweets (2018</xref>
          )
which received submissions from 43 teams worldwide for
the binary classification task A, deep learning algorithms
were further explored and optimized for irony detection
tasks. The best ranked system submitted by team THU NGN
(2018) consisted of densely connected LSTM network with
multi-task learning strategy. Another system from one of the
top teams,
          <xref ref-type="bibr" rid="ref4">NTUA-SLP (2018</xref>
          ), which used an ensemble of
two bi-directional LSTM network-based models, achieved
comparable results. The submissions represented a variety
of neural network-based approaches and other popular
classification algorithms including SVM, Random Forest, and
Na¨ıve Bayes
          <xref ref-type="bibr" rid="ref29">(Van Hee, Lefever, and Hoste 2018)</xref>
          .
Overall, the approaches with ensemble learners were the current
trend to tackle the challenges in sarcasm detection.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Proposed Method</title>
      <sec id="sec-3-1">
        <title>Dataset Preprocessing</title>
        <p>In the majority of recent studies applying machine
learning methods to text classification, the datasets are
usually used in their most basic form, namely, represented
as tokens (words, punctuation, etc.), despite a wide
variety of knowledge-based NLP systems (e.g., stemmers,
partof-speech taggers, etc.) capable of initial preprocessing of
datasets, thus providing more informative features to ML
algorithms. Therefore in this research we performed
additional preprocessing to the dataset to verify usefulness of
such knowledge-base systems in ML.</p>
        <p>For the implemented dataset, each tweet was first
transformed into lowercase and emojis were represented with
their corresponding labels (e.g. :smileyface:) using Emoji
for Python (2019). All tagged users (e.g. @user123) and
URLs (e.g. http://google.com/) appearing in the text were
replaced with specific neutral labels, such as ” tagged ” and
” url .” The first dataset preprocessing technique to be used
in this study is shown below.
1. Only basic preprocessing.</p>
        <p>
          To verify the depth of dependence of sarcasm detection on
hashtags, all of the hashtags (e.g. #sarcasm) in the next 5
versions of the dataset shown below were replaced with a
general label, e.g., “ hashtag .”
2. URLs, tagged users and hashtags replaced with labels.
Furthermore, we applied the knowledge-based tools for
language processing provided by NLTK (2019).
3. Stemming of all words using
          <xref ref-type="bibr" rid="ref20">Porter Stemmer (2019</xref>
          )
4. Stopwords removal with NLTK built-in Stopwords
Filtering Tool
5. Stemming of all words after stopwords removal
6. PoS tagging using NLTK Universal Part-of-Speech Tagset
Finally we have our last dataset 7 to have its social media
markers such as hashtags, URLs, and tagged users removed
instead of being replaced with labels.
7. Tagged users, URLs, and hashtags removed
Below are three examples of a tweet, with hashtags (dataset
1), with hashtags replaced with labels (dataset2), and with
hashtags removed (dataset7).
monday morning is my favorite! #sarcasm
monday morning is my favorite! _hashtag_
monday morning is my favorite!
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Feature Weighting</title>
        <p>Traditional weight calculation scheme was applied to all
versions of the dataset. In particular, we used term frequency
with inverse document frequency (tf*idf). Term frequency
tf (t,d) refers here to the traditional raw frequency, which is
the number of times a term t (word, token) occurs in a
document d. Inverse document frequency idf (t,D) is the
logarithm of the total number of documents D containing the
term t. Finally tf*idf refers to the term frequency multiplied
by inverse document frequency as in equation 1.
idf (t; D) = log jDj
nt
(1)</p>
      </sec>
      <sec id="sec-3-3">
        <title>Applied Classifier</title>
        <p>
          Based on our previous work
          <xref ref-type="bibr" rid="ref10 ref23 ref24 ref8">(Chia, Ptaszynski, and Masui
2019; Ptaszynski, Eronen, and Masui 2017)</xref>
          , in this study
we propose to use Convolutional Neural Networks (CNN)
due to it having the best result for classifying tweets without
ironic hashtags when compared to other classifiers.
        </p>
        <p>
          CNN are a type of feed-forward artificial neural network
which is an improved neural network model originally
designed for image recognition. CNN performance has been
proved useful in various classification tasks including
sentence classification and NLP
          <xref ref-type="bibr" rid="ref10 ref15 ref23 ref24">(Kim 2014; Ptaszynski,
Eronen, and Masui 2017)</xref>
          .
        </p>
        <p>In the proposed CNN we applied Rectified Linear Units
(ReLU) as neuron activation function which is a piece-wise
linear function that will output the input directly if positive,
zero if negative. We also applied dropout regularization. The
CNN consisted of two hidden convolutional layers,
containing 20 and 100 feature maps, respectively, with both layers
having 5x5 patch size and 2x2 max-pooling.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation Experiment</title>
      <sec id="sec-4-1">
        <title>Dataset Description</title>
        <p>
          The dataset used in this research was the publicly
available sarcasm detection dataset collected by
          <xref ref-type="bibr" rid="ref10">Ghosh and
Veale (2017)</xref>
          and consists of 51,189 tweets (24,453
sarcastic tweets and 26,736 non-sarcastic tweets) in which
sarcastic tweets were automatically collected from Twitter using
user’s self-declaration of sarcasm/irony with sarcastic and
ironic hashtags (e.g. #irony, #sarcasm) and annotated for
confirmation. All seven dataset versions were implemented
with different data preprocessing methods.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Experiment Setup</title>
        <p>All seven separate versions of the dataset (represented with
various preprocessing techniques) were analysed in the
experiment using the proposed CNN method in the setting of
a 10-fold cross validation procedure. The results were
calculated using standard balanced F-score (F1) which is the
harmonic mean of Precision and Recall.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Results and Discussion</title>
        <p>
          Table 1 shows the summary of all results from the 7 datasets
with different preprocessing techniques applied. Dataset 1
which is the dataset with all the hashtags included yielded
an F1 score of 0.997. Compared to our previous work
          <xref ref-type="bibr" rid="ref8">(Chia,
Ptaszynski, and Masui 2019)</xref>
          which tested on a smaller data
set with only 4,618 tweets and attained an F1 score of 0.844
with similar settings (hashtags included), this shows the
significant increase in the performance of the CNN model with
the increase of the size of the dataset. This suggests that the
model is tied to the size of the implemented dataset and the
number of extracted features.
        </p>
        <p>The results of dataset 1 (hashtags included) also enhance
our understanding of the impact of hashtags, which make
a great difference in sarcasm and irony detection, especially
in Twitter messages. However, due to the natural
characteristics of deliberate sarcastic hashtags in Twitter, classification
of tweets with hashtags included does not contribute much
to the study of sarcasm detection from linguistic point of
view. However, as the results show, hashtags can be a very
useful practical mean to handle sarcasm detection with high
performance.</p>
        <p>While the remaining datasets were stripped of their
hashtags (replaced with labels), data set 2 has no further
preprocessing while data set 3 to 6 were further processed with
different methods. Interestingly, data set 2 still attained the
highest F1 score among all the data sets without hashtags
included. This discovery highlights the importance of
linguistic features in irony detection and shows that increment
in data preprocessing does not always provide better results.
This is due to the oversimplification of data with many vital
and important features manipulated or removed while
classification tasks such as irony detection heavily depended on
them.</p>
        <p>However, further preprocessed data sets have their own
value despite attaining lower F1 score. From our observation
on the attributes extracted from their confusion matrices in
Table 1, their true positive rate is higher than the data set 2
which scored the highest F1 score among the datasets. Data
set 5 which implemented both stemming and stop-word
removal has obtained the highest true positive rate with only
290 false positive. This shows the implementation of further
data preprocessing is crucial to the sensitivity of the data.</p>
        <p>Finally for the last dataset 7 which had all of its social
media markers, such as tagged users (e.g. @user123), URLs,
and hashtags completely removed, Table 1 shows that the
result dropped significantly to an F1 score of 0.665 comparing
to other datasets. This case has shown the impact of the
labels which were supposed to be neutral to the classification.
Comparing to dataset 2 which had the social media markers
replaced with labels, the significant increase in false
negatives shows that the presence of the labels provides heavy
contribution to the precision of the classification.</p>
      </sec>
      <sec id="sec-4-4">
        <title>Error Analysis</title>
        <p>Table 2 shows the occurrences of top 6 error features
extracted from dataset 1 (with hashtags), dataset 2 (hashtags
replaced with labels) and dataset 7 (hashtags, URLs, and
tagged users removed) after removing prepositions,
conjunctions, and pronouns which do not contribute much to
the classifications. For dataset 1, the error feature which
occurred the most is the #sarcasm following the word sarcasm.
This shows that even the sarcastic hashtags cannot assist the
model to achieve 100% sensitivity.</p>
        <p>For the dataset 2 results in the second column, the label
hashtag appeared 5445 times out of the 5466 misclassified
instances (99.62%). Coming up next is the label tagged
which appeared 1639 times while the remaining words such
as “love”, “great”, “not”, and “best”, which are popular
errors in all the 3 implemented datasets. As previously noticed,
the supposedly neutral labels, in fact contribute heavily to
the precision of the classification. Therefore, removing them
does not provide improvement to the results.</p>
        <p>The evidences so far provide further support for the
hypothesis that deliberate sarcastic hashtags play a significant
role in sarcasm detection in tweets. Taken together, these
results also suggest that hashtag is the product of authors who
understand that their sarcastic phrases alone may not be
sufficient for the audience to figure out the intended irony or
sarcasm. However, these findings do not completely solve
the general sarcasm detection nor do they redefine sarcasm
or irony in textual communication especially on social
network service.</p>
      </sec>
      <sec id="sec-4-5">
        <title>Application in Automatic Cyberbullying Detection</title>
        <p>
          Although the number of research on sarcasm and irony
detection grows each year, practical implementation of such
models have not been widely discussed. Ptaszynski et al.
          <xref ref-type="bibr" rid="ref21">(Ptaszynski et al. 2010)</xref>
          mentions, that sarcasm poses a
problem in cyberbullying (CB) detection. Therefore, aiming to
improve their expert system for automated Internet Patrol,
we propose a practical implementation of sarcasm detection
in cyberbullying detection.
        </p>
        <p>
          To quantify the extent to how such model would be
useful, we applied the model trained on sarcastic dataset 2 and
tested on the cyberbullying detection dataset provided by
          <xref ref-type="bibr" rid="ref22">Ptaszynski et al. (2018)</xref>
          which consists of 12,728 data
samples. The result attained an F-score of 0.889 which is
comparable to the result of dataset 2 with an F-score of 0.898
above. Interestingly, it was also much higher than models
trained on purely cyberbullying-related data
          <xref ref-type="bibr" rid="ref22">(Ptaszynski et
al. 2018)</xref>
          . This observation shows the prevalence of sarcasm
in cyberbullying, and proves the practical applicability of
sarcasm detection in other tasks.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this paper, to find practical solutions for sarcasm
detection on Twitter, we compared various dataset preprocessing
methods and observed the impact of the preprocessed labels.</p>
      <p>We firstly reviewed previous related works on text-based
sarcasm detection, where we covered various types of
systems, such as rule-based, statistical, or deep learning-based.
Next, we compared datasets with various preprocessing on
the proposed CNN model.</p>
      <p>The first dataset with hashtags included scored an F1 of
0.9965, thus proving the dependence on hashtags in
sarcasm detection. Next, the dataset with the least
preprocessing ranked the highest among all datasets without hashtags
included. However, we observed that data preprocessing is
still crucial to the sensitivity of data Lastly, this research
serves as a base for future studies on application of sarcasm
detection in other tasks, such as cyberbullying detection.</p>
      <p>In the future, we also plan to further improve the proposed
method with more and diverse features and test it on larger
datasets, also with other preprocessing techniques. We will
also focus on optimizing the feature extraction and the
classifier model.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Amir</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wallace</surname>
            ,
            <given-names>B. C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lyu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ; Carvalho,
          <string-name>
            <surname>P.</surname>
          </string-name>
          ; and Silva,
          <string-name>
            <surname>M. J.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>Modelling context with user embeddings for sarcasm detection in social media</article-title>
          .
          <source>In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning (CoNLL)</source>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Barbieri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Machine learning methods for understanding social media communication: Modeling irony and emojis</article-title>
          . Department DTIC.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Baziotis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Athanasiou</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Papalampidi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kolovou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Paraskevopoulos</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Ellinas,
          <string-name>
            <given-names>M.</given-names>
            ; and
            <surname>Potamianos</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>Ntua-slp at semeval-2018 task 3: Tracking ironic tweets using ensembles of word and character level attentive rnns</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <source>In Proceedings of the 12th International Workshop on Semantic Evaluation(SemEval-</source>
          <year>2018</year>
          ).
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Bird</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Loper</surname>
          </string-name>
          , E.; and
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Natural language toolkit</article-title>
          . https://www.nltk.org/.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Bouazizi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Otsuki</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>A pattern-based approach for sarcasm detection on twitter</article-title>
          .
          <source>In Digital Object.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Chia</surname>
            ,
            <given-names>Z. L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ptaszynski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Masui</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>Exploring machine learning techniques for irony detection</article-title>
          .
          <source>In Proceedings of The 33rd Annual Conference of the Japanese Society for Artificial Intelligence (JSAI</source>
          <year>2019</year>
          ).
          <source>Japanese Society of Artificial Intelligence.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Davidov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Tsur</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ;
          <article-title>and</article-title>
          <string-name>
            <surname>Rappoport.</surname>
          </string-name>
          <year>2010</year>
          .
          <article-title>Semisupervised recognition of sarcastic sentences in twitter and amazon</article-title>
          .
          <source>In Proceedings of the 14th Conference on Computational Natural Language Learning. Association of Computational Linguistics.</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Veale</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Fracking sarcasm using neural network</article-title>
          .
          <source>In Proceedings of NAACL-HLT</source>
          <year>2016</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Gonzalez-Ibanez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ; Muresan,
          <string-name>
            <given-names>S.</given-names>
            ; and
            <surname>Wacholder</surname>
          </string-name>
          ,
          <string-name>
            <surname>N.</surname>
          </string-name>
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <article-title>Identifyinng sarcasm in twitterl: A closer look</article-title>
          .
          <source>In Proceedings of the 49th Annual Meeting of the Association For Computational Linguistics</source>
          .
          <article-title>Assosiation for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Hee</surname>
            ,
            <given-names>C. V.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Can machine sense irony? exploring automatic irony detection on social media</article-title>
          . University Gent.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1997</year>
          .
          <article-title>Long short-term memory</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Convolutional neural networks for sentence classification</article-title>
          .
          <source>In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Somani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Bhattacharyya,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Approaches for computational sarcasm detection: A survey</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.;
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ; and Haffner,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          Oxford.
          <year>2019</year>
          .
          <article-title>Oxford learner's dictionary.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Poria</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Cambria</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hazarika</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and Vij,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>A deeper look into sarcastic tweets using deep convolutional neural networks</article-title>
          .
          <source>COLING</source>
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Porter</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>The porter stemming algorithm</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Ptaszynski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dybala</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; Matsuba,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Masui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ;
            <surname>Rzepka</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          ; Araki,
          <string-name>
            <surname>K.</surname>
          </string-name>
          ; and Momouchi,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <year>2010</year>
          .
          <article-title>In the service of online order: Tackling cyber-bullying with machine learning and affect analysis</article-title>
          .
          <source>In International Journal of Computational Linguistics Research</source>
          . Hokkaido University.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Ptaszynski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Leliwa</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Piech,
          <string-name>
            <given-names>M.</given-names>
            ; and
            <surname>Smywinski-Pohl</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <year>2018</year>
          . Cyberbullying detection -
          <source>technical report 2/</source>
          <year>2018</year>
          , department of computer science agh, university of science and technology.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Ptaszynski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Eronen</surname>
            ,
            <given-names>J. K. K.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Masui</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Learning deep on cyberbullying is always better than brute force</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>In IJCAI 2017 3rd Workshop on Linguistic and Cognitive</source>
          Approaches to Dialogue Agents (LaCATODA
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Reyes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; and Veale,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2013</year>
          .
          <article-title>A multidimensional approach for detecting irony in twitter</article-title>
          .
          <source>Lang Resources and Evaluation.</source>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Riloff</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Qadir</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Surve</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Silva</surname>
            ,
            <given-names>L. D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gilber</surname>
            , N.; and Huang,
            <given-names>R.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Sarcasm as contrast between a positive sentiment and negative situation</article-title>
          .
          <source>In Proceddings of the 2013 Conference on Empirical Methods in Natural Language Processing(EMNLP</source>
          <year>2013</year>
          ). EMNLP.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Tepperman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Yeah right: Sarcasm recognition for spoken dialogue system</article-title>
          .
          <source>In Interspeech</source>
          <year>2006</year>
          . ICSLP.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Tsur</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Davidov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Rappoport</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2010</year>
          .
          <article-title>Icwsm - a great catchy name: Semi-supervised recognition of sarcastic sentences in online product reviews</article-title>
          .
          <source>In Proceedings of the 4th International AAAI Conference on Weblogs and Social Media.</source>
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Van Hee</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lefever</surname>
          </string-name>
          , E.; and
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Semeval-2018 task 3: Irony detection in english tweets</article-title>
          .
          <source>In Proceedings of the 12th International Workshop on Semantic Evaluation (SemEval2018)</source>
          .
          <article-title>Association for Computational Linguistic</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          2018.
          <article-title>Thu ngn at semeval-2018 task 3: Tweet irony detection with densely connected lstm and multi-task learning</article-title>
          .
          <source>In Proceedings of the 12th International Workshop on Semantic Evaluation(SemEval-</source>
          <year>2018</year>
          ).
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>