<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>U O U P V2 at HAHA 2019: BiGRU Neural Network Informed with Linguistic Features for Humor Recognition</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Reynier Ortega-Bueno</string-name>
          <email>reynier.ortega@cerpamid.co.cu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Paolo Rosso</string-name>
          <email>prosso@dsic.upv.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jose E. Medina Pagola</string-name>
          <email>jmedinap@uci.cu</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Pattern Recognition and Data Mining, Universidad de Oriente</institution>
          ,
          <addr-line>Santiago de Cuba</addr-line>
          ,
          <country country="CU">Cuba</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>PRHLT Research Center, Universitat Politecnica de Valencia</institution>
          ,
          <addr-line>Valencia</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Informatics Sciences</institution>
          ,
          <addr-line>Havana</addr-line>
          ,
          <country country="CU">Cuba</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>212</fpage>
      <lpage>221</lpage>
      <abstract>
        <p>Verbal humor is an illustrative example of how humans use creative language to produce funny content. We, as human being, access to humor or comicality with the purpose of projecting more complex meanings which, usually, represent a real challenge, not only for computers, but for humans as well. For that, understanding and recognizing humorous content automatically has been and continue being an important issue in Natural Language Processing (NLP) and even more in Cognitive Computing. In order to addressing this challenge, in this paper we describe our U O U P V2 system developed for participating in the second edition of the HAHA (Humor Analysis based on Human Annotation) task proposed at IberLEF 2019 Forum. Our starting point was the UO UPV system we participated in HAHA 2018 with some modi cation in its architecture. This year we explored other way to inform our Attention based Recurrent Neural Network model with linguistic knowledge. Experimental results show that our system achieves positive results ranked 7th out of 18 teams.</p>
      </abstract>
      <kwd-group>
        <kwd>Spanish Humor Classi cation</kwd>
        <kwd>BiGRU Neural Network</kwd>
        <kwd>Social Media</kwd>
        <kwd>Linguistic Features</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Natural language systems have to deal with many problems related with texts
comprehension, but these problems become very hard when creativity and
gurative devices are used in verbal and written communication. Human can easily
understand the underlying meaning of such texts but, for a computer to
disentangle the meaning of creative expressions such as irony and humor, it requires
much additional knowledge, and complex methods of reasoning.</p>
      <p>Humor is an illustrative example of how humans use creative language devices
in social communication. Humor not only serves to interchange information or
share implicit meaning, but also engages a relationship between those exposed
to the funny message. It can help people see the amusing side of problems and
can help them distance themselves from stressors. In the same way, it helps to
regulate ours emotions. Moreover, the manners in which people produce funny
content also reveal insight about their genre and personal traits.</p>
      <p>
        From a computational linguistics points of view many methods have been
proposed to tackle the task of recognizing humor from texts [
        <xref ref-type="bibr" rid="ref1 ref15 ref16 ref20 ref22">16,15,22,20,1</xref>
        ]. These
focus the attention on investigating linguistic features which can be considered as
markers and indicators of verbal humor. Also, due to the closely relation between
irony and humor, other works have studied these phenomena with the goal of
shed some light about what is common and what is distinct from linguistic point
of view. [
        <xref ref-type="bibr" rid="ref1 ref20 ref9">9,1,20</xref>
        ]
      </p>
      <p>
        Other methods focused on recognizing humor on messages from Twitter
based on supervised learning [
        <xref ref-type="bibr" rid="ref1 ref10 ref23 ref26 ref7">1,23,7,10,26</xref>
        ]. Deep Neural Networks based
methods have obtained competitive results in humor recognition on tweets. Among
them, Recurrent Neural Networks (RNN) models and their bidirectional
variant capture relevant information like long term dependencies. Also, attention
mechanism have become to be a strong tool, that has endowed the RNN model
with the capability of paying more attention to those elements that increase the
e ectiveness of these networks in several tasks of NLP [
        <xref ref-type="bibr" rid="ref13 ref24 ref27 ref28">13,24,28,27</xref>
        ]
      </p>
      <p>
        Previous researches have focused on English language; however, for Spanish,
the availability of corpora is scarce, which limits the amount of research done
for this language. HAHA 2018 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] became the rst shared task addressing the
problem of humor recognition in Spanish content in social media. Three systems
were proposed to solve the task. The best ranked approach used a model based on
the EvoMSA tool with uses a EvoDAG method (Evolutionary Algorithm) [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
This is a steady-state Genetic Programming system with tournament selection.
The main characteristic of EvoDAG is that the genetic operation is performed
at the root. EvoDAG was inspired by the geometric semantic crossover. The
second system [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] proposed a model based on Bidirectional Long Short Term
Memory (BiLSTM) neural networks with attention mechanism. The authors
used word2vec as input for the network and also a set of linguistically motivated
features (stylistic, structural and content, and a ective ones). The linguistic
information was combined with the deep representation learned in the next to the
last layer. The results showed that incorporating linguistic knowledge improves
the overall performance. The third system presented to the shared task, trained a
method based on SVM using a bag of character n-grams of sizes 1 to 8 character
models.
      </p>
      <p>
        Considering the advantages of linguistic features for capturing deep
linguistics aspects of the language also the capability of RNN for learning deep
representation and long term dependencies from sequential data, in this paper, we
present a method that combines the linguistic features used for humor
recognition and an Attention based Bidirectional Gated Recurrent Unit (BiGRU)
model. The system works with an attention layer which is applied at the top of
a BiGRU to generate a context vector for each word embedding which is then
fed to another BIGRU network. Finally, the learned representation is fed to a
Feed Forward Network (FNN) to classify whether the tweet is humorous or not.
Motivated by the results shown in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], we explore to incorporate the linguistic
information to the model through initial hidden state in the rst BiGRU layer.
      </p>
      <p>The paper is organized as follows. Section 2 presents a brief description of the
HAHA task. Section 3 introduces our system for humor detection. Experimental
results are subsequently discussed in Section 4. Finally, in Section 5 we present
our conclusions and attractive directions for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>HAHA Task and Dataset</title>
      <p>HAHA 2019 is the second edition of the rst shared task that addresses the
problem of recognizing humor in Spanish tweets. Similar to the rst edition,
in the HAHA 2019 task, two subtasks were proposed. The rst one, \Humor
Detection", aims at predicting whether a tweet is a joke or not (intended humor
by the author or not) and the second one \Funniness Score Prediction", is for
predicting a score value into 5-star ranking, supposing it is a joke.</p>
      <p>
        Participants were provided with a human-annotated corpus of 30000
Spanish tweets [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], divided in 24000 and 6000 for training and test respectively. The
training subset contains 9253 tweets with funny content and 14747 tweets
considered as non humorous. As could be observed, the classes distribution are slightly
unbalanced, hence a di culty is added to learn automatically the models.
      </p>
      <p>System evaluation metrics were used and reported by the organizers. They
use F1 measure on humor class for the subtask of \Humor Detection", moreover,
precision, recall and accuracy were also reported.
3</p>
      <p>Our U O</p>
    </sec>
    <sec id="sec-3">
      <title>U P V2 System</title>
      <p>
        The motivation behind of our approach are rstly to investigate the capability
of Recurrent Neural Network, speci cally, the Gated Recurrent Unit (GRU) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
to capture long-term dependencies. They showed to be able to learn the
dependencies in lengths of considerably large sequences. GRU networks simpli ed the
complexity of the LSTM networks [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], being computationally more e cient.
Moreover, attention mechanisms have endowed these networks with a
powerful strategy to increase their e ectiveness achieving better results [
        <xref ref-type="bibr" rid="ref12 ref24 ref27 ref29">27,29,24,12</xref>
        ].
Recently, the initial hidden state of the recurrent neural network has been a
successful explored way to inform the networks with contextual information [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
Secondly, humor recognition based on features engine and supervised learning
have been well studied in previous research papers. These features have proved
to be good indicators and markers of humor in text. For these reasons, in this
approach we propose a method that enrich the Attention-based GRUs model
with linguistic knowledge which is passed to the network using the initial
hidden state. In Section 3.1 we describe the tweets preprocessing phase. Following,
in Section 3.2 we present the linguistic features used for encoding humorous
content. Finally, in Section 3.3 we introduce the neural network model and the
way in which linguistic features are introduced. The Figure 1 shows the overall
architecture of our system.
In the preprocessing step, the tweets are cleaned. Firstly, the emoticons, urls,
hashtags, mentions, twitter-reserve words as RT (for retweet) and FAV (for
favorite) are recognized and replaced by a corresponding wildcard which encodes
the meaning of these special words. Afterwards, tweets are morphologically
analyzed by FreeLing [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. In this way, for each resulting token, its lemma is assigned.
Then, the tweets are represented as vectors with a word embedding model. This
embedding was generated by using the FastText algorithm [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] from the Spanish
Billion Words Corpus [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and an in-house background corpus of 9 millions of
Spanish tweets.
3.2
      </p>
      <sec id="sec-3-1">
        <title>Linguistic Features</title>
        <p>
          In our work, we explored several linguistic features useful for humor recognition
in texts [
          <xref ref-type="bibr" rid="ref1 ref15 ref16 ref20 ref22 ref5">16,15,22,20,1,5</xref>
          ] which can be grouped in three main categories: Stylistic,
Structural and Content, and A ective. Particularly, we considered stylistic
features such as: length, dialog markers, quotation, punctuation marks, emphasized
words, url, emoticons, hashtag, etc. Features for capturing lexical and semantic
ambiguity, sexual, obscene, animal and human-related terms, etc., were
considered as Content and Structural. Finally, due to the relation of humor with
expressions of sentiment and emotions we used features for capturing a ects,
attitudes, sentiments and emotions. For more details about the features see [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>
          Notice that, our proposal did not consider the positional features used in
[
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. Moreover taking into account the close relation between irony and humor
and motivated by the results presented in [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] we include psycho-linguistic
features extracted from the LIWC [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. This resource contains about 4,500 entries
distributed in 65 categories. Speci cally, for this work we decided to use all
categories as independent features.
        </p>
        <p>Taking into account the previous features, we represent each message by one
vector VTi with dimensionality equal to 165. Also, in order to reduce and improve
this representation we applied a feature selection method. Speci cally we use the
Wilcoxon Rank-sum test for paired samples. By using this test all features were
ranked considering their p value.
3.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Recurrent Network Architecture</title>
        <p>We propose a model that consists in a BiGRU neural network at the word level.
Each time step t the BiGRU gets as input a word vector wt. Afterward, an
attention layer is applied over each hidden state ht. The attention weights are
learned using the concatenation of the current hidden state ht of the BiGRU and
the past hidden state st 1 in the second BiGRU layer. Finally, the target humor
of the tweet is predicted by an FFN with one hidden layer, and an output layer
with two neurons. Our overall architecture is described in the following sections.
3.4</p>
      </sec>
      <sec id="sec-3-3">
        <title>First BiGRU Layer</title>
        <p>In NLP problems, standard GRU receives sequentially (left to right order) at
each time step a word embedding wt and produces a hidden state ht. Each
hidden state ht is calculated as follow:
zt = (W (z)xt + U (z)ht 1 + b(z))</p>
        <p>(W (r)xt + U (r)ht 1 + b(r))
rt =
h^t = tanh(W (h^)xt + U (h^)ht 1 + b(h^))
ht = zt ht 1 + (1 zt) h^t
(update gate)
(reset gate)
(memory cell)</p>
        <p>Where all W ( ), U ( ) and b( ) are parameters to be learned during training.
Function is the sigmoid function and stands for element-wise multiplication.</p>
        <p>Bidirectional GRU, on the other hand, makes the same operations as
standard GRU, but processing the incoming text in a left-to-right and a right-to-left
order in parallel. Thus, it outputs two hidden state at each time step !ht and
ht . The proposed method uses a BiGRU network which considers each new
hid!
den state as the concatenation of these two h^t = [ht ; ht ]. The idea behind this
BiGRU layer is to capture long-range and backwards dependencies
simultaneously. In this layer is where the linguistic information is passed throughout the
model. We initialized both initial hidden state [h!0 = g(Ti); h0 = g(Ti)] where
g(:) receives a tweet and returns a vector which encodes contextual and linguistic
knowledge g(Ti) = VTi .
3.5</p>
      </sec>
      <sec id="sec-3-4">
        <title>Attention Layer</title>
        <p>With an attention mechanism we allow the BiGRU to decide which segment
of the sentence should \attend". Importantly, we let the model learn what to
attend on the basic of the input sentence and what it has produced so far.
Let H 2 R2 Nh T the matrix of hidden states [h^1; h^2; : : : ; h^T ] produced by the
rst BiGRU layer, where Nh is the size of the hidden state and T is the length of
the given sequence. The goal is then to derive a context vector ct that captures
relevant information and feeds it as input to the next BiGRU layer. Each ct is
calculate as follow:</p>
        <p>T
ct = X
t0=1
t;t0 h^t0</p>
        <p>t;i =
(si; hj ) = VaT
tanh(Wa
exp( (st 1; h^i))
T
X exp( (st 1; h^j ))
j=1
[si; h^j ])</p>
        <p>Where Wa and Va are the trainable attention parameters, st 1 is the past
hidden state of the second BiGRU layer and h^t is the current hidden state. The
idea of the concatenation layer is to take into account not only the input sentence
but also the past hidden state to produce the attention weights.
3.6</p>
      </sec>
      <sec id="sec-3-5">
        <title>Second BiGRU Layer</title>
        <p>The goal of this layer is to obtain a deep dense representation of the message
with the intention to determine whether the tweet is humorous or not. This
network at each time step receives the context vector ct which is propagated until
the nal hidden state sT . This vector is a high level representation of the tweet.
Afterwards, it is passed to a feed forward network (FFN) with 3 hidden layers,
and we use a softmax layer at the end as follow:
y^ = sof tmax(W0 dense1 + b0)
dense1 = relu(W1 dense2 + b1)
dense2 = relu(W2 dense3 + b2)</p>
        <p>dense3 = relu(W3 sT + b3)</p>
        <p>Where Wj and bj (j = 0; :::3) denote the weight matrices and bias vectors
for the last three layers with a softmax at the end. Finally, cross entropy is used
as the loss function, which is de ned as:</p>
        <p>L = X yi log(y^i)</p>
        <p>i</p>
        <p>Where yi is the ground true classi cation of the tweet (humor vs. not humor)
and y^i is the predicted value by the model.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments and Results</title>
      <p>For tuning some parameters of the proposed model we used a strati ed k-fold
cross validation with 5 partitions on the training dataset. At this time, a
features selection process was not performed, therefore we consider all linguistic
features. During the training phase, we xed some hyper-parameters, concretely:
the batch size =256, epochs=10, units of the GRU cell was de ned as 256,
optimizer=\Adam" and dropout in the GRU cells=0.3. After that, we evaluated
di erent subsets of linguistic features, particularly, ve setting of features were
explored. The considered subsets were: N o F ea (linguistic information is not
considered), 64 F ea (the 64 best ranked features according to p value), 128 F ea
(the 128 best ranked features according to p value), 141 F ea (all features with
p value 0:05) and All F ea (all linguistic features).</p>
      <p>As can be observed in the Table 1, a slight improvement is obtained when
linguistic features were passed to the model. Particularly, the subset of F ea 128
achieved the best F1 score (F1 h=0.785) in the humor class. Also, when linguistic
information was missing a gradual drop of 3.5%, in term of F1-score was observed
in the humor class.</p>
      <p>Regarding o cial evaluation and results, for the system's submission,
participants were allowed to send more than one model till a maximum of 10 possible
runs. Taking into account the results showed in the Table 1 we submitted three
runs. The di erence among them is the number of linguistic features
considered for informing the U O U P V2 model. In Run1 we use the subset of features
F ea 64, for Run2 we used the subset of features F ea 128 while in Run3 the
subset of features F ea 141 was considered. We achieved 0.765, 0.773, and 0.765,
in terms of F1 score in the humor class, for Run1, Run2 and Run3 respectively.
These values are consistent with the results obtained in the training phase.
In this paper we presented our modi cation of the UO UPV system (U O U P V2)
for the task of humor recognition (HAHA) at IberLEF 2019. We only
participated in the \Humor Detection" subtask and ranked 7th out of 18 team. Our
proposal combines linguistic features with an Attention-based BiGRU Neural
Network. The model consists of a Bidirectional GRU neural network with an
attention mechanism that allows to estimate the importance of each word and
then, this context vector is used with another BiGRU model to estimate whether
the tweet is humorous or not. Regarding the feature selection, the best result was
achieved when the 128 best ranked (according to p value) features were
considered. The results, also shown that adding linguistic information through initial
hidden state caused an improvement in the e ectiveness based on F1-measure.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>The work of the second author was partially funded by the Spanish MICINN
under the research project MISMIS-FAKEnHATE on Misinformation and
Miscommunication in social media: FAKE news and HATE speech
(PGC2018-096212B-C31).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Barbieri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saggion</surname>
          </string-name>
          , H.:
          <article-title>Automatic Detection of Irony and Humour in Twitter</article-title>
          .
          <source>In: Fifth International Conference on Computational Creativity</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Enriching Word Vectors with Subword Information</article-title>
          .
          <source>Transactions of the ACL. 5</source>
          ,
          <issue>135</issue>
          {
          <fpage>146</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cardellino</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Spanish Billion Words Corpus and Embeddings (</article-title>
          <year>2016</year>
          ). [Online]. Available: http://crscardellino.me/SBWCE/.
          <source>Retrieved May 4</source>
          ,
          <year>2018</year>
          , http: //crscardellino.me/SBWCE/
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of the HAHA Task : Humor Analysis based on Human Annotation at IberEval 2018</article-title>
          . In: Rosso,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Mart</surname>
          </string-name>
          <string-name>
            <given-names>nez</given-names>
            , R.,
            <surname>Montalvo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Carrillo De Albornoz Cuadrado</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . (eds.)
          <source>Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval</source>
          <year>2018</year>
          )
          <article-title>co-located with 34th Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2018</year>
          ). pp.
          <volume>187</volume>
          {
          <fpage>194</fpage>
          . CEUR-WS.org, Sevilla, Spain (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moncecchi</surname>
          </string-name>
          , G.:
          <article-title>Is This a Joke? Detecting Humor in Spanish Tweets</article-title>
          .
          <source>In: Ibero-American Conference on Arti cial Intelligence</source>
          . pp.
          <volume>139</volume>
          {
          <issue>150</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Castro</surname>
          </string-name>
          , Santiago and Chiruzzo, Luis and Rosa, Aiala and Garat, Diego and Moncecchi, G.:
          <article-title>A Crowd-Annotated Spanish Corpus for Humor Analysis</article-title>
          .
          <source>In: Proceedings of SocialNLP</source>
          <year>2018</year>
          ,
          <article-title>The 6th International Natural Language Processing for Social Media (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Cattle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bay</surname>
            ,
            <given-names>C.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kong</surname>
          </string-name>
          , H.:
          <article-title>SRHR at SemEval-2017 Task 6: Word Associations for Humour Recognition</article-title>
          .
          <source>In: 11th International Workshop on Semantic Evaluations (SemEval-2017)</source>
          . pp.
          <volume>401</volume>
          {
          <issue>406</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Chung</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Gulcehre,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          :
          <article-title>Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling</article-title>
          .
          <source>CoRR abs/1412</source>
          .3 (
          <issue>2014</issue>
          ), http://arxiv.org/abs/1412.3555
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Gibbs</surname>
            ,
            <given-names>R.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bryant</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Colston</surname>
            ,
            <given-names>H.L.</given-names>
          </string-name>
          :
          <article-title>Where is the humor in verbal irony ?</article-title>
          <source>Humor</source>
          <volume>27</volume>
          (
          <issue>4</issue>
          ),
          <volume>575</volume>
          {
          <fpage>595</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toner</surname>
          </string-name>
          , G.:
          <article-title>QUB at SemEval-2017 Task 6: Cascaded Imbalanced Classi cation for Humor Analysis in Twitter</article-title>
          .
          <source>In: 11th International Workshop on Semantic Evaluations (SemEval-2017)</source>
          . pp.
          <volume>380</volume>
          {
          <issue>384</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Hochreiter</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural computation 9(8)</source>
          ,
          <volume>1735</volume>
          {
          <fpage>1780</fpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Sentiment Analysis Model Based on Structure Attention Mechanism</article-title>
          .
          <source>In: UK Workshop on Computational Intelligence</source>
          . pp.
          <volume>17</volume>
          {
          <fpage>27</fpage>
          . Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Luong</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pham</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.:
          <article-title>E ective approaches to attention-based neural machine translation</article-title>
          .
          <source>arXiv preprint arXiv:1508.04025</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Mar</surname>
          </string-name>
          <article-title>a del Pilar Salas-Zarate , Mario Andres Paredes-Valverde , Miguel Angel Rodriguez-Garc a , Rafael Valencia-Garc a</article-title>
          ,
          <string-name>
            <surname>G.A.H.</surname>
          </string-name>
          :
          <article-title>Automatic Detection of Satire in Twitter: A psycholinguistic-based approach</article-title>
          .
          <source>Knowledge-Based Systems</source>
          (
          <year>2017</year>
          ), http://dx.doi.org/10.1016/j.knosys.
          <year>2017</year>
          .
          <volume>04</volume>
          .009
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pulman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Characterizing Humour:
          <article-title>An Exploration of Features in Humorous Texts</article-title>
          .
          <source>In: International Conference on Intelligent Text Processing and Computational Linguistics</source>
          . pp.
          <volume>337</volume>
          {
          <issue>347</issue>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strapparava</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Learning to laugh (automatically): computational models for humor recognition</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>22</volume>
          (
          <issue>2</issue>
          ),
          <volume>126</volume>
          {
          <fpage>142</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Ortega-Bueno</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Mun~iz,
          <string-name>
            <given-names>C.E.</given-names>
            ,
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Medina-Pagola</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.E.: UO UPV</surname>
          </string-name>
          :
          <article-title>Deep Linguistic Humor Detection in Spanish Social Media</article-title>
          . In: Rosso,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Mart</surname>
          </string-name>
          <string-name>
            <given-names>nez</given-names>
            , R.,
            <surname>Montalvo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Carrillo-de Albornoz</surname>
          </string-name>
          , J. (eds.)
          <source>Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval</source>
          <year>2018</year>
          )
          <article-title>co-located with 34th Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2018</year>
          ). pp.
          <volume>203</volume>
          {
          <fpage>213</fpage>
          . CEUR-WS.org, Sevilla, Spain (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Padro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stanilovsky</surname>
          </string-name>
          , E.:
          <article-title>FreeLing 3.0: Towards Wider Multilinguality</article-title>
          .
          <source>In: Proceedings of the (LREC</source>
          <year>2012</year>
          )
          <article-title>(</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Pennebaker</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Francis</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Booth</surname>
          </string-name>
          , R.J.:
          <article-title>Linguistic inquiry and word count: LIWC 2001</article-title>
          . Mahway: Lawrence Erlbaum Associates
          <volume>71</volume>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Reyes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buscaldi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>From humor recognition to irony detection: The gurative language of social media</article-title>
          .
          <source>Data and Knowledge Engineering</source>
          <volume>74</volume>
          ,
          <issue>1</issue>
          {
          <fpage>12</fpage>
          (
          <year>2012</year>
          ), http://dx.doi.org/10.1016/j.datak.
          <year>2012</year>
          .
          <volume>02</volume>
          .005
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Salgado</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.:</given-names>
          </string-name>
          <article-title>INGEOTEC at IberEval 2018 Task HaHa: TC and EvoMSA to Detect and Score Humor in Texts</article-title>
          . In: Rosso,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Mart</surname>
          </string-name>
          <string-name>
            <given-names>nez</given-names>
            , R.,
            <surname>Montalvo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Carrillo De Albornoz Cuadrado</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . (eds.)
          <source>Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval</source>
          <year>2018</year>
          )
          <article-title>co-located with 34th Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2018</year>
          ). pp.
          <volume>195</volume>
          {
          <fpage>202</fpage>
          . CEUR-WS.org, Sevilla, Spain (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Sjobergh</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Araki</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Recognizing Humor Without Recognizing Meaning</article-title>
          .
          <source>In: International Workshop on Fuzzy Logic and Applications</source>
          . pp.
          <volume>469</volume>
          |-
          <fpage>476</fpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Turcu</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alexa</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Amarandei</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Herciu</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scutaru</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iftene</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>#WarTeam at SemEval-2017 Task 6: Using Neural Networks for Discovering Humorous Tweets</article-title>
          .
          <source>In: 11th International Workshop on Semantic Evaluations (SemEval-2017)</source>
          . pp.
          <volume>407</volume>
          {
          <issue>410</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <article-title>Others: Attention-based lstm for aspect-level sentiment classi cation</article-title>
          .
          <source>In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>606</volume>
          {
          <issue>615</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Wenke</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fleming</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Contextual Recurrent Neural Networks</article-title>
          . CoRR abs/
          <year>1902</year>
          .0 (
          <issue>2019</issue>
          ), http://arxiv.org/abs/
          <year>1902</year>
          .03455
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pedersen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Duluth at SemEval-2017 Task 6: Language Models in Humor Detection</article-title>
          .
          <source>In: 11th International Workshop on Semantic Evaluations (SemEval2017)</source>
          . pp.
          <volume>385</volume>
          {
          <fpage>389</fpage>
          . No.
          <volume>2</volume>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Attention Based LSTM for Target Dependent Sentiment Classi cation</article-title>
          .
          <source>In: AAAI</source>
          . pp.
          <volume>5013</volume>
          {
          <issue>5014</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smola</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hovy</surname>
          </string-name>
          , E.:
          <article-title>Hierarchical attention networks for document classi cation</article-title>
          .
          <source>In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          . pp.
          <volume>1480</volume>
          {
          <issue>1489</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          :
          <article-title>Attention-based LSTM with Multi-task Learning for Distant Speech Recognition</article-title>
          .
          <source>Proc. Interspeech</source>
          <year>2017</year>
          pp.
          <volume>3857</volume>
          {
          <issue>3861</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>