<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Generating Document Embeddings for Humor Recognition using Tensor Decomposition</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Evangelos Papalexakis</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Hong Kong University of Science and Technology Clear Water Bay</institution>
          ,
          <country>Kowloon Hong Kong</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of California, Riverside 900 University Ave Riverside</institution>
          ,
          <addr-line>CA 92521</addr-line>
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>151</fpage>
      <lpage>159</lpage>
      <abstract>
        <p>This paper details our submission to the HAHA 2019 (5) group task on humor recognition. We propose a novel humor recognition system based on tensor embeddings, capable of being trained without the need for any external training corpora. Our model achieves an F1-score of 0.736 on a binary humor classi cation task and a root-mean-squarederror of 0.963 on a humor scoring task, both using a Spanish-language Twitter corpus (2). While our experiments are performed on Spanish documents, our approach is truly language agnostic and can be applied to any language with minimal adaptation.</p>
      </abstract>
      <kwd-group>
        <kwd>Computational Humor</kwd>
        <kwd>Humor Recognition</kwd>
        <kwd>Tensor Decomposition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Humor is an integral part of human interaction. It can be used to defuse tense
situations, increase likeability, or even for pure entertainment. As such, the
automatic recognition of humor represents an important step for natural
humancomputer interaction (
        <xref ref-type="bibr" rid="ref18">18</xref>
        ). While early works tended to approach humor
recognition as a binary classi cation task (11; 21), ne-grained humor evaluation has
gained recent attention (3; 14; 18).
      </p>
      <p>
        This paper details our submission to the HAHA 2019 (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) group task on
humor recognition. We propose a novel humor recognition system based on tensor
embeddings, capable of being trained without the need for any external training
corpora. While most humor recognition works focus on English corpora, HAHA
2019 utilizes a Spanish-language corpora (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ). As such, our model prioritizes
language-agnosticism and thus can be applied to any language with minimal
adaptation. The code used to create our model is available for download1.
2
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <sec id="sec-2-1">
        <title>Humor Recognition</title>
        <p>
          Although humor recognition has typically been framed as a binary classi cation
task (1; 11; 21), recent works have moved beyond mere humor detection and
toward humor evaluation by framing humor recognition as a relative ranking
task (3; 14; 18). Generally, these works have focused on documents generated
around a common prompt such as cartoon captions (
          <xref ref-type="bibr" rid="ref18">18</xref>
          ) or Twitter Hashtag
Wars (3; 14).
        </p>
        <p>
          Humor is a complex phenomenon, incorporating aspects of phonology, style,
semantics, and word-choice. As such, existing humor recognition works have
tended to use a variety of features designed to capture this complexity. (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
extract acoustic features from sitcom audio tracks while (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) incorporate \phonetic
embeddings" generated using a character-to-phoneme LSTM encoder-decoder.
(
          <xref ref-type="bibr" rid="ref11">11</xref>
          ) look for alliteration, rhyming, negative sentiment, and adult slang to aid
humor recognition while (
          <xref ref-type="bibr" rid="ref17">17</xref>
          ) add emotional scenarios. Inspired by incongruity
theory (
          <xref ref-type="bibr" rid="ref16">16</xref>
          ), several works attempt to measure the amount of incongruity in a
document using various lexical similarity metrics (4; 18; 21). Other works
represent semantics using word (1; 6) or document (
          <xref ref-type="bibr" rid="ref21">21</xref>
          ) embeddings as model inputs.
(
          <xref ref-type="bibr" rid="ref15">15</xref>
          ) apply document centrality as de ned by the graph-based text
summarization model LexRank (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ). A more basic approach is to use word frequency (
          <xref ref-type="bibr" rid="ref11">11</xref>
          )
and n-gram probability (
          <xref ref-type="bibr" rid="ref20">20</xref>
          ) as indications of humor.
        </p>
        <p>
          The majority of humor recognition works focus on English language
humor. Given the dependence of many models on language dependant resources,
adapting these models for non-English contexts can be challenging. (
          <xref ref-type="bibr" rid="ref21">21</xref>
          ) and
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) make use of the semantic ontology WordNet (
          <xref ref-type="bibr" rid="ref12">12</xref>
          ). Similarly, (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) train their
phonetic embeddings on the CMU pronouncing dictionary (
          <xref ref-type="bibr" rid="ref10">10</xref>
          ). More
fundamentally, many existing systems achieve their results using high quality, large
scale pre-trained word embedding models such as Google word2vec2 or Stanford
GloVe 3. As such, adapting such models for use with other languages would be
dependant on the availability of equivalent resources.
2.2
        </p>
        <p>
          Tensor Embeddings
State-of-the-art document embedding approaches like doc2vec (
          <xref ref-type="bibr" rid="ref9">9</xref>
          ) or sent2vec
(
          <xref ref-type="bibr" rid="ref13">13</xref>
          ) are capable of encoding the meaning of a document but often require large
        </p>
        <sec id="sec-2-1-1">
          <title>1 https://github.com/acattle/HumourTools/</title>
        </sec>
        <sec id="sec-2-1-2">
          <title>2 https://code.google.com/archive/p/word2vec/</title>
        </sec>
        <sec id="sec-2-1-3">
          <title>3 https://nlp.stanford.edu/projects/glove/</title>
          <p>
            amounts of training data. By comparison, tensor decomposition is capable of
generating low-rank embeddings of sentences that capture the similarity of
contextual patterns without the need for large training corpora (
            <xref ref-type="bibr" rid="ref8">8</xref>
            ).
          </p>
          <p>
            Such \tensor embeddings", combined with a label propagation technique,
have been shown e ective in semi-supervised fake news detection (
            <xref ref-type="bibr" rid="ref8">8</xref>
            ). The fact
that high quality embeddings can be created from relatively small corpora makes
tensor embeddings an attractive option for training models for low-resource
languages where large scale corpora may not be available.
3
3.1
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Method</title>
      <sec id="sec-3-1">
        <title>Tensor Embeddings</title>
        <p>For each document in the corpus we compute a tensor embedding based on word
co-occurrence. That is, for a corpus D = {s1; s2; : : : ; sD} with D sentences, we
rst extract a vocabulary w1; w2; : : : ; wV , where V is the number of words. For
each sentence s in D, we count the word-word co-occurrences within a small
window H. This results in a frequency matrix Ws ∈ ZV ×V , where Z denotes
the set of integers. In particular, Ws(i; j) indicates the frequency that word wi
and wj co-occur in s within the window H. This allows us to encode the lexical
patterns of s in Ws. All Ws are then stacked, creating a three-dimensional tensor
W ∈ ZV ×V ×D. The objective of tensor decomposition is to nd an approximation
W^ of W such that:</p>
        <p>
          R
^
W = Q vr ⊗ vr ⊗ dr;
r=1
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
where vr ∈ RV , dr ∈ RD, R is the pre-de ned rank parameter, and ⊗ is the outer
product, namely, vr ⊗ vr ⊗ dr being a three-dimensional tensor, and
vr ⊗ vr ⊗ dr(i; j; k) = vr(i) ⋅ vr(j) ⋅ dr(k):
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
        </p>
        <p>In particular, C = [d1; d2; : : : ; dR] ∈ RD×R, where the s-row of C is the
embedding vector of sentence s.</p>
        <p>
          The low-rank sentence embeddings are calculated using the alternating least
squares method of the CANDECOMP/PARAFAC tensor decomposition
implementation in Matlab tensor toolbox 4 (
          <xref ref-type="bibr" rid="ref19">19</xref>
          ). H was set as 5 and R set as 50.
3.2
        </p>
        <p>
          Humor Recognition
After extracting tensor embeddings for each document, we performed a simple
baseline experiment using the (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) corpora. We trained both classi cation and
regression models in a supervised manner on the tensor embeddings corresponding
to the documents in the training set along with their labels. These models were
        </p>
        <sec id="sec-3-1-1">
          <title>4 https://www.tensortoolbox.org/</title>
          <p>
            then used to predict the labels and scores of documents in the test set as part
of HAHA 2019's (
            <xref ref-type="bibr" rid="ref5">5</xref>
            ) binary classi cation and humor scoring tasks, respectively.
          </p>
          <p>
            Second, inspired by (
            <xref ref-type="bibr" rid="ref8">8</xref>
            )'s work on fake news detection, we performed applied
label propagation (
            <xref ref-type="bibr" rid="ref22">22</xref>
            ), a semi-supervised labeling technique, to the extracted
tensor embeddings. In this method, labels are propagated to their neighbors in
a weight average way.
          </p>
          <p>
            Third, inspired by (
            <xref ref-type="bibr" rid="ref15">15</xref>
            ), we compute a lexical centrality feature. While (
            <xref ref-type="bibr" rid="ref15">15</xref>
            )
uses a graph-based approach, we instead take a vector-space approach.
Operating on the assumption that funnier documents are more central, we calculate
a tensor embedding centroid as the average of all the tensor embeddings. The
Euclidean distance of each tensor embedding from the centroid is then taken
as an indication of humor. While this distance-based metric is a type of
humor score in-and-of-itself, it is not appropriate for the HAHA 2019 (
            <xref ref-type="bibr" rid="ref5">5</xref>
            ) humor
scoring task. While the (
            <xref ref-type="bibr" rid="ref2">2</xref>
            ) dataset rates each document on a ve point scale,
with funnier tweets receiving higher scores, this distance-based metric has no
upper bound and is inversely proportional to the perceived humor. Instead, we
trained further so-called \enhanced" classi er and regression models on a
combination our lexical centrality, label propagation, and tensor embeddings. We
hoped that the addition of the label propagation and lexical centrality features
would help capture complex pattern that may not be extracted from the raw
tensor embeddings.
          </p>
          <p>Except where otherwise noted, all results reported in this paper were obtained
using the Random Forest classi er and regressor implementations in scikit-learn5
with default parameters. These results are comparable to results we obtained
using Support Vector Machine and Extra Trees models. Label propagation was
also performed using the implementation in scikit-learn with default parameters.
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results and Discussion</title>
      <p>The performance of our humor recognition models is shown in Table 1. Despite
the simplicity of our models, we are able to achieve reasonable performance.
Our tensor embedding only models outperforms both our label propagation and
enhanced models across the board. Despite the performance of a similar model</p>
      <sec id="sec-4-1">
        <title>5 https://scikit-learn.org/</title>
        <p>
          on fake news detection (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ), our label propagation model performs surprisingly
poorly; correctly identifying less than 10% of true positives in the test set.
        </p>
        <p>
          It is important to note at while the (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) dataset is in Spanish, our tensor
embeddings, and thus all of our models, are completely language agnostic. Because
our embeddings are generated only from the dataset itself, without the need for
any external training corpora, our models can be readily applied to any corpus
regardless of language. The only requirement is for reliable word segmentation,
which may be an issue for languages with optional or inconsistent whitespace
(e.g. Chinese).
        </p>
        <p>
          Another advantage of our approach is its relative simplicity. The most
computationally expensive aspect of our model is computing the tensor decomposition.
This makes approach quite scalable when compared with complex
sequencedbased neural models such as those used by (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) or (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) and better-suited than
such systems for small datasets.
        </p>
        <p>
          Part of the reason for the poor label propagation performance may be the
(
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) dataset's unbalanced nature. Of the 24,000 training examples, only 9,253
have a positive label. Since our label propagation system employed a K Nearest
Neighbors kernel (the default in scikit-learn), the larger number of negative
humor labels may have overwhelmed any positive labels. A di erent choice of
kernel or further hyperparameter tuning may lead to better results.
        </p>
        <p>
          The semi-supervised nature of label propagation may further explain this
lack of performance. Given the di culty in obtaining reliable humor judgments,
it should come as no surprise that most humor datasets tend to be relatively
small. Some datasets, like Pun of the Day (
          <xref ref-type="bibr" rid="ref21">21</xref>
          ), are as small as a few thousand
documents. With so few training examples, fully supervised approaches run the
risk of either failing to extract meaningful patterns or placing too much
emphasis on patterns in the training set that may not generalize. Thus, we expected
a semi-supervised approach, like label propagation, would help mitigate these
risks. However, the (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) dataset used in this paper is relatively large for a humor
dataset, containing 24,000 training examples and 6,000 test, diminishing these
advantages.
        </p>
        <p>
          The performance of label propagation may also be a ecting the performance
of the enhanced model. Another potential limitation the enhanced model is our
lexical centrality feature. One major di erence between the (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) dataset and the
dataset used by (
          <xref ref-type="bibr" rid="ref15">15</xref>
          ) is that (
          <xref ref-type="bibr" rid="ref15">15</xref>
          ) compared documents generated in response to
a common prompt (i.e. captions submitted to the same New Yorker Cartoon
Caption Contest). By comparison, (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) uses tweets sampled from Twitter with
no regard for common prompts. As such, our centrality assumption may not
hold. One possible improvement would be to cluster the tweets and computing
multiple centroids. Unfortunately, we were unable to run a test using lexical
centrality only due to limitations imposed as part of the HAHA 2019 (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) group
task in terms of time and numbers of submissions.
        </p>
        <p>
          Another potential area for improvement is related to the hyper parameters
used to generate our tensor embeddings. The window size of 5 and tensor rank
of 50 was chosen empirically due to their high performance on smaller,
Englishlanguage datasets. The larger size of the (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) dataset may allow for higher ranks
while di erences between English and Spanish may favor di erent window sizes.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this paper we have shown that tensor embeddings are capable of producing
reasonable performance for both binary humor classi cation and humor scoring.
Furthermore, we identify some key advantages of this tensor embedding approach
including its simplicity and language agnosticism. Finally, we o er several
potential avenues for improving the performance of our models including better
hyperparamter tuning.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Bertero</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fung</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A long short-term memory framework for predicting humor in dialogues</article-title>
          .
          <source>In: 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          ,
          <fpage>2016</fpage>
          -Proceedings of the Conference. pp.
          <volume>130</volume>
          {
          <issue>135</issue>
          (
          <year>2016</year>
          ), https://pdfs. semanticscholar.org/015c/26ca824f9d20b6523f44e6b9dc3b3dd65b2d.pdf
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moncecchi</surname>
          </string-name>
          , G.:
          <article-title>A crowdannotated spanish corpus for humor analysis</article-title>
          .
          <source>In: Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media</source>
          . pp.
          <volume>7</volume>
          {
          <issue>11</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Cattle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ma</surname>
          </string-name>
          , X.:
          <article-title>E ects of semantic relatedness between setups and punchlines in twitter hashtag games</article-title>
          .
          <source>In: Proceedings of the Workshop on Computational Modeling of People's Opinions, Personality, and Emotions in Social Media (PEOPLES)</source>
          . pp.
          <volume>70</volume>
          {
          <fpage>79</fpage>
          .
          <string-name>
            <surname>The</surname>
            <given-names>COLING</given-names>
          </string-name>
          2016
          <string-name>
            <given-names>Organizing</given-names>
            <surname>Committee</surname>
          </string-name>
          , Osaka,
          <source>Japan (Dec</source>
          <year>2016</year>
          ), https://www.aclweb.org/anthology/ W16-4308
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Cattle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ma</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Recognizing humour using word associations and humour anchor extraction</article-title>
          .
          <source>In: Proceedings of the 27th International Conference on Computational Linguistics</source>
          . pp.
          <year>1849</year>
          {
          <year>1858</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          , Santa Fe, New Mexico, USA (Aug
          <year>2018</year>
          ), https: //www.aclweb.org/anthology/C18-1157
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etcheverry</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prada</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : Overview of HAHA at IberLEF 2019:
          <article-title>Humor analysis based on human annotation</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          , CEUR-WS, Bilbao,
          <source>Spain (9</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Donahue</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Romanov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>HumorHawk at SemEval-2017 task 6: Mixing meaning and sound for humor recognition</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval2017)</source>
          . pp.
          <volume>98</volume>
          {
          <fpage>102</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (Aug
          <year>2017</year>
          ). https://doi.org/10.18653/v1/
          <fpage>S17</fpage>
          -2010, https://www. aclweb.org/anthology/S17-2010
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Erkan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Radev</surname>
            ,
            <given-names>D.R.</given-names>
          </string-name>
          : Lexrank:
          <article-title>Graph-based lexical centrality as salience in text summarization</article-title>
          .
          <source>Journal of Arti cial Intelligence Research</source>
          <volume>22</volume>
          ,
          <volume>457</volume>
          {
          <fpage>479</fpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Hosseinimotlagh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papalexakis</surname>
            ,
            <given-names>E.E.</given-names>
          </string-name>
          :
          <article-title>Unsupervised content-based identi cation of fake news articles with tensor decomposition ensembles</article-title>
          .
          <source>In: WSDM 2018 Workshop on Misinformation and Misbehavior Mining on the Web (MIS2)</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of sentences and documents</article-title>
          .
          <source>In: Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32</source>
          . pp.
          <article-title>II{1188{II{1196. ICML'14, JMLR</article-title>
          .org (
          <year>2014</year>
          ), http://dl.acm.org/citation.cfm?id=
          <volume>3044805</volume>
          .
          <fpage>3045025</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Lenzo</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>The CMU pronouncing dictionary (</article-title>
          <year>1998</year>
          ), http://www.speech. cs.cmu.edu/cgi-bin/cmudict
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strapparava</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Making computers laugh: Investigations in automatic humor recognition</article-title>
          .
          <source>In: Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>531</volume>
          {
          <fpage>538</fpage>
          .
          <article-title>Association for Computational Linguistics (</article-title>
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          :
          <article-title>WordNet: a lexical database for English</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>38</volume>
          (
          <issue>11</issue>
          ),
          <volume>39</volume>
          {
          <fpage>41</fpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Pagliardini</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jaggi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Unsupervised Learning of Sentence Embeddings using Compositional n-Gram Features</article-title>
          . In: NAACL 2018 -
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Potash</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Romanov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>SemEval-2017 task 6: #HashtagWars: Learning a sense of humor</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017)</source>
          . pp.
          <volume>49</volume>
          {
          <fpage>57</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (Aug
          <year>2017</year>
          ). https://doi.org/10.18653/v1/
          <fpage>S17</fpage>
          -2004, https://www.aclweb. org/anthology/S17-2004
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Radev</surname>
            ,
            <given-names>D.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stent</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tetreault</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pappu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iliakopoulou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanfreau</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , de Juan,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Vallmitjana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Jaimes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Manko</surname>
          </string-name>
          , R.:
          <article-title>Humor in collective discourse: Unsupervised funniness detection in the new yorker cartoon caption contest</article-title>
          .
          <source>CoRR abs/1506</source>
          .08126 (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Raskin</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Semantic theory of humor</article-title>
          .
          <source>In: Semantic Mechanisms of Humor</source>
          . Springer (
          <year>1985</year>
          ). https://doi.org/10.1007/
          <fpage>978</fpage>
          -94-009-6472-34, http: //dx.doi.org/10.1007/
          <fpage>978</fpage>
          -94
          <source>-009-6472-3 4</source>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Reyes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buscaldi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>From humor recognition to irony detection: The gurative language of social media</article-title>
          .
          <source>Data &amp; Knowledge Engineering</source>
          <volume>74</volume>
          ,
          <issue>1</issue>
          {
          <fpage>12</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Shahaf</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horvitz</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manko</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>: Inside jokes: Identifying humorous cartoon captions</article-title>
          .
          <source>In: Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>1065</volume>
          {
          <fpage>1074</fpage>
          . KDD '15,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2015</year>
          ). https://doi.org/10.1145/2783258.2783388, http://doi.acm.
          <source>org/10</source>
          . 1145/2783258.2783388
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Sidiropoulos</surname>
            ,
            <given-names>N.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lathauwer</surname>
            ,
            <given-names>L.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Papalexakis</surname>
            ,
            <given-names>E.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Faloutsos</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Tensor decomposition for signal processing and machine learning</article-title>
          .
          <source>IEEE Transactions on Signal Processing</source>
          <volume>65</volume>
          (
          <issue>13</issue>
          ),
          <volume>3551</volume>
          {
          <fpage>3582</fpage>
          (
          <year>July 2017</year>
          ). https://doi.org/10.1109/TSP.
          <year>2017</year>
          .2690524
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pedersen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Duluth at semeval-2017 task 6: Language models in humor detection</article-title>
          .
          <source>In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017)</source>
          . pp.
          <volume>384</volume>
          {
          <fpage>388</fpage>
          . Association for Computational Linguistics, Vancouver, Canada (
          <year>August 2017</year>
          ), http: //www.aclweb.org/anthology/S17-2064
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lavie</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dyer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hovy</surname>
          </string-name>
          , E.:
          <article-title>Humor recognition and humor anchor extraction</article-title>
          .
          <source>In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>2367</volume>
          {
          <fpage>2376</fpage>
          . Association for Computational Linguistics, Lisbon, Portugal (Sep
          <year>2015</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D15</fpage>
          -1284, https://www.aclweb.org/ anthology/D15-1284
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bousquet</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lal</surname>
            ,
            <given-names>T.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Scholkopf, B.:
          <article-title>Learning with local and global consistency</article-title>
          .
          <source>In: Proceedings of the 16th International Conference on Neural Information Processing Systems</source>
          . pp.
          <volume>321</volume>
          {
          <fpage>328</fpage>
          . NIPS'03, MIT Press, Cambridge, MA, USA (
          <year>2003</year>
          ), http://dl.acm.org/citation.cfm? id=
          <volume>2981345</volume>
          .
          <fpage>2981386</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>