<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Multi-task Learning for Cross-Lingual Sentiment Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gaurish Thakkar[</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nives Mikelic Pr</string-name>
          <email>nmikelic@m</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>rko T</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Humanities and Social Sciences, University of Zagreb</institution>
          ,
          <addr-line>Zagreb 10000</addr-line>
          ,
          <country country="HR">Croatia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents a cross-lingual sentiment analysis of news articles using zero-shot and few-shot learning. The study aims to classify the Croatian news articles with the positive, negative, and neutral sentiment using the Slovene dataset. The system is based on a trilingual BERT-based model trained in three languages: English, Slovene, Croatian. The paper analyses di erent setups of using datasets in two languages and proposes a simple multi-task model to perform sentiment classi cation. The evaluation is performed using the few-shot and zeroshot scenarios in single-task and multi-task experiments for Croatian and Slovene.</p>
      </abstract>
      <kwd-group>
        <kwd>sentiment analysis</kwd>
        <kwd>cross-lingual</kwd>
        <kwd>transfer learning</kwd>
        <kwd>multitask learning</kwd>
        <kwd>news sentiment</kwd>
        <kwd>under-resourced languages</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Sentiment analysis is one of the most exciting applications of Natural Language
Processing. This eld encompasses text analysis ranging from the average
customer reviews of online products and movies to user-generated text from social
media platforms. Other applications include the analysis of nancial news [
        <xref ref-type="bibr" rid="ref2 ref7">2,7</xref>
        ]
for stock market movement. While the eld has a vast span across multiple
subareas, here we are interested in sentiment analysis of news articles. This paper
focuses on improving sentiment analysis of Croatian news articles tagged with
coarse-grained sentiment tags. Since Slovene and Croatian belong to the same
(sub)family of Slavic languages, Slovene was chosen as the hub language and in
the context of rather a small dataset available for Croatian.
      </p>
      <p>We seen the following items as the main contributions of our work:a) the
mutual dependence of the sentiment analysis task at three levels (i.e.,
documentlevel, paragraph-level, and sentence-level) is leveraged to improve the performance
of cross-lingual sentiment analysis; b) shared language encoder representations
across both languages are proposed and c) a language from the same language
family is used for cross-lingual sentiment transfer1.</p>
      <p>
        Copyright © 2021 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
1 Source code: https://github.com/cleopatra-itn/SentimentAnalyserSLHRNews
The previous state-of-the-art computational processes for sentiment analysis
relied on sentiment lexicons [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] as well as other classical methods like TF-IDF
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and dealing with various features along with SVM [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. As automatic feature
extraction became a common trend with the usage of deep learning techniques
[
        <xref ref-type="bibr" rid="ref15 ref18">15,18</xref>
        ], the Convolutional Neural Nets [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] took the lead, and then gradually have
been replaced by various Recurrent Neural Networks approaches [
        <xref ref-type="bibr" rid="ref12 ref16 ref23 ref9">9,16,23,12</xref>
        ].
      </p>
      <p>
        The machine-translation, as one of the well-studied cross-lingual techniques,
also aids to sentiment labelling. Its application ranges from lexicon translation
[
        <xref ref-type="bibr" rid="ref1 ref3">1,3</xref>
        ] up to instance translations [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Several recent works have explored the
use of transformers in multi-task learning setup for mono-lingual [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and
multilingual setups [
        <xref ref-type="bibr" rid="ref22 ref6">6,22</xref>
        ] for sentiment classi cation.
      </p>
      <p>
        The closest work upon which our research is built [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] performs zero-shot
sentiment learning on Croatian news articles by enriching a masked language
model (mBERT) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] using sentiment tags from the SentiNews and Croatian
news sentiment dataset. Our work is novel because of the way we use the dataset.
Previous work does not utilise paragraph-level and sentence-level annotations for
document-level classi cation but uses them for pre-training a masked-language
model. We utilise these annotations in a multi-task setup for aiding the sentiment
classi cation task for Croatian.
3
      </p>
    </sec>
    <sec id="sec-2">
      <title>Datasets</title>
      <p>We use two datasets in our experiments.</p>
      <p>
        SentiNews Dataset in Slovene This is a manually annotated dataset
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]2 in the domain of news. It contains 10,427 documents. All annotations have
three levels of granularity, i.e., document, paragraph, and sentence-level. The
dataset covers news from economics, nance, and politics published between 1
September 2007 and 31 December 2013. It contains all instance annotations of
each annotator, along with the news content and the nal sentiment label. The
dataset has been annotated at a ve-point Likert scale and has mapping onto
three labels (Positive, Negative, and Neutral) using the scale's average score.
The overall distribution of this dataset is given in Table 1.
      </p>
      <sec id="sec-2-1">
        <title>2 https://www.clarin.si/repository/xmlui/handle/11356/1110</title>
        <p>Sentiment Dataset in Croatian The Croatian dataset 3 was created
using guidelines similar to those of the SentiNews dataset. The text content comes
from the Croatian 24sata daily news portal. It covers topics such as health,
lifestyle, and automotive news. Table 1 shows the statistics for this dataset. Like
the Slovenian corpus, this dataset is also annotated with 3-class sentiment
labels and covers the same domain. However, it does not contain paragraph and
sentence-level annotations. Both datasets present an imbalanced-class
distribution phenomenon.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>
        We strive to leverage contextual information from the Slovene dataset accessible
at three distinct levels and dataset from Croatian in our proposed system to
promote knowledge transfer between two languages. As Croatian and Slovene
proved to have the highest level of mutual intelligibility among three South
Slavic languages (Croatian, Slovene and Bulgarian) represented in the mutual
intelligibility study [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], we hypothesise Slovene could be the right candidate
as a hub language for cross-lingual knowledge transfer to Croatian. Also, both
datasets belong to the same text type { news. A simple multi-task learning setup
with three di erent task heads is employed. Here, each task head represents the
classi cation layer responsible for classifying the given instance of a particular
type. The instances are of three types, namely document-level, paragraph-level,
and sentence-level. The document-classi cation layer is trained by concatenating
the Croatian and Slovene dataset. The paragraph-classi cation and the
sentenceclassi cation layer are only fed with Slovene instances since the Croatian dataset
does not provide this information.
      </p>
      <p>A shared encoder is used for feature extraction, enabling feature sharing
across all three tasks and both languages. Our model is trained in ve di erent
scenarios in total, while the overall architecture is presented in Fig. 1. In the
zeroshot learning setting, we do not use the Croatian data in training. However, use
the Croatian test for reporting the performance.
1. Single-task-SL-Zero-shot-HR - Train only Slovene document-level data.</p>
      <p>In this setting, the Slovene data is used in the zero-shot setting for the
Croatian sentiment analysis.
2. Multi-task-SL-Zero-shot-HR - Train using all three levels on Slovene
data. This setting is the same as the previous one, except that all three
classi cation heads were learning respective tasks using Slovene data. We
did not use any Croatian data in this or previous setting.
3. Single-task-HR - Train only using Croatian document-level data. This
setting is a classic ne-tuning setup involving a single head learning classi
cation on a single dataset.</p>
      <sec id="sec-3-1">
        <title>3 https://www.clarin.si/repository/xmlui/handle/11356/1342</title>
        <p>
          4. Multi-task-HR+SL - Train with Croatian (document-level) and Slovene
data (all three levels). We concatenated the respective document-level
instances from respective languages at the same time. The other heads were
trained using compatible instances from Slovene datasets.
5. Single-task-HR+SL - Train with Slovene and Croatian data
(documentlevel). This approach is similar to the previous one, but the single
documentlevel classi cation head is trained.
This section presents a brief description of the pre-processing steps, followed by
the details of the experiments.
1. The empty string values, which have neutral tags, were dropped. As null
strings have no content, we decided to drop these instances.
2. The strings based on content were de-duplicated. Many strings in the dataset
were duplicates. We performed this step in order to prevent leakage of the
instances into validation or test set split.
Our pipeline used a shared encoder based on a BERT-based masked language
model. The tri-lingual model named CroSloEngual [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] trained with three
languages (Slovene, Croatian, and English) on 5.9 billion tokens altogether. This
model was chosen as it considers two languages that we are interested in
transferring knowledge. The model outperforms the mBERT [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] for the task of Part
of Speech (POS) tagging, Named Entity Recognition (NER) and dependency
parsing for these three languages.
5.3
        </p>
        <p>Experiments
In the datasets de ned in Table 2, the data is split into 80:20 train-test split
in a strati ed fashion. The 10% or a proportionate train set is kept aside as a
development set, since all the datasets di er in size. All datasets are combined in
a single collection. The combined datasets behave as our size-proportional
population. Our model is trained sequentially in a single-task setting by randomly
sampling tasks from this collection. Using the test set from Croatian and Slovene,
we evaluated our proposed approach. Table 3 shows the model con guration. All
our models were trained using Nvidia RTX 3090 (24GB) with a batch size of 32.
All hyperparameters were constant during the whole experiments except epoch
which varied from 5 for a single task to 3 for a multi-task setup. These values
were chosen to prevent over tting on the train set. The overall training time for
the MTL setup was 3 hours. We evaluated the performance on the development
set and chose the best model for reporting test performance.</p>
        <p>
          Table 3 reports all the results of our experiment.
The single-task (STL) results and the multi-task (MTL) setup results are
reported in Table 4. Precision, recall, and F1 are macro averaged for all the
experiments. We used a simple majority-class classi er as our baseline. Our MTL
setup with Croatian and Slovene dataset outperforms the other settings for
Croatian sentiment classi cation. However, it does not perform best when tested on
document-level classi cation for Slovene. There is a small drop in the
performance similar to what was previously reported by [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. The second-best
performing model for Croatian is the single-task variant with Slovene and Croatian
data. The worst performing model on Croatian was the MTL with only Slovene
data and the STL model, which was performing zero-shot learning. Nevertheless,
the SL MTL performs better on Slovene test-sets. The HR STL, which used the
least amount of data compared to other settings, seemed to perform at par with
SL STL in F1(55.61 vs 56.95), but both had contrasting precision and recall.
        </p>
        <p>For paragraph-level, SL MTL has similar performance to SL+HR MTL, but
we see a slight drop in performance in the latter case. A similar observation can
be made for sentence-level classi cation.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>
        The research presented in this paper has revealed that having a small amount of
target language (Croatian) data helps in overall cross-lingual transfer learning.
Adding data from another language hinders the source language task's
performance but improves task performance in the target language. Comparing our
work with [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], which proposes joint optimisation of language modelling task
and paragraph-level sentiment analysis for the document-level sentiment
classication, we did not perform any intermediate training but utilised the available
annotation directly at the ne-tuning stage. Another signi cant di erence is the
use of a trilingual shared encoder which performs better than mBERT. We have
qualitatively analysed the MTL model's errors on the test set and discovered
that the model does pick up cues from the sentiment bearing words from the
input. However, the analysis of these errors would be beyond the scope of this
paper. The current model tags the news document as positive or negative when
it nds positive or negative words in neutral news. The articles, which are
advertisements or recipes, contain words with positive sentiment. Most of the errors
belong to this class. Since our encoder has xed-length input, the truncated text
prevents correct classi cation. We can solve this problem by performing a
sliding window sampling over the text or using the beginning and the end of each
article.
7
      </p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>We presented an overall setup for cross-lingual sentiment analysis (SA) of
Croatian news documents using a Multi-task learning approach. The goal was to
perform knowledge transfer using existing datasets and models to aid SA in
Croatian. For this purpose, we used a large Slovene SentiNews corpus created
similarly to the Croatian corpus.</p>
      <p>The publicly available trained trilingual BERT-based language model for
feature representation was utilised and used as a shared encoder for the downstream
tasks. We combined the Croatian and Slovene datasets for document-level
classi cation and trained three di erent classi cation heads. The results show that
the MTL setup outperformed the STL setup for Croatian.</p>
      <p>In the future, we would like to experiment more in an under-resourced setting
for Slovene and balance the dataset among distinct levels. Another interesting
approach would be to use each level of the datasets individually in a
hierarchical fashion to help with the next level's classi cation task. For example, using
sentence-level features to aid the paragraph classi cation could be fused into the
document-level prediction process. We believe that processing documents from
di erent text types and topics separately would be an easier task. Therefore,
in our future work, we plan to cluster documents into text-types and topics
like recipes, advertisements or obituaries and process them separately. Also, our
experiment setting could be further checked by running it for other language
pairs supported by this CroSloEngual BERT-based model: English-Slovene and</p>
      <p>English-Croatian. In this way, we could verify whether the genetically and
typologically distant language, as English here is, would contribute to the
performance.
8</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>The work presented in this paper has received funding from the European
Union's Horizon 2020 research and innovation program under the Marie
SklodowskaCurie grant agreement no. 812997 and under the name CLEOPATRA
(Crosslingual Event-centric Open Analytics Research Academy).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Abdalla</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hirst</surname>
          </string-name>
          , G.:
          <article-title>Cross-lingual sentiment analysis without (good) translation</article-title>
          .
          <source>In: Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers)</source>
          . pp.
          <volume>506</volume>
          {
          <fpage>515</fpage>
          .
          <article-title>Asian Federation of Natural Language Processing</article-title>
          , Taipei,
          <source>Taiwan (Nov</source>
          <year>2017</year>
          ), https://www.aclweb. org/anthology/I17-1051
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Agic</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ljubesic</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tadic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Towards sentiment analysis of nancial texts in croatian</article-title>
          .
          <source>In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Balahur</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steinberger</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kabadjov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zavarella</surname>
          </string-name>
          , V.,
          <string-name>
            <surname>van der Goot</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>Halkia</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pouliquen</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Belyaeva</surname>
          </string-name>
          , J.:
          <article-title>Sentiment analysis in the news</article-title>
          .
          <source>In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bucar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Znidarsic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Povh</surname>
          </string-name>
          , J.:
          <article-title>Annotated news corpora and a lexicon for sentiment analysis in slovene</article-title>
          .
          <source>Language Resources and Evaluation</source>
          <volume>52</volume>
          (
          <issue>3</issue>
          ),
          <volume>895</volume>
          {
          <fpage>919</fpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Cer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kong</surname>
          </string-name>
          , S.y.,
          <string-name>
            <surname>Hua</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Limtiaco</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>St</surname>
          </string-name>
          . John, R.,
          <string-name>
            <surname>Constant</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guajardo-Cespedes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tar</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strope</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurzweil</surname>
          </string-name>
          , R.:
          <article-title>Universal sentence encoder for English</article-title>
          .
          <source>In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          . pp.
          <volume>169</volume>
          {
          <fpage>174</fpage>
          . Association for Computational Linguistics, Brussels, Belgium (Nov
          <year>2018</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D18</fpage>
          -2029, https://www.aclweb. org/anthology/D18-2029
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Chidambaram</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sung</surname>
            ,
            <given-names>Y.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strope</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurzweil</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Learning cross-lingual sentence representations via a multi-task dual-encoder model</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>12836</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Day</surname>
            ,
            <given-names>M.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>C.C.</given-names>
          </string-name>
          :
          <article-title>Deep learning for nancial sentiment analysis on nance news providers</article-title>
          .
          <source>In: 2016 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM)</source>
          . pp.
          <volume>1127</volume>
          {
          <fpage>1134</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Dong</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Adaptive recursive neural network for target-dependent twitter sentiment classi cation</article-title>
          . In:
          <article-title>Proceedings of the 52nd annual meeting of the association for computational linguistics (volume 2: Short papers)</article-title>
          . pp.
          <volume>49</volume>
          {
          <issue>54</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Golubovic</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gooskens</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Mutual intelligibility between west and south slavic languages</article-title>
          .
          <source>Russian linguistics 39(3)</source>
          ,
          <volume>351</volume>
          {
          <fpage>373</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>K.H.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>H.H.</given-names>
          </string-name>
          :
          <article-title>Emotion classi cation of online news articles from the reader's perspective</article-title>
          .
          <source>In: 2008 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology</source>
          . vol.
          <volume>1</volume>
          , pp.
          <volume>220</volume>
          {
          <fpage>226</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Majumder</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poria</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hazarika</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cambria</surname>
          </string-name>
          , E.:
          <article-title>Dialoguernn: An attentive rnn for emotion detection in conversations</article-title>
          .
          <source>In: Proceedings of the AAAI Conference on Arti cial Intelligence</source>
          . vol.
          <volume>33</volume>
          , pp.
          <volume>6818</volume>
          {
          <issue>6825</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Pang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vaithyanathan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Thumbs up? sentiment classi cation using machine learning techniques</article-title>
          .
          <source>In: Proceedings of the 2002 Conference on Empirical Methods in Natural Language Processing (EMNLP</source>
          <year>2002</year>
          ). pp.
          <volume>79</volume>
          {
          <fpage>86</fpage>
          . Association for Computational Linguistics (
          <year>Jul 2002</year>
          ). https://doi.org/10.3115/1118693.1118704, https://www.aclweb.org/anthology/ W02-1011
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Pelicon</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pranjic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miljkovic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skrlj</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pollak</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Zero-shot learning for cross-lingual news sentiment classi cation</article-title>
          .
          <source>Applied Sciences</source>
          <volume>10</volume>
          (
          <issue>17</issue>
          ),
          <volume>5993</volume>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perelygin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chuang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potts</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Recursive deep models for semantic compositionality over a sentiment treebank</article-title>
          .
          <source>In: Proceedings of the 2013 conference on empirical methods in natural language processing</source>
          . pp.
          <volume>1631</volume>
          {
          <issue>1642</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vinyals</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.V.</given-names>
          </string-name>
          :
          <article-title>Sequence to sequence learning with neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1409.3215</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Taboada</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brooke</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , To loski, M.,
          <string-name>
            <surname>Voll</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stede</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Lexicon-based methods for sentiment analysis</article-title>
          .
          <source>Computational linguistics 37(2)</source>
          ,
          <volume>267</volume>
          {
          <fpage>307</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Tai</surname>
            ,
            <given-names>K.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C.D.:
          <article-title>Improved semantic representations from treestructured long short-term memory networks</article-title>
          .
          <source>In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)</source>
          . pp.
          <volume>1556</volume>
          {
          <issue>1566</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Ulcar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robnik-Sikonja</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Finest bert and crosloengual bert</article-title>
          . In: Sojka,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Kopecek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Pala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Horak</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . (eds.) Text, Speech, and Dialogue. pp.
          <volume>104</volume>
          {
          <fpage>111</fpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Wan</surname>
            ,
            <given-names>X.:</given-names>
          </string-name>
          <article-title>Co-training for cross-lingual sentiment classi cation</article-title>
          .
          <source>In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP</source>
          . pp.
          <volume>235</volume>
          {
          <issue>243</issue>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Zhang, Y.:
          <article-title>Personalized microblog sentiment classi cation via adversarial cross-lingual multi-task learning</article-title>
          .
          <source>In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>338</volume>
          {
          <fpage>348</fpage>
          . Association for Computational Linguistics, Brussels, Belgium (Oct-Nov
          <year>2018</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D18</fpage>
          -1031, https://www. aclweb.org/anthology/D18-1031
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ahmad</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Law</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Constant</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abrego</surname>
            ,
            <given-names>G.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tar</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sung</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strope</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurzweil</surname>
          </string-name>
          , R.:
          <article-title>Multilingual universal sentence encoder for semantic retrieval</article-title>
          . CoRR abs/
          <year>1907</year>
          .04307 (
          <year>2019</year>
          ), http://arxiv. org/abs/
          <year>1907</year>
          .04307
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Zadeh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>P.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poria</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vij</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cambria</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morency</surname>
            ,
            <given-names>L.P.:</given-names>
          </string-name>
          <article-title>Multiattention recurrent network for human communication comprehension</article-title>
          .
          <source>In: Proceedings of the AAAI Conference on Arti cial Intelligence</source>
          . vol.
          <volume>32</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>