<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Forum for Information Retrieval Evaluation, December</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>YouTube Com ments and Posts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sanjeepan Sivapiran</string-name>
          <email>sanjeepan.18@cse.mrt.ac.lk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Charangan Vasantharajan</string-name>
          <email>charangan.18@cse.mrt.ac.lk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Uthayasanker Thayasivam</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Sentiment Analysis„ Code-Mixed, Transformers, Tamil, ULMFiT</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, University of Moratuwa</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>1</volume>
      <fpage>3</fpage>
      <lpage>17</lpage>
      <abstract>
        <p>This paper presents the methodologies implemented while doing Sentiment Analysis on Dravidian code-mixed comments and posts collected from social media. With a dataset of code-mixed Tamil, We experimented with transformer-based models such as multilingual BERT and DistilBERT and ULMFiT. This work submitted to the track 'Sentiment Analysis for Dravidian Languages in Code-Mixed Text' organized by the Forum for Information Retrieval Evaluation. Although it received the seventh rank for the Tamil task in the benchmark, This paper improves upon the results by a margin to attain the final weighted F1 score of 0.61 for the Tamil task.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the past few years, usage of social media platforms has drastically increased. With this trend,
cyberbullying and hate speech also increased and created a need to analyze comments/posts
on social media. Sentimental Analysis is a study that uses Natural Language Processing in
identifying subjective opinions or emotional responses about a given topic.[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] There are already
multiple steps taken to make use of sentimental Analysis in monolingual texts. But there has
been an indispensable demand for sentimental Analysis in code-mixed Dravidian languages
(Tamil, Malayalam, and Kannada) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Code-mixing is a prevalent phenomenon in a multilingual
community, and the code-mixed texts sometimes write in non-native scripts.[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] Systems trained
on monolingual data fail on code-mixed data due to the complexity of code-switching at diferent
linguistic levels in the text.
      </p>
      <p>
        The objective of our study is to classify YouTube comments into positive, negative, neutral,
mixed emotions or if the word is not in Tamil, which is in code-mixed form [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. For this task,
transformer architecture models Like multilingual BERT and DistilBERT yielded good results
since they optimized for low-resourced languages like Tamil. Yet ULMFiT made the best results
compared to transformer models. Since data was in code-mixed form, models had dificulty
https://rtuthaya.lk/ (U. Thayasivam)
https://www.linkedin.com/in/sanjeepan/ (S. Sivapiran); https://chaarangan.github.io/ (C. Vasantharajan);
© 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
understanding semantic relationships and their respective contexts. We used the translation and
transliteration technique to convey a word from one writing system to another while preserving
the context and semantics to overcome this issue.
      </p>
      <p>
        The rest of the sections in the paper are as follows. Section 2 reviews related experiment
works in Sentiment Analysis. Section 3 describes the given dataset in the Shared Task[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The
fourth section(4) presents the system description and conducted experiments using diferent
approaches and features as well as the results reaped from the experiments of our proposed
system. Benchmark results are discussed in section 4.5 and finally, the conclusion.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Cyberbullying and hateful speech are unpleasant parts of social media. To ensure the
wellbeing of the social media users from cyberbullying, social media companies always had to
invest/contribute in sentimental analysis research. Due to that, an adequate amount of studies
has been already done. Historically, there have been two approaches to solve sentimental
analysis problems lexicon-based and machine learning approaches [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Even though they produce
moderately quality results, they failed against human-generated data. Due to that, new deep
learning models such as Bidirectional Recurrent Neural Network(RNN)[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and Long Short-Term
Memory(LSTM) network [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] were introduced. On the other hand, [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] conducted experiments
in Kannada-English using the traditional learning approaches such as Logistic Regression(LR),
Support Vector Machine(SVM), Multinomial Naive Bayes, K-Nearest Neighbors(KNN), Decision
Trees(DT), and Random Forest (RF).
      </p>
      <p>
        To address the sentiment analysis problem using the above techniques, We need a corpus.
Since social-media comments/posts do not follow the strict grammar rules and also they are
always in non-native scripts as well as code-mixed [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] created a gold standard
TamilEnglish code-switched, sentiment-annotated corpus containing 15,744 comment posts from
YouTube to overcome the above situation. Moreover, Chakravarthi et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] created a standard
corpus for Malayalam-English to increase the sentiment analysis tasks in the code-mixed
contents.
      </p>
      <p>[13] explored in Tamil-English, Kannda-English, and [14]Malayalam-English by using the
transformer-based model mBERT. The model performed well but failed in some text where
code-mixed comes[15]. As an extension work of this research work, [16] conducted experiments
on diferent kinds of models such as Bidirectional LSTM, mBERT, DistilBERT, and ULMFiT [ 17]
to overcome this issue. Moreover, they developed a standard Translation and Transliteration
algorithm to convert the corpus into monolingual. From this approach, they could be able to
improve their system’s performance.</p>
      <p>Over the past decade, diferent kinds of models introduced, but contrasted to conventional
Recurrent Neural Network models (RNNs), the eficiency and performance of the transformer
models such as BERT[18], DistilBERT[19], mBERT[20] are remarkably distinguished. BERT
[21]) models designed to contextualize the text by jointly conditioning on both left and proper
contexts. Due to that, transformer models can be used to produce a state-of-the-art result by
just fine-tuning the output layer. After studying the above research studies, we decided to go
with transformer models and ULMFiT.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>The Tamil-English data set is provided by the Dravidian-CodeMix-FIRE 2021 organizing
committee, which extracted from Tamil YouTube comments/posts that contains three parts(Train,
Validation, Test). The training, validation and testing datasets have 35,656, 3962, and 4392
comments, respectively, with annotated labels. The dataset consists of texts in five diferent
classes as follows:</p>
      <sec id="sec-3-1">
        <title>Text</title>
        <p>Vijay Annaa Ur Maasssss Therrrrriii
நம்ப நேட நாசாமா தான் ேபாச்ேச
Thala’s hardwork + dedication in the movie next level #Thalaaaaaaaaaaa
மனிதனாய் வாழ்வதற்கு ேதைவ மனிதாபிமானம் மட்டுேம... ஜாதி இல்ைல...!
Subtitle me traller dekhne wale like karo</p>
      </sec>
      <sec id="sec-3-2">
        <title>Label</title>
        <p>Positive
Negative
Mixed Feelings
Unknown State
Not in Tamil</p>
        <p>The data set contains three code-mixed sentences: Inter-Sentential switch, Intra-Sentential
switch, and Tag switching. They wrote in either native Tamil script or English grammar with
Tamil. Some comments wrote in Tamil script with English words between them. Table 2
describes the dataset statistics and it is visualized in Figure 1. The following items show the
ifve diferent classes of comments with a definition:
• Positive: Comments which are not ofensive</p>
        <p>e.g: ennaya trailer Ku mudi Ellam nikkudhu... Vera level trailer..
• Negative: Comments which are ofensive
e.g: எந்ெதந்த youtube channel காரங்க எல்லாம் இைத ஜாதி ெவறி படம்குறாங்கேளாேளா
அவங்ெகல்லாம் அந்த ஜாதி என்றறிக
• Mixed Feelings: Comments which are both negative and positive
e.g:Kaagam karaindhu koodi unnum, Manidham ennum moodar koodam koodi serdhu
pagaimai kollum... Idil yaar uyarthinai yaar agrinai
• Unknown State: Comments which are not determined</p>
        <p>e.g:Vandha raja vah dhaan varuven Vera level str
• Not in Tamil: Comments which are not in native Tamil</p>
        <p>e.g:Subtitle me traller dekhne wale like karo</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. System Description and Result Analysis</title>
      <sec id="sec-4-1">
        <title>4.1. Preprocessing</title>
        <p>Since the dataset collected from YouTube does not follow any grammar rules and is in code-mixed
form. The dataset undergoes the Following steps to use the dataset eficiently.
• The first step is to stemming and lemmatization the words and lower casing the only
romanized words as there is no such thing in Tamil script.</p>
        <sec id="sec-4-1-1">
          <title>Label</title>
          <p>positive
negative
unknown_state
mixed-feelings
Not-Tamil
Total</p>
          <p>• The next step is to remove all emojis, special characters, numbers, and punctuations as
they do not carry any meaning to the sentence.
• Finally, we applied the algorithm introduced by [16] to do translation and transliteration
on the comments and posts to create a monolingual corpus.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Translation</title>
        <p>After loading the dataset, we used an extensive corpus of English words from NLTK-corpus 1 to
detect English words in a sentence; if the word is in the English dictionary, then we translated
the word into native Tamil script; otherwise, we ignored the word. For this purpose, We used
Google Translate API2.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Transliteration</title>
        <p>Most of the comments are in code mixed form. Comments should be in the native script to get
state-of-the-art results from transformers models. Transliteration is the process of transferring
a word from the alphabet of one language to another. All non-native Tamil words converted
1https://www.nltk.org/
2https://pypi.org/project/googletrans/
into the same meaning Tamil words using transliteration. To achieve this, we used AI4Bharat
Transliteration3.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Models</title>
        <p>Recently released transformer models such as BERT achieves a state of the art results in text
classification tasks. Considering the performance of transform models, we choose to start
with multilingual BERT and DistilBERT. All of our transformer-based models are culled from
HuggingFace4 transformers library and the models’ parameters are as stated in Table 3. Figure
2 depicts the architecture of our best-performed model(ULMFiT).</p>
        <sec id="sec-4-4-1">
          <title>Parameters</title>
          <p>LSTM Units
Dropout
Activation Function
Max Len
Learning Rate
Optimizer
Loss Function
Batch Size
Epochs</p>
          <p>DistilBERT model is a small, fast, and light transformer-based model trained on the Wikipedia
dataset. It has 40% fewer parameters than BERT, runs 60% faster while preserving over 95%
of BERT’s performances. Since our purpose is to train a model in Tamil(non-Latin script), we
selected the distilbert-base-multilingual-cased model, which has six layers, 768 dimensions,
12 heads, and tantalizing 134M parameters.</p>
          <p>We also experimented with bert-base-multilingual-cased as our pre-trained multilingual
model having approximately 110M parameters with 12-layers and 768 hidden states.</p>
        </sec>
      </sec>
      <sec id="sec-4-5">
        <title>4.5. Results and Analysis</title>
        <p>Teams were ranked by the weighted average F1 score of their model, and we received 7th
rank. Even though our model got above rank, the F1-score diference between the first team is
relatively low.</p>
        <p>In the beginning, we start with our BERT model and it doesn’t perform well. It may have
happened due to the lack of BERT multilingual based model training in the Tamil language. In
the next step, We approached the problem with the ULMFiT model, a transfer learning technique
[22]. ULMFiT’s model architecture is diferent from transformer models, and it is an efective
transfer learning method that can apply to any task in NLP. The table shows that ULMFiT
3https://pypi.org/project/ai4bharat-transliteration/
4https://github.com/huggingface/
yielded an F1-Score of 0.6101, and DistilBert, mBERT yielded 0.60104 and 0.5963, respectively.
Precision and recalls of the above models showen in Table 4.</p>
        <sec id="sec-4-5-1">
          <title>Models</title>
          <p>ULMFiT
DistilBert
mBERT</p>
        </sec>
        <sec id="sec-4-5-2">
          <title>Precision</title>
          <p>0.6075
0.5978
0.5782</p>
        </sec>
        <sec id="sec-4-5-3">
          <title>Recall</title>
          <p>0.6045
0.5984
0.5627</p>
        </sec>
        <sec id="sec-4-5-4">
          <title>F1-Score</title>
          <p>0.6101
0.6014
0.5963</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>
        In this research, we have analyzed diferent NLP techniques to classify ofensive language in
Tamil code-mixed YouTube comments[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. We used a novel technique, transliteration, which
leverages the accuracy across all three models.Also, We experimented with transformer models
and transfer learning technique(ULMFiT) models. Even though transformer models are more
advanced, To our task, ULMFiT yields the best results. Since Tamil is a low-resourced language
[23], our research also can be applied to other low-resourced languages without much dificulty.
dataset for code-mixed Malayalam-English, in: Proceedings of the 1st Joint Workshop on
Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration
and Computing for Under-Resourced Languages (CCURL), European Language Resources
association, Marseille, France, 2020, pp. 177–184. URL: https://aclanthology.org/2020.sltu-1.
25.
[13] C. Vasantharajan, U. Thayasivam, Hypers@DravidianLangTech-EACL2021: Ofensive
language identification in Dravidian code-mixed YouTube comments and posts, in:
Proceedings of the First Workshop on Speech and Language Technologies for Dravidian
Languages, Association for Computational Linguistics, Kyiv, 2021, pp. 195–202. URL:
https://www.aclweb.org/anthology/2021.dravidianlangtech-1.26.
[14] B. R. Chakravarthi, P. K. Kumaresan, R. Sakuntharaj, A. K. Madasamy, S. Thavareesan,
P. B, S. Chinnaudayar Navaneethakrishnan, J. P. McCrae, T. Mandl, Overview of the
HASOC-DravidianCodeMix Shared Task on Ofensive Language Detection in Tamil and
Malayalam, in: Working Notes of FIRE 2021 - Forum for Information Retrieval Evaluation,
CEUR, 2021.
[15] R. Priyadharshini, B. R. Chakravarthi, S. Thavareesan, D. Chinnappa, T. Durairaj, E. Sherly,
Overview of the dravidiancodemix 2021 shared task on sentiment detection in tamil,
malayalam, and kannada, in: Forum for Information Retrieval Evaluation, FIRE 2021,
Association for Computing Machinery, 2021.
[16] C. Vasantharajan, U. Thayasivam, Towards ofensive language identification for tamil
code-mixed youtube comments and posts, 2021. a r X i v : 2 1 0 8 . 1 0 9 3 9 .
[17] J. Howard, S. Ruder, Universal language model fine-tuning for text classification, 2018.
      </p>
      <p>a r X i v : 1 8 0 1 . 0 6 1 4 6 .
[18] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, 2019. a r X i v : 1 8 1 0 . 0 4 8 0 5 .
[19] V. Sanh, L. Debut, J. Chaumond, T. Wolf, Distilbert, a distilled version of bert: smaller,
faster, cheaper and lighter, 2020. a r X i v : 1 9 1 0 . 0 1 1 0 8 .
[20] T. Pires, E. Schlinger, D. Garrette, How multilingual is multilingual bert?, 2019.</p>
      <p>a r X i v : 1 9 0 6 . 0 1 5 0 2 .
[21] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I.
Polosukhin, Attention is all you need, 2017. a r X i v : 1 7 0 6 . 0 3 7 6 2 .
[22] F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, Q. He, A comprehensive survey
on transfer learning, 2020. a r X i v : 1 9 1 1 . 0 2 6 8 5 .
[23] B. R. Chakravarthi, R. Priyadharshini, V. Muralidaran, S. Suryawanshi, N. Jose, E. Sherly,
J. P. McCrae, Overview of the track on sentiment analysis for dravidian languages in
code-mixed text, in: Forum for Information Retrieval Evaluation, 2020, pp. 21–24.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Jose</surname>
          </string-name>
          , A.
          <string-name>
            <surname>Kumar</surname>
            <given-names>M</given-names>
          </string-name>
          , T. Mandl,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. R L</given-names>
            ,
            <surname>J. P. McCrae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherly</surname>
          </string-name>
          ,
          <article-title>Findings of the shared task on ofensive language identification in Tamil, Malayalam, and Kannada</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, Association for Computational Linguistics</source>
          , Kyiv,
          <year>2021</year>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>145</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          . dravidianlangtech-
          <volume>1</volume>
          .
          <fpage>17</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. Raja</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <article-title>Comparison of pretrained embeddings to identify hate speech in indian code-mixed text</article-title>
          ,
          <source>in: 2020 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>25</lpage>
          .
          <source>doi:1 0 . 1 1</source>
          <volume>0</volume>
          <fpage>9</fpage>
          <string-name>
            <surname>/ I C A C C C N</surname>
          </string-name>
          <article-title>5 1</article-title>
          <volume>0 5 2 . 2 0 2 0 . 9 3 6 2 7 3 1 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Kumar</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Overview of the hasoc track at ifre 2020: Hate speech and ofensive language identification in tamil, malayalam, hindi, english and german</article-title>
          ,
          <source>in: Forum for Information Retrieval Evaluation</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryawanshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Findings of the shared task on troll meme classification in Tamil</article-title>
          ,
          <source>in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, Association for Computational Linguistics</source>
          , Kyiv,
          <year>2021</year>
          , pp.
          <fpage>126</fpage>
          -
          <lpage>132</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .dravidianlangtech-
          <volume>1</volume>
          .
          <fpage>16</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Thavareesan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chinnappa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hande</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Vasantharajan</surname>
          </string-name>
          ,
          <source>Findings of the Sentiment Analysis of Dravidian Languages in Code-Mixed Text</source>
          <year>2021</year>
          , in: Working Notes of FIRE 2021 -
          <article-title>Forum for Information Retrieval Evaluation</article-title>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>O.</given-names>
            <surname>Habimana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. X.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Sentiment analysis using deep learning approaches: an overview</article-title>
          ,
          <source>Science China Information Sciences</source>
          <volume>63</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. Paliwal,</surname>
          </string-name>
          <article-title>Bidirectional recurrent neural networks</article-title>
          ,
          <source>IEEE Transactions on Signal Processing</source>
          <volume>45</volume>
          (
          <year>1997</year>
          )
          <fpage>2673</fpage>
          -
          <lpage>2681</lpage>
          .
          <source>doi:1 0 . 1 1 0 9 / 7 8 . 6 5</source>
          <volume>0 0 9 3 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sherstinsky</surname>
          </string-name>
          ,
          <article-title>Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network, Physica D: Nonlinear Phenomena 404 (</article-title>
          <year>2020</year>
          )
          <article-title>132306</article-title>
          . URL: http://dx.doi.org/10.1016/j.physd.
          <year>2019</year>
          .
          <volume>132306</volume>
          .
          <source>doi:1 0 . 1 0</source>
          <volume>1 6</volume>
          / j . p
          <source>h y s d . 2 0</source>
          <volume>1 9 . 1 3 2 3 0 6 .</volume>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hande</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>KanCMD: Kannada CodeMixed dataset for sentiment analysis and ofensive language detection</article-title>
          ,
          <source>in: Proceedings of the Third Workshop on Computational Modeling of People's Opinions</source>
          , Personality, and
          <article-title>Emotion's in Social Media, Association for Computational Linguistics</article-title>
          , Barcelona,
          <source>Spain (Online)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>54</fpage>
          -
          <lpage>63</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .peoples-
          <volume>1</volume>
          .6.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jayapal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Thavareesan</surname>
          </string-name>
          ,
          <article-title>Nuig-shubhanker@dravidian-codemix- fire2020: Sentiment analysis of code-mixed dravidian text using xlnet</article-title>
          ,
          <source>in: FIRE</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Muralidaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <article-title>Corpus creation for sentiment analysis in code-mixed Tamil-English text</article-title>
          ,
          <source>in: Proceedings of the 1st Joint Workshop on Spoken Language Technologies</source>
          for
          <article-title>Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), European Language Resources association</article-title>
          , Marseille, France,
          <year>2020</year>
          , pp.
          <fpage>202</fpage>
          -
          <lpage>210</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .sltu-
          <volume>1</volume>
          .
          <fpage>28</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          , N. Jose,
          <string-name>
            <given-names>S.</given-names>
            <surname>Suryawanshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sherly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>McCrae</surname>
          </string-name>
          ,
          <article-title>A sentiment analysis</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>