CoSaD- Code-Mixed Sentiments Analysis for Dravidian Languages Fazlourrahman Balouchzahi1 , Hosahalli Lakshmaiah Shashirekha2 and Grigori Sidorov1 1 Instituto Politécnico Nacional (IPN), Centro de Investigación en Computación (CIC), Mexico City, Mexico 2 Department of Computer Science, Mangalore University, Mangalore, India Abstract Analyzing sentiments or opinions in code-mixed languages is gaining importance due to increase in the use of social media and online platforms especially during the Covid-19 pandemic. In a multilingual society like India, code-mixing and script mixing is quite common as people especially the younger gen- eration are quite familiar in using more than one language. In view of this, the current paper describes the models submitted by our team MUCIC for the shared task in ’Sentiments Analysis (SA) for Dravid- ian Languages in Code-Mixed Text’. The objective of this shared task is to develop and evaluate models for code-mixed datasets in three Dravidian languages, namely: Kannada, Malayalam, and Tamil mixed with English language resulting in Kannada-English (Ka-En), Malayalam-English (Ma-En), and Tamil- English (Ta-En) language pairs. N-grams of char, char sequences, and syllables features are transformed into feature vectors and are used to train three Machine Learning (ML) classifiers with majority voting. The predictions on the Test set obtained average weighted F1-scores of 0.628, 0.726, and 0.619 securing 2nd , 4th , and 5th ranks for Ka-En, Ma-En, and Ta-En language pairs respectively. Keywords Code-Mixing, Sentiments Analysis, Dravidian Languages, n-grams, Machine Learning 1. Introduction The task of analyzing the opinions, feelings, and reviews posted on social media or online markets to identify the sentiments of users about a given topic, movie, song, product, etc. is called as Sentiments Analysis (SA). For example, a video on Instagram or a product in e-markets can be viral and popular based on its reviews and sentiments posted by the customers/users [1, 2]. Lately, the demand for SA of social media data has increased both in academia and industry, especially for the code-mixed data [3]. Code-mixed data are common in multilingual communities such as India where people use more than one languages’ words, grammar, and phrases in their communication/ posts/ comments in social media or reviews in online shopping websites [4]. Code-mixed content in Dravidian languages is usually a combination of a native language FIRE 2021, Forum for Information Retrieval Evaluation, December 13-17, 2021, India " frs_b@yahoo.com (F. Balouchzahi); hlsrekha@gmail.com (H. L. Shashirekha); sidorov@cic.ipn.mx (G. Sidorov) ~ https://mangaloreuniversity.ac.in/dr-h-l-shashirekha (H. L. Shashirekha); http://www.cic.ipn.mx/~sidorov/ (G. Sidorov)  0000-0003-1937-3475 (F. Balouchzahi); 0000-0003-3901-3522 (G. Sidorov) © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) Table 1 Samples of code-mixed text in Dravidian languages such as Kannada, Tamil or Malayalam and English language at different linguistic units such as sentence, phrase, word, morpheme and sub-word. The code-mixed text will either be in a single script which is usually a Roman script or in multi-script i.e., a combination of Roman and native script may be with few words of the native language in Roman script. Table 1 presents some examples of single and multi-scripts code-mixed contents in Ta-En, Ma-En, and Ka-En language pairs from the datasets used in the shared task. Dravidian languages in general are under-resourced languages and code-mixing adds a further dimension mainly due to the problems with collecting and annotating code-mixed data for various applications. ’Sentiment Analysis for Dravidian Languages in Code-Mixed Text’ is a shared task in Dravidian-CodeMix-FIRE20211 with the aim of promoting SA of code-mixed texts in Ka-En, Ma-En, and Ta-En language pairs [5, 6]. This shared task is an extension of previous shared task of SA in Ta-En and Ma-En in FIRE 2020 [3] with the addition of Ka-En language pair [6]. The objective of the shared task is to identify the opinion/sentiment of the comments posted by the users on a given topic and classifying them further into one of the following categories: • Positive: comments contain positive contents or justify that speaker is in a positive state • Negative: comments contain negative contents or justify that speaker is in a negative state • Mixed_Feelings: comments contain positive as well as negative contents and hence cannot be explicitly categorized into one of the two classes mentioned earlier • Unknown_state: emotional state of a speaker is not clear or comments does not contain positive or negative contents explicitly • Not in indented language: comments are not written in the intended language In the earlier works, i) Balouchzahi et al. [1] experimented various features such as Skipgram word embedding, BPEmb2 sub-word embedding, and a combination of word and char n-grams to train ML classifiers for SA, and ii) Balouchzahi et al. [2] also explored and compared different learning approaches such as ML, Deep Learning (DL), and Transfer Learning (TL) for SA. In continuation of these works in SA in Dravidian languages, this paper describes the models 1 https://dravidian-codemix.github.io/2021/index.html 2 https://nlp.h-its.org/bpemb/ Table 2 Statistics of the datasets used in Dravidian-CodeMix-FIRE2020 Mixed_ Unknown_ Other Class Positive Negative Total Feelings state languages Ta-En 10,559 2,037 1,801 850 497 15,744 Dataset Ma-En 2,811 738 403 1,903 884 6,739 submitted by our team MUCIC to the Dravidian-CodeMix-FIRE2021 shared task. Three different feature sets, namely: char, char sequences, and syllables are explored to check the effectiveness of char level (characters) and sub-word level (char sequences and syllables) n-grams for code- mixed SA task. Each feature set is individually used to train three ML classifiers, namely: Linear Support Vector Machine (LSVM), Logistic Regression (LR) and Multi-Layer Perceptron (MLP) and the majority voting of the predictions of all the classifiers is used to classify the given sentiment. The code of the proposed methodology is available in our GitHub link3 . The rest of paper is organized as follows: Section 2 gives a summary of the best models submitted to the Sentiment Analysis for Dravidian Languages in Code-Mixed Text in Dravidian- CodeMix-FIRE20204 shared task and the Methodology is described in Section 3. Section 4 describes the results obtained and the paper concludes in Section 5. 2. Related Work Researches had submitted several models to ’Sentiment Analysis for Dravidian Languages in Code-Mixed Text’ shared task in Dravidian-CodeMix-FIRE2020 organized by Chakravarthi et al. [3, 7]. The shared task consists of similar sentiments categories (as mentioned in Section 1) in two language pairs, namely: Ta-En and Ma-En. Authors collected the Youtube comments to develop datasets consisting of 15,744 and 6,739 comments in Ta-En and Ma-En language pairs respectively, and provided the same to the participants of the shared task as Train, Dev and Test set. The label distribution of the comments in the dataset shown in Table 2 (borrowed from [2]) illustrates that the dataset is imbalanced for both the language pairs. Participants were supposed to train and evaluate their models locally on Train and Dev set respectively and then predict the class label of the Test set. These predictions were submitted to the shared task organizers for final evaluation and ranking which is based on average weighted scores. The brief descriptions of the models which exhibited good performance in this shared task are given below: Most of successful teams have utilized Multilingual BERT (mBERT5 ) [8] and XLM-Roberta [9] - the multilingual transformer based models for SA similar to that of code-mixed Offensive Language Identification (OLI) in Dravidian languages [10]. With the objective of using Masked Language Modeling (MLM), mBERT was trained on the top 104 languages that have largest Wikipedia including Kannada, Malayalam, and Tamil. Pires et al. [11] describe that mBERT can be employed for cross-lingual generalization. Moreover, based on the authors’ experiments, 3 https://github.com/fazlfrs/CoSaD 4 https://dravidian-codemix.github.io/2020/index.html 5 https://github.com/google-research/bert/blob/master/multilingual.md despite the high lexical overlap among different languages, mBERT is capable of transfering between languages with different scripts by capturing multilingual representations. XLM- Roberta also relay on MLM objective and cross-lingual transfer. By using the large-scale multilingual pre-training model trained on 2.5 TB of clean CommonCrawl data in 100 languages [12], XLM-Roberta has overcome the limitation of XLM [13] and mBERT in learning useful representations for under-resourced languages. Sun et al. [12] proposed a XLM-Roberta based model by extracting the abundant semantic information from the hidden layer state of XLM-Roberta, which is then fed as input into convolution and max pooling. Further, they concatenated the top hidden states and pooler to improve performances and reported that the proposed model without any pre-processing obtained better results. The proposed model outperformed all other models submitted to the shared task by securing 1st ranks (for both the language pairs) with average weighted F1-scores of 0.74 and 0.65 for Ma-En and Ta-En language pairs respectively. Ou et al. [14] developed a XLM-Roberta based model similar to the work of Sun et al. [12]. Here, the authors obtained the pooler output and the sequence of hidden states of the last layer of XLM-Roberta and concatenated the pooler output with the average-pooling and max-pooling of hidden-states of XLM-Roberta into a classifier. They merged and shuffled the Train and Dev sets and used k-fold cross validation to enhance the performances of the system. They obtained 1st rank for Ma-En language pair and average weighted F1-scores of 0.74 and 0.63 for Ma-En and Ta-En language pairs respectively. In a simple way, Sun et al. [15] proved the efficiency of multilingual transformers by fine-tuning the pre-trained multilingual BERT adopted from multi_cased_L12_H-768_A-126 . They secured 2nd and 4th ranks with average weighted F1-scores of 0.73 and 0.62 for Ma-En and Ta-En language pairs respectively. Huang et al. [16] proposed a multi-step integration of fine-tuned XLM-Roberta and mBERT transformers for the shared task and obtained average weighted F1-scores of 0.73 and 0.63 for Ma-En and Ta-En language pairs respectively. They used mBERT as binary classifier and XLM-Roberta as quaternary classifier and intertwined both model’s predictions for final decision. Zhu et al. [17] experimented an mBERT-based model along with BiLSTM by feeding the hidden state of the last layer of mBERT model to BiLSTM. Further, they set weights for each hidden state layer in BiLSTM and the weighted sum of hidden states is concatenated with the original output of mBERT. The results reported in leaderboard shows 2nd rank for both language pairs with average weighted F1-scores of 0.73 and 0.64 for Ma-En and Ta-En language pairs respectively. In addition to transformers, several models based on ML classifiers have also obtained promis- ing results in the shared task. Kanwar et al. [18] adopted under-sampling technique from TOMEK [19] to train several ML classifiers with various syntax based n-grams features. The best performance obtained was using LR classifier with word and char n-grams features for Ma-En and Ta-En language pairs with 0.71 and 0.62 average weighted F1-scores respectively. Balouchzahi et al. [2] submitted a majority voting of ML classifiers (Multinomial Naïve Bayes trained on Skipgram word embedding and Multi-Layer Perceptron (MLP) trained on the combi- nation of word and char n-grams) and BiLSTM model (with training a sub-word embedding using BPEmb library that is used as weight later in BiLSTM) for the shared task. The proposed model obtained 0.68 and 0.62 average weighted F1-scores for Ma-En and Ta-En language pairs 6 https://github.com/google-research/bert respectively. Researchers have explored several models based on ML and DL approaches with a combination of different embeddings and feature sets. Balouchzahi et al. [1] explored ML, DL and TL approaches by proposing (i) a ML-based voting classifier trained on a feature set of char sequences along with BPEmb sub-words ngrams and syntactic ngrams [20, 21] with three estimators, namely: LR, MLP, and eXtreme Gradient Boosting (XGB); (ii) A Keras sequential classifier trained on earlier feature set; and (iii) A Universal Language Model Fine-Tuning (ULMFiT) for SA. They used Dakshina7 dataset as raw text to train a tokenizer, universal Language Model (LM) for fine tuning and fast.ai8 library for training LM and SA classification model. Using ML-based voting classifier they obtained 0.72 and 0.62 average weighted F1-scores for Ma-En and Ta-En language pairs respectively. 3. Methodology The proposed methodology contains: • pre-processing texts • extracting char, char sequences and syllables as features from the texts • obtaining the corresponding n-grams from the n-gram generator • vectorizing the n-grams using TfidfVectorizer • training the ML classifiers • predict the labels of the Test set The pre-processing module adopted from Balouchzahi et al. [2] includes converting Emojis to text, removing punctuation, numbers, unnecessary characters and words of length less than 2 and lower casing the words written in Roman script. Words are split by a simple strategy of using attributes of string data type in a ’for’ loop to obtain char features. Char sequences are extracted as sub-word level features using everygrams9 library from NLTK. Syllable which comprises of vowels and consonants [22] is a smallest unit used to organize sequences of sounds and are considered as the building blocks in Text To Speech (TTS) tasks. Sidorov [23] proposed using syllables as features for Text Classification (TC) tasks. Syllable features are extracted using the Syllablizer10 library. Though the library works better for native scripts, results for code-mixed texts are also encouraging. The n-gram generator accepts a list of chars/char sequences/syllables of a word as input and will generate the corresponding n-grams which are vectorized using TfidfVectorizer11 to train the ML classifiers. The overview of feature engineering which includes the procedures to pre-process, extract features, generate n-grams, and obtaining TFIDF vectors is shown in Figure 1 and the range of n-grams for each feature type is given in Table 3. 7 https://github.com/google-research-datasets/dakshina 8 https://nlp.fast.ai 9 https://tedboy.github.io/nlps/generated/generated/nltk.everygrams.html 10 https://github.com/libindic/syllabalizer.git 11 https://scikit-learn.org/stable/modules/generated/sklearn.feature_extraction.text.TfidfVectorizer.html Figure 1: Feature Engineering Table 3 n-grams range for feature sets Feature Char Char sequences Syllables n-grams range (1, 5) (1, 6) (1, 8) Figure 2: Training the classifiers The parameters of LSVM and LR classifiers are set to default and that of MLP classifier are set as: hidden_layer_sizes = (150, 100, 50), max_iter = 300, activation = ’relu’, solver = ’adam’, random_state = 1. Each classifier is trained separately with the three feature sets mentioned earlier. The best performing feature and classifier pairs are selected manually based on their performances on Dev set and majority voting of the predictions on the Test set were submitted for final evaluation to the shared task organizers. Figure 2 presents the steps for training the individual classifier for each feature set. 4. Experiments and Results A post/comment in each language pair should be classified into one of the five categories as described in Section 1. The dataset provided by the shared task organizers [5, 6] includes a collection of code-mixed text from social media in three language pairs, namely: Ma-En, Ta-En, and Ka-En. These datasets were split into Train, Dev and Test sets and provided to the participants of the shared task to train and evaluate the models. The statistics of the datasets are given in Table 4. Similar to Dravidian-CodeMix-FIRE2020 shared task, the label distribution over the datasets illustrate that the datasets are highly imbalanced. The observation of the datasets in Table 4 illustrate that, for each class, Ta-En language pair consists of more samples and Ka-En language pair consists of less samples and this could affect the performance of the Table 4 Statistics of the Datasets Datasets Class Ta-En Ma-En Ka-En train dev test train dev test train dev test Positive 20,070 2,257 2,546 6,421 706 780 2,823 321 374 Negative 4,271 480 477 2,105 237 258 1,188 139 157 Mixed_Feelings 4,020 438 470 926 102 134 574 52 65 Unknown_state 5,628 611 665 5,279 580 643 711 69 62 Other languages 1,667 176 244 1,157 141 147 916 110 110 Total 35,657 3,962 4,402 15,888 1,766 1,962 6,212 691 768 Table 5 Results for the Development set Feature set Language pair Classifiers F1-score LR 0.59 Ta-En LSVM 0.60* MLP 0.58 LR 0.75* Char n-grams Ma-En LSVM 0.74 MLP 0.71 LR 0.65 Ka-En LSVM 0.66* MLP 0.63 Ta-En LR 0.59 LR 0.73 Char sequences Ma-En LSVM 0.74 MLP 0.69 n-grams LR 0.63 Ka-En LSVM 0.64 MLP 0.60 Ta-En LR 0.60* LR 0.74 Ma-En LSVM 0.73 Syllable n-grams MLP 0.70 LR 0.66* Ka-En LSVM 0.64 MLP 0.62 classifiers for Ka-En language pair. The predictions on the Test set submitted by the participants were evaluated based on the average weighted F1-scores. Organizers had encouraged the teams to evaluate the models locally on Dev set and then to submit the predictions on the Test set. Table 5 gives the performances of proposed methodology on the Dev set for all the three feature sets using the three classifiers for Table 6 Results for the Test set Language Feature set and Classifier Precision Recall F1-score Rank Char + LSVM 0.620 0.655 0.616 - Char seq. + LR 0.598 0.609 0.603 - Ta-En Syllable + LR 0.602 0.622 0.609 - Majority Voting 0.612 0.644 0.619 5 Char + LR 0.723 0.728 0.721 - Char seq. + LSVM 0.719 0.725 0.720 - Ma-En Syllable + LR 0.715 0.720 0.712 - Majority Voting 0.726 0.733 0.726 4 Char + LSVM 0.622 0.650 0.622 - Char seq. + LSVM 0.614 0.652 0.624 - Ka-En Syllable + LR 0.615 0.634 0.618 - Majority Voting 0.622 0.655 0.628 2 all the three language pairs. Observation of the results on the Dev set shows that LR and LSVM outperform each other for various feature sets and language pairs, while MLP always obtained the lowest results for all feature sets and all language pairs. The highlighted content in Table 5 correspond to the best performing classifier and the starred (*) score indicates the best feature set and classifier pair for the language pair. It can be seen that most of high performances are obtained with char n-grams followed by syllable n-grams. However, results using char sequences are interesting as well. The good performance of syllable n-grams reveals that they can be effectively used as features in TC tasks as well and it is expected that they perform much better for native scripts as compared to code-mixed texts. Due to hardware resource constraints, only LR classifier was trained with char sequences and syllable n-grams for Ta-En language pair. According to the performances of the models on the Dev set (highlighted scores in Table 5), the best feature set and classifier pair are selected and applied on the Test sets. The results of the best individual classifier and feature set pairs and their majority voting on the Test sets are given in Table 6. It can be observed that the performance of the majority voting of the predictions outperformed the performances of the individual classifiers. The results released by the shared task organizers in the leaderboard12 reveals that our proposed methodology using majority voting of the predictions obtained 2nd , 4th , and 5th ranks with average weighted F1-scores of 0.628, 0.726, and 0.619 for Ka-En, Ma-En and Ta-En language pairs respectively. The confusion matrix for each language pair based on the best performances as mentioned in Table 6 are presented in Figure 3. For both Ka-En (Figure 3a) and Ta-En (Figure 3c) language pairs, the weakest performances are for predicting "Mixed_feelings" comments and the best performances are for predicting "Positive" comments. Similarly, for Ma-En (Figure 3b) language pair, the weakest performance is for predicting "Mixed_feelings" comments and the good performance is for predicting "not-Malayalam" comments along with "Positive" comments. Though predicting "Mixed_feelings" comments exhibits weakest performance in all the three language pairs, the results of Ma-En language pair are higher compared to that in other two 12 https://dravidian-codemix.github.io/2021/proceedings.html Figure 3: Confusion matrix language pairs. The comparison of the performances of the proposed methodology with that of the top performing models in the shared task shown in Figure 4 illustrates that the performances are quite competitive for all the language pairs. Ka-En language pair which has a smaller dataset compared to other language pairs also has given good performance. Figure 4: Comparison of the performances of the proposed models with the top performing models in shared task 5. Conclusion and Future Work This paper describes the participation of our team MUCIC in SA shared task at Dravidian- CodeMix-FIRE2021. Three types of features, namely: char, char sequences and syllables are extracted from the given texts. These features are used to generate corresponding n-grams which are then transformed to TFIDF vectors for training the classifiers. According to the performances of the models on the Dev set, the best feature set and classifier pair are selected and applied on the Test sets and the majority voting of their predictions were submitted to the shared task organizers for evaluation. The results on the leaderboard reveals that our proposed strategy obtained promising results and secured 2nd , 4th , and 5th ranks with average weighted F1-scores of 0.628, 0.726, and 0.619 for Ka-En, Ma-En and Ta- En language pairs respectively. Other features and feature selection algorithms will be explored further for code-mixed low resource Dravidian languages. Acknowledgments Team MUCIC sincerely appreciate the organizers for their efforts to conduct this shared task. References [1] F. Balouchzahi, H. Shashirekha, LA-SACo: A Study of Learning Approaches for Sentiments Analysis in Code-Mixing Texts, in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, 2021, pp. 109–118. [2] F. Balouchzahi, H. L. Shashirekha, MUCS@Dravidian-CodeMix-FIRE2020: SACO- Sentiments Analysis for CodeMix Text, in: P. Mehta, T. Mandl, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2020 - Forum for Information Retrieval Evaluation, Hyderabad, India, December 16-20, 2020, volume 2826 of CEUR Workshop Proceedings, CEUR-WS.org, 2020, pp. 495–502. [3] B. R. Chakravarthi, R. Priyadharshini, V. Muralidaran, S. Suryawanshi, N. Jose, E. Sherly, J. P. McCrae, Overview of the Track on Sentiment Analysis for Dravidian Languages in Code-mixed Text, in: Working Notes of FIRE 2020 - Forum for Information Retrieval Evaluation, Hyderabad, India, December 16-20, 2020, volume 2826 of CEUR Workshop Proceedings, 2020, pp. 21–24. [4] N. Jose, B. R. Chakravarthi, S. Suryawanshi, E. Sherly, J. P. McCrae, A Survey of Cur- rent Datasets for Code-Switching Research, in: 2020 Sixth International Conference on Advanced Computing and Communication Systems (ICACCS), IEEE, 2020, pp. 136–141. [5] B. R. Chakravarthi, R. Priyadharshini, S. Thavareesan, D. Chinnappa, D. Thenmozhi, E. Sherly, J. P. McCrae, A. Hande, R. Ponnusamy, S. Banerjee, C. Vasantharajan, Findings of the Sentiment Analysis of Dravidian Languages in Code-Mixed Text, in: Working Notes of FIRE 2021 - Forum for Information Retrieval Evaluation, CEUR, 2021. [6] R. Priyadharshini, B. R. Chakravarthi, S. Thavareesan, D. Chinnappa, T. Durairaj, E. Sherly, Overview of the Dravidian CodeMix 2021 Shared Task on Sentiment Detection in Tamil, Malayalam, and Kannada, in: Forum for Information Retrieval Evaluation, FIRE 2021, Association for Computing Machinery, 2021. [7] B. R. Chakravarthi, N. Jose, S. Suryawanshi, E. Sherly, J. P. McCrae, A Sentiment Analysis Dataset for Code-Mixed Malayalam-English, in: Proceedings of the First Joint Workshop on Spoken Language Technologies for Under-resourced Languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), European Language Resources Association, Marseille, France, 2020, pp. 177–184. URL: https://aclanthology.org/2020.sltu-1. 25. [8] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2019, pp. 4171–4186. [9] A. Conneau, K. Khandelwal, N. Goyal, V. Chaudhary, G. Wenzek, F. Guzmán, É. Grave, M. Ott, L. Zettlemoyer, V. Stoyanov, Unsupervised Cross-lingual Representation Learning at Scale, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 8440–8451. [10] B. R. Chakravarthi, R. Priyadharshini, N. Jose, T. Mandl, P. K. Kumaresan, R. Ponnusamy, R. Hariharan, J. P. McCrae, E. Sherly, et al., Findings of the Shared Task on Offensive Language Identification in Tamil, Malayalam, and Kannada, in: Proceedings of the First Workshop on Speech and Language Technologies for Dravidian Languages, 2021, pp. 133–145. [11] T. Pires, E. Schlinger, D. Garrette, How Multilingual is Multilingual BERT?, in: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019, pp. 4996–5001. [12] R. Sun, X. Zhou, SRJ@ Dravidian-CodeMix-FIRE2020: Automatic Classification and Identification Sentiment in Code-mixed Text, in: FIRE (Working Notes), 2020, pp. 548–553. [13] A. Conneau, G. Lample, Cross-lingual Language Model Pretraining, Advances in Neural Information Processing Systems 32 (2019) 7059–7069. [14] X. Ou, H. Li, YNU@ Dravidian-CodeMix-FIRE2020: XLM-RoBERTa for Multi-language Sentiment Analysis, in: FIRE (Working Notes), 2020, pp. 560–565. [15] H. Sun, J. Gao, F. Sun, HIT_SUN@ Dravidian-CodeMix-FIRE2020: Sentiment Analysis on Multilingual Code-Mixing Text Base on BERT, in: FIRE (Working Notes), 2020, pp. 517–521. [16] B. Huang, Y. Bai, LucasHub@ Dravidian-CodeMix-FIRE2020: Sentiment Analysis on Multilingual Code Mixing Text with M-BERT and XLM-RoBERTa, in: FIRE (Working Notes), 2020, pp. 574–581. [17] Y. Zhu, K. Dong, YUN111@ Dravidian-CodeMix-FIRE2020: Sentiment Analysis of Dravid- ian Code Mixed Text, in: FIRE (Working Notes), 2020, pp. 628–634. [18] N. Kanwar, M. Agarwal, R. K. Mundotiya, PITS@ Dravidian-CodeMix-FIRE2020: Tradi- tional Approach to Noisy Code-Mixed Sentiment Analysis, in: FIRE (Working Notes), 2020, pp. 541–547. [19] I. TOMEK, Two Modifications of CNN, IEEE Trans. Systems, Man and Cybernetics 6 (1976) 769–772. [20] J. P. Posadas-Durán, I. Markov, H. Gómez-Adorno, G. Sidorov, I. Batyrshin, A. Gelbukh, O. Pichardo-Lagunas, Syntactic N-grams as Features for the Author Profiling Task, in: CEUR Workshop Proceedings, 2015. [21] G. Sidorov, F. Velasquez, E. Stamatatos, A. Gelbukh, L. Chanona-Hernández, Syntactic Dependency-based n-grams: More Evidence of Usefulness in Classification, in: Interna- tional Conference on Intelligent Text Processing and Computational Linguistics, Springer, 2013, pp. 13–24. [22] K. De Jong, Temporal Constraints and Characterising Syllable Structuring, Phonetic Interpretation. Papers in Laboratory Phonology VI (2003) 253–268. [23] G. Sidorov, Automatic Authorship Attribution using Syllables as Classification Features, Rhema. (2018).