<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Sexism Prediction in Spanish and English Tweets Using Monolingual and Multilingual BERT and Ensemble Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Angel</forename><surname>Felipe</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Magnossão</forename><surname>De Paula</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Roberto</forename><surname>Fray Da Silva</surname></persName>
							<email>roberto.fray.silva@gmail.com</email>
						</author>
						<author>
							<persName><forename type="first">Ipek</forename><forename type="middle">Baris</forename><surname>Schlicht</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Escola Politécnica da Universidade de São Paulo</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Sexism Prediction in Spanish and English Tweets Using Monolingual and Multilingual BERT and Ensemble Models</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">826F9BE7048F44B479F0FEB7CC0A3B2A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:22+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sexism identification</term>
					<term>Sexism classification</term>
					<term>BERT</term>
					<term>Deep learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The popularity of social media has created problems such as hate speech and sexism. The identification and classification of sexism in social media are very relevant tasks, as they would allow building a healthier social environment. Nevertheless, these tasks are considerably challenging. This work proposes a system to use multilingual and monolingual BERT and data points translation and ensemble strategies for sexism identification and classification in English and Spanish. It was conducted in the context of the sEXism Identification in Social neTworks shared 2021 (EXIST 2021) task, proposed by the Iberian Languages Evaluation Forum (IberLEF). The proposed system and its main components are described, and an in-depth hyperparameters analysis is conducted. The main results observed were: (i) the system obtained better results than the baseline model (multilingual BERT); (ii) ensemble models obtained better results than monolingual models; and (iii) an ensemble model considering all individual models and the best standardized values obtained the best accuracies and F1-scores for both tasks. This work obtained first place in both tasks at EXIST, with the highest accuracies (0.780 for task 1 and 0.658 for task 2) and F1-scores (F1-binary of 0.780 for task 1 and F1-macro of 0.579 for task 2).</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The emergence of social networks and microblogs has created a new medium for people to express themselves, providing freedom of speech and the possibility for quickly spreading opinions, news, and information <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b13">14]</ref>. This has impacted considerably on peoples' lives, by increasing access to all kinds of information.</p><p>Nevertheless, a small part of the users employs those media for spreading hate messages, increasing the impacts of racism, sexism, and others types of prejudices and hate speech <ref type="bibr" target="#b31">[32,</ref><ref type="bibr" target="#b9">10]</ref>.</p><p>One crucial problem faced by the different stakeholders related to social media platforms is detecting hate speech <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b34">35,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b4">5]</ref>, both in general and issuespecific forms. Also, some types of hate speech tend to be more challenging to identify, as they present characteristics such as irony or sarcasm, among others <ref type="bibr" target="#b34">[35,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b22">23]</ref>. Sexism is a type of toxic language that could be used both as hate speech and, in a much more subtle way, as sarcasm. Sexism is related to all kinds of behaviors and content that aim to spread prejudice against women, reduce their importance in society, or behave aggressively or offensively <ref type="bibr" target="#b26">[27,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b22">23]</ref>. There are several forms of sexism, and identifying them in social media messages is a fundamental challenge among the various natural language processing (NLP) tasks <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b34">35,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b25">26]</ref>.</p><p>The detection of sexism can be broken into two main tasks: (i) sexism identification, which aims to identify if a message or post containing sexist contents (regardless of the type of sexism contained in it); and (ii) sexism classification, which aims to classify the type of sexism contained in a given sexist message or post <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b25">26,</ref><ref type="bibr" target="#b10">11]</ref>. Both are very relevant, and the second task is dependent on the first, as it needs posts that are confirmed as sexist as inputs for the different classification models. Additionally, the difficulty of using data-driven models may increase for languages that are more complex or that have fewer resources available, such as high-quality word embeddings, pre-trained language-specific models, task-specific lexicons, among others.</p><p>To advance the state-of-the-art knowledge in both sexism identification and classification on social media messages, the Iberian Languages Evaluation Forum (IberLEF) proposed the sEXism Identification in Social neTworks shared 2021 (EXIST 2021) shared task. For the rest of this work, this challenge will be referred to as EXIST shared task. The main goal of the IberLEF forum is to promote scientific advances towards innovative solutions for detecting sexism on social media platforms <ref type="bibr" target="#b18">[19]</ref>. For this reason, the 2021 shared task provided datasets in English and Spanish for both tasks labeled by experts, following state-of-the-art data collection and labeling procedures <ref type="bibr" target="#b18">[19]</ref>. Those datasets are expected to become benchmarks for state-of-the-art research on sexism identification and classification on social media messages.</p><p>Therefore, a relevant gap in the literature is to develop data-driven models that better identify and classify sexist content on social media messages, considering the implementation in different languages. This would: (i) advance both the knowledge on the use of artificial intelligent models for data-driven sexism identification and detection; (ii) provide a better methodology for identifying and classifying sexist content, which is highly relevant for identifying unacceptable user behavior; and (iii) address the problem of generalizing the model throughout different languages. Related to this gap, it is vital to observe that identifying online sexism can be considerably challenging because posts may have several forms: they may sound hateful and offensive, or friendly and funny, misleading the current classifying models used for this task <ref type="bibr" target="#b26">[27]</ref>.</p><p>State of the art systems for addressing those tasks for multiple languages uses the Bidirectional Encoder Representations from Transformers (BERT) multilingual model, an NLP model that uses transformers and is pre-trained on a comprehensive text corpora <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b15">16]</ref>. This model is trained on datasets of multiple languages, but it is not language-specific. The pre-trained models are then fine-tuned on task-specific datasets on the target language.</p><p>The main goal of this work is to propose and evaluate a system to identify and classify sexist content in social media messages in multiple languages, using the EXIST 2021 shared task dataset <ref type="bibr" target="#b18">[19]</ref> for implementation and evaluation. The official shared task metrics were used: accuracy for task 1 (sexism detection) and F1-macro for task 2 (sexism classification). However, we also implemented other relevant metrics for NLP tasks to better evaluate the different models in relation to the state of the art baseline model, the multilingual BERT: precision and recall.</p><p>The three main research questions that are going to be addressed in this work are: (i) does the use of monolingual BERT models provides better results than the multilingual BERT model to identify and classify sexist content on social media messages in English and Spanish?; (ii) does the use of an ensemble strategy improves the results of the individual models?; and (iii) does the results differ between the English and Spanish languages? Besides answering those three questions, this work also conducts an in-depth analysis of the main hyperparameters for the implemented models for both languages.</p><p>The main contribution of this work is to propose and evaluate the sexism identification and classification system in multiple languages considering different components: monolingual BERT models, multilingual BERT, data points translation, and different ensemble strategies. We also explore the main hyperparameters of the implemented models in-depth, comparing the final models with the state-of-the-art BERT multilingual model. This work obtained first place in both sexism identification and classification tasks at the EXIST shared task <ref type="bibr" target="#b18">[19]</ref>.</p><p>This work is organized in the following sections: section 2 describes the main concepts and models used for sexism prediction in social media messages; section 3 contains the main steps of the methodology used; section 4 describes the proposed system for addressing both sexism identification and classification; section 5 contains the main results of the system's implementation on the EXIST shared task dataset; section 6 contains a discussion of relevant topics on the system's use, modification, and potential improvements; and section 7 concludes this paper.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Sexism identification and classification using artificial intelligence models</head><p>The works by <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b5">6]</ref> explore in-depth the impacts of the different types of sexism on social media platforms, describing several important classes of sexism.</p><p>As sexism is an important type of hate speech, we also refer the reader to the works by <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b34">35]</ref> for excellent reviews on identifying and classifying the different forms of hate speech. The main concepts observed in those works were considered in our approach. This work addresses two very relevant tasks: (i) sexism identification in natural language texts; and (ii) classification of types of sexism in natural language texts. Some examples of works that addressed the first task are <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b23">24]</ref>. Examples that addressed the second task are <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b28">29]</ref>.</p><p>It is essential to observe that the second task is considerably more complex because different languages can be used in the different classes (as well as the traditional problems related to social media messages: abbreviations, emojis, misspellings, memes, among others).</p><p>Although there are a variety of different models and strategies used for sexism detection and classification, the most traditionally used models are: support vector machines (SVM), convolutional neural networks (CNN), long short-term neural networks (LSTM), and BiLSTM <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b34">35]</ref>. In the last years, the BERT has been widely used <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b26">27,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b34">35]</ref>. This model (and its variations) have presented the best results on those tasks, as observed in the works by <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b30">31,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b24">25]</ref>.</p><p>The NLP literature addresses several identification and classification tasks related to extracting and evaluating opinions from natural language texts. In general, those tasks are addressed using three main approaches <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b29">30,</ref><ref type="bibr" target="#b21">22]</ref>: (i) lexical-based, in which specific dictionaries (lists of words with corresponding values on important dimensions for the task) are used to classify the input text; (ii) statistical learning or machine learning-based, in which machine learning and deep learning models are used, generally with word embeddings or bag of words models, to classify the text; and (iii) hybrid, in which both lexicons and machine learning models are used.</p><p>However, it is essential to note that: (i) lexical-based systems are not able to learn (and could be improved by using a deep learning model, such as BERT); (ii) deep learning models, especially the multilingual BERT, are state-of-theart on sexism identification and classification <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b15">16]</ref>; (iii) lexicons tend to be language-specific, making it more challenging to apply the solutions for multiple languages; and (iv) few works use BERT with domain-specific lexicons for sexism identification and classification. One of those lexicons that are highly relevant in this context is the Hurtlex lexicon <ref type="bibr" target="#b2">[3]</ref>. This was used in the works by <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b15">16]</ref>, among several others.</p><p>The BERT model was proposed by <ref type="bibr" target="#b8">[9]</ref> and can be described as a language learning model aimed at providing a general structure that can be further refined with fine-tuning on specific tasks and domains. Its main objective is to learn the main features and semantics of a language, based on semi-supervised learning on a vast text corpora (such as the BookCorpus and the Wikipedia database) <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b27">28,</ref><ref type="bibr" target="#b0">1]</ref>. Its architecture and training workflow is composed of three main components: transformers (which is an advanced deep learning model), bidirectional training, and use of encoder representations <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b27">28,</ref><ref type="bibr" target="#b0">1]</ref>.</p><p>In this work, we use the multilingual BERT <ref type="bibr" target="#b8">[9]</ref>, the English version of the model <ref type="bibr" target="#b8">[9]</ref>, and the Spanish version of the model, called BETO <ref type="bibr" target="#b3">[4]</ref>. For an indepth analysis of how the BERT model works, we refer the reader to <ref type="bibr" target="#b27">[28]</ref>. For an in-depth comparison with multilingual BERT with other models, as well as an in-depth description of how they work, we refer the reader to <ref type="bibr" target="#b33">[34]</ref>.</p><p>However, very few works in the literature consider dealing with datasets with multiple languages. This work addressed this gap by proposing a system that contains multiple models and ensemble strategies.</p><p>This paper aims to fulfill the gap of evaluating monolingual and multilingual BERT models for identifying and classifying sexism in texts in multiple languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Methodology</head><p>The methodology used in this work was composed of six steps. Figure <ref type="figure" target="#fig_0">1</ref> illustrates the strategy used to tackle each of the tasks. For task 1 (sexism identification), the classification models considered two labels: 0 (non-sexist) and 1 (sexist). For task 2 (sexism classification), the tweets labeled as non-sexist (from task 1) were eliminated. Then, the classification models were used to predict the following sexism categories on the remaining tweets: ideological and inequality; stereotyping and dominance; objectification; sexual violence; and misogyny and non-sexual violence. For a thorough description of those classes, we refer to the EXIST shared task at IberLEF 2021 <ref type="bibr" target="#b18">[19]</ref>, which developed and labeled the dataset that was used in this research.</p><p>The steps of the methodology were:</p><p>1. Data collection: we used the dataset developed for the EXIST shared task at IberLEF 2021 <ref type="bibr" target="#b18">[19]</ref>. This dataset contained labeled data from two social media platforms: Twitter and Gab. For an in-depth description of this dataset, we refer the reader to Section 5 of this work;</p><p>2. Data processing: for both tasks, we used the following processing techniques: separation of the dataset between languages (English and Spanish), tokenization, lemmatization, and elimination of stop words. These are widely used in the literature for the implementation of machine learning models on NLP tasks, such as hate speech detection, sexism identification, sentiment analysis, among others <ref type="bibr" target="#b30">[31,</ref><ref type="bibr" target="#b24">25,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b19">20,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b21">22]</ref>. There was no need to eliminate data points from the datasets, as the shared task organizers had already thoroughly curated them. The training subset was then divided into training (80%) and validation (20%) for cross-validation purposes. Additionally, one of the training strategies used for some of the models implemented involved translating the social media messages from one language to the other (for example, from English to Spanish to train a Spanish language model). This strategy doubled the number of data points available for the single language models (even if part of the meaning may have been lost during the translation process). The googletrans (https://github.com/ssut/py-googletrans) library was used for the translation process; 3. Exploratory data analysis: in this step, an exploratory analysis of the dataset was conducted to understand better the different class distributions on both tasks throughout the training dataset. No data imbalance problems were observed;</p><p>4. Model implementation and hyperparameters analysis: in this research, we implemented the following models: (i) the BERT Multilingual model or mBERT <ref type="bibr" target="#b8">[9]</ref> (named M1 in this research); (ii) single language models (one for English and one for Spanish, named M2-English and M2-Spanish); (iii) single language models with translated data points (one for English and one for Spanish, named M3-English and M3-Spanish); and (iv) ensemble models (used only for the test subset). All the implementations were conducted with the Hugging Face BERT implementation library (https://huggingface.co/transformers/index.html) <ref type="bibr" target="#b32">[33]</ref>, with 10-fold cross-validation on the training stage. A thorough hyperparameters analysis was conducted, considering the following hyperparameters and values: output BERT type (hidden or pooler), batch size (32 and 64), learning rate (0.00002, 0.00003, and 0.00005), and number of epochs (1 to 8). Following the official metrics of the EXIST 2021 shared task, accuracy was used as the quality metric for model training on task 1, and F1-macro was used on task 2. Besides this metric, an analysis of model overfitting was conducted for each model, based on charts that contained the models' accuracies on the different epochs;</p><p>5. Final models implementation: the final models and model ensembles were built using the best hyperparameters identified in Step 4. They were then trained on the whole training datasets (training plus validation subsets). Table <ref type="table" target="#tab_0">1</ref> contains all the final models implemented: (i) M1: multilingual model; (ii) separated single language models without translation on the training datasets (M2, composed of M2-English and M2-Spanish) and with translation on the training datasets (M3, composed of M3-English and M3-Spanish); (iii) English single language model with translation only on the test subset (M4) and on training and test subsets (M5), both derived from the M3-English model; (iv) Spanish single language model with translation only on the test subset (M6) and on training and test subsets to Spanish (M7), both derived from the M3-Spanish model; (v) ensembles considering only the best models: E1 (majority vote), E2 (higher unstandardized value), and E3 (higher standardized value); and (vi) ensembles considering all the models: E4 (majority vote), E5 (higher unstandardized value), and E6 (higher standardized value); 6. Models comparison: the final comparison of all models was then conducted on the test subsets. The official metrics for the EXIST shared task at IberLEF 2021 <ref type="bibr" target="#b18">[19]</ref> were considered the quality metrics for both sexism identification and classification tasks. For task 1, we evaluated the accuracy, precision, recall, and F1-binary metrics. For task 2, we evaluated the accuracy, precision, recall, and F1-macro metrics). Additionally, an analysis of a sample of correctly and incorrectly classified data points on the test set was conducted, aiming to better understand each model's main strengths, weaknesses, and opportunities for future improvements. Lastly, the best model was chosen.</p><p>The implementation was done using Python on a Google Collaboratory Pro (https://colab.research.google.com/) TPU , with the following technical specifications: Intel(R) Xeon(R) CPU @ 2.30GHz CPU, 26GB of RAM, and TPU v2. The code implemented is available on an open Github repository (https:// github.com/AngelFelipeMP/BERT-tweets-sexims-classification). Section 5 presents the main results of the exploratory data analysis and the model implementations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Proposed system: components and implementation</head><p>The proposed system considered two separate workflows: training and testing. The objective of the training workflow was to fine-tune the pre-trained BERT models. It considered three options: (i) using a multilingual BERT model (illustrated in Figure <ref type="figure" target="#fig_1">2</ref>), which was also considered our baseline, since it is the state of the art for multilingual NLP classification tasks; (ii) using monolingual BERT models without data points translation (illustrated in Figure <ref type="figure" target="#fig_2">3</ref>); and (iii) using monolingual BERT models with data points translation (illustrated in Figure <ref type="figure" target="#fig_2">3</ref>). It is essential to observe that those three options considered 10-fold crossvalidation on training to identify the best hyperparameter values for each model. The result from the first option was the M1 model. The results from the second option were the models M2-English and M2-Spanish, which would be used as components of the M2 model. The results from the third option were the models M3-English and M3-Spanish, which would be used as components of the M3 model.</p><p>Figure <ref type="figure" target="#fig_3">4</ref> illustrates the test workflow. The objective of this workflow was to use the previously tested models as components for the final models, their training on the whole training dataset (training plus validation subsets), and testing on the test subset. This workflow also introduces the six ensemble models implemented, considering different model configurations and rules for generating the models. It is vital to observe that this model could be easily expanded for other languages, quality metrics, and data sources. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Results</head><p>This section contains the main research results and is divided into three subsections: 5.1 contains a description of the dataset used; 5.2 contains the main results and observations related to the hyperparameters and 5.3 contains the final models' comparison on the test subset, considering four metrics: accuracy, precision, recall, and F1-score (F1-binary for task 1 and F1-macro for task 2).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Description of the EXIST 2021 shared task dataset</head><p>The dataset from EXIST 2021 shared task at IberLEF 2021 <ref type="bibr" target="#b18">[19]</ref> was used in this work. This dataset contained labeled data from two social media platforms: (i) Twitter, with 6,977 tweets for training and 3,386 tweets for testing (both subsets equally distributed between English and Spanish); and (ii) Gab, with 492 gabs in English and 490 gabs in Spanish (used only for testing purposes). It is important to note that Gab is an uncensored social media website with considerably fewer users than Twitter.</p><p>It is vital to observe that the labeling procedure adopted by the shared task organizers considered both experts and crowdsourcing labeling (considering a specific procedure developed by experts in this domain). The dataset distribution was balanced on the training and test subsets. For a thorough description of the dataset, we refer the readers to IberLEF 2021 <ref type="bibr" target="#b18">[19]</ref>.</p><p>The five classes that were used in this work for the sexism classification task (also referred to by the organizers of the dataset as sexism categorization) are the ones provided by the EXIST challenge dataset <ref type="bibr" target="#b18">[19]</ref>. These classes contain, as described by <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b26">27]</ref>:</p><p>-Ideological and inequality: texts that affirm that the feminist movement deserves no credits, rejects the existence of inequality between genders, or claims that men are oppressed gender; -Stereotyping and dominance: texts that claim that women are inappropriate for specific tasks, suitable only for specific roles, or that men are superior to women; -Objectification: texts that claim that women should have certain physical qualities or that separate women from their dignity and personal aspects; -Sexual violence: texts that contain sexual suggestions or sexual harassment; -Misogyny and non-sexual violence: texts that express different forms of hatred and violence towards women.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Hyperparameters analysis</head><p>Due to the considerable difference between the identification and classification tasks, this section will analyze both separately and then conclude with a comparison of the best hyperparameters values for all models for both tasks. The results observed in this subsection can be used as a guide for further implementations of BERT models for sexism identification and classification, considering multiple languages.</p><p>Table <ref type="table" target="#tab_1">2</ref> illustrates the best hyperparameters values for task 1 for each of the five models implemented on the training step of the proposed system (M1, M2-Eng, M2-Sp, M3-Eng, and M3-Sp), as well as their associated quality on the validation subset. Based on the analysis of Table <ref type="table" target="#tab_1">2</ref>, it is important to observe that: (i) the M2-Eng (monolingual without translation) presented better results for the English language; (ii) the M2-Sp (monolingual without translation) presented better results for the Spanish language; (iii) the F1-binary for the Spanish language models (M2-Sp and M3-Sp) presented better results than for the English language models (M2-Eng and M3-Eng); (iv) most of the models presented better results by using the hidden output BERT type; (v) all models presented a learning rate of 0.00005; (vi) most models presented better results with a batch size of 32; and (vii) most models presented better results with 6 or more epochs. We have focused the analysis and model choice considering the accuracy, as it was the official metric for the EXIST shared task.</p><p>Table <ref type="table" target="#tab_2">3</ref> illustrates the best hyperparameters values for task 2 for each of the five models implemented on the training step of the proposed system and their quality metrics on the validation subset. It is possible to observe that: (i) similar to task 1, the M1 model (multilingual) did not present better results than any of the languages; (ii) the M3-Eng (monolingual with translation) presented the best results for the English language; (iii) the M3-Sp (monolingual with translation) presented better results for the Spanish language; (iv) most of the models presented better results by using the hidden output BERT type; (v) like in task 1, all models presented better results by using a learning rate of 0.00005; (vi) most models presented better results with a batch size of 32; and (vii) most models presented better results with 7 or 8 epochs.</p><p>Lastly, Table <ref type="table" target="#tab_3">4</ref> contains a cross-model and cross-language analysis of the results on the validation subset. It presents the hyperparameter values of the best models in each category (M1, M2-Eng, M2-Sp, M3-Eng, and M3-Sp), as a percentage of the total number of models, for each task. For example, on the first cell, it is possible to observe that, for the output BERT type on task 1, 60% of the final models contained a hidden output BERT, while 40% used the pooler type. Based on an analysis of this table, it is possible to conclude that: (i) for both tasks, the hidden output BERT type provided the best results; (ii) higher learning rates (0.00005) presented the best results for both tasks; (iii) the best batch size for both tasks was 32; and (iv) the best number of epochs differed among tasks, probably due to their different nature. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Final models comparison</head><p>Table <ref type="table" target="#tab_4">5</ref> contains the results of the thirteen final models on the test subset for both tasks, considering the following metrics: accuracy, precision, recall, and F1score (F1-binary for task 1 and F1-macro for task 2). For both tasks, it is vital to observe that: (i) the ensemble models presented a better F1-score than the monolingual models and the multilingual model; and (ii) the E6 (ensemble model considering all individual models and the best standardized values) obtained the best accuracy and F1-score. For task 1, it is essential to observe on Table <ref type="table" target="#tab_4">5</ref> that: (i) the baseline model (M1) presented better accuracy results than the M4, M6, and M7 models; (ii) the E4 model presented results that were comparable to the E6 model in terms of accuracy, both being considered the best models for this task; (iii) the E1 model presented the best precision; and (iv) the M2 model presented the best recall. For task 2, it can be observed that: (i) with an exception for M4, all models presented a better F1-macro than the M1 model, indicating that the use of monolingual may provide significantly better results than multilingual models for sexism classification; and (ii) the E6 model presented the best results for all metrics, indicating that it outperformed all other models for this task.</p><p>Table <ref type="table" target="#tab_5">6</ref> presents a comparison of the best individual and ensemble models for tasks 1 and 2, considering the two official metrics of the EXIST shared task: accuracy and F1-score. Considering the differences of the F1-scores of each model and the best model, it is possible to conclude that: (i) the differences are significantly higher for the task of sexism classification; (ii) the baseline model (M1) obtained the worst F1-score among those models (around 3% lower for task 1 and 11% for task 2 in comparison to the E6 model); (iii) the baseline model (M1) obtained the worst accuracy for both tasks; and (iv) although the E4 model presented similar results for task 1, it observed a 6.39% lower F1-score in comparison to the E6 model for task 2. The analysis of the models' accuracies lead to the same conclusions.</p><p>It is also possible to observe that the same overall conclusions hold for the accuracy metric. Our approach ranked first in both sexism identification and classification tasks at EXIST, with the highest accuracies (0.780 for task 1 and 0.658 for task 2) and F1-scores (F1-binary of 0.780 for task 1 and F1-macro of 0.579 for task 2), considering the E6 model. We also observed that ensemble models provide a better generalization.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Discussions</head><p>This section briefly explores several important aspects related to the system proposed in this work and its results, encompassing the following topics: implementation aspects, system design, use of ensembles, system adaptation for other languages, results obtained concerning the literature, impacts of the different system components, and the use of the proposed system in real scenarios.</p><p>It is vital to note that the system proposed in this work can be extended using additional components with few adaptations to the code. Some additional interesting components to explore are lexicons (both generalist, such as Vader <ref type="bibr" target="#b11">[12]</ref> and domain-specific, such as Hurtlex <ref type="bibr" target="#b2">[3]</ref>), word embeddings, and transfer learning (via training on multiple datasets). Additional models could also be implemented to improve feature engineering (such as unsupervised learning models) or improve prediction quality (such as different weak models used in an ensemble strategy).</p><p>Although ensemble models are relatively common in other domains, such as price prediction <ref type="bibr" target="#b1">[2]</ref> and sentiment analysis <ref type="bibr" target="#b6">[7]</ref>, they are not widely spread on sexism identification and classification, as this is a new task. In general, if the behavior of the weak models can capture different aspects of the task, an ensemble strategy could improve the final prediction results <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b7">8]</ref>. In this work, we evaluated several simple average ensemble strategies. However, an in-depth analysis of more complex ensemble strategies with the proposed system could be conducted in future works. As was observed in this work, the use of ensembles can significantly improve the results obtained by the individual models.</p><p>Another important aspect is related to adapting the proposed system for other languages. Concerning this aspect, it is vital to separate the languages into two main groups: (i) languages with individual BERT models already implemented; and (ii) languages that currently have no widely accepted individual BERT model implemented. In the first group, the system allows for easy implementation with minimal coding needed: the only components needed are the BERT individual model pre-trained on that language and the task-specific dataset for fine-tuning and testing.</p><p>In the second group, it is necessary to train a language-specific BERT model before using the proposed system. This task demands a considerable amount of computation power and resources, demanding processing clusters and large text corpora in the target language (such as the Wikipedia text database). However, adaptations can be implemented in the proposed system to use different models that are easier to implement and demand fewer data, such as recurrent neural networks or convolutional neural networks with language-specific word embeddings or lexicons. The ensemble component can then be used, considering the BERT multilanguage (if it encompasses the target language) and the implemented models.</p><p>Lastly, the proposed system can be implemented and used in real case scenarios to improve sexism identification on social media platforms. After the hyperparameters and final models are chosen, as described and explored in this work, the prediction process is considerably fast. The system has the potential to be implemented as a separate service on the social media platform, analyzing the content published by its users and pointing out sexist messages and their respective classes (considering the five classes studied in this work).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusion and future work</head><p>As was explored throughout this work, a widespread problem on social networks and microblogs is the misuse of these tools to spread toxic language and sexist content. Identifying and detecting sexism in those media is considerably challenging, especially in a scenario with multiple languages. This paper explored the fine-tuning of multilingual and monolingual BERT models for English and Spanish and the use of different ensemble configurations to identify and classify sexism on tweets and gabs. The dataset used was provided by the EXIST shared task, which contained two tasks: sexism identification and sexism classification.</p><p>The proposed system in this research considered the use of fine-tuning of pretrained BERT models and the translation of the training dataset (to increase the number of data points used by the model for learning) and ensemble models with different characteristics. Our central hypothesis was that using this system would provide better results than the traditional use of the multilingual BERT model. Our results have shown that the use of ensembles provided better results for both tasks, primarily the ensemble that considered all trained models and the higher standardized label values as the final predictions. This model obtained significantly better results than the baseline multilingual BERT model, with an F1-score around 3% higher for the sexism identification task and 11% higher for the sexism classification task. Those results and models and the in-depth hyperparameters analysis that was conducted can be used as a guide for future research on both tasks.</p><p>Future works are related to: (i) conducting an analysis considering additional datasets; (ii) implementing additional models; (iii) implementing different ensemble configurations; (iv) implementing unsupervised models for feature engineering; (v) analyzing the impacts on the models' results of using lexicons (both general and domain-specific) as features; (vi) analyzing the impacts on the models' results of using word embeddings as features; and (vii) implementing and evaluating the use of deep reinforcement learning to improve the models' results, especially on the sexism classification problem.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Illustration of the two main tasks evaluated in this research.</figDesc><graphic coords="6,218.63,115.83,178.10,272.35" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Workflow for training the Multilingual model (M1).</figDesc><graphic coords="8,134.77,361.35,345.83,114.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 3 .</head><label>3</label><figDesc>Fig.3. Workflow for training the monolingual models (M2-English and M2-Spanish) and the translated languages models (M3-English and M3-Spanish), considering the English models as an example.</figDesc><graphic coords="9,134.77,115.83,345.83,128.30" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Workflow for testing the final models on the test subset.</figDesc><graphic coords="9,175.18,364.05,265.00,252.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Final models implemented and their characteristics</figDesc><table><row><cell cols="2">Model Multilingual</cell><cell cols="3">Translation of data points Observations Training Test</cell></row><row><cell>M1</cell><cell>X</cell><cell></cell><cell></cell><cell>Baseline model</cell></row><row><cell>M2</cell><cell></cell><cell></cell><cell></cell><cell>Separated single language models</cell></row><row><cell>M3</cell><cell></cell><cell>X</cell><cell></cell><cell>Separated single language models</cell></row><row><cell>M4</cell><cell></cell><cell></cell><cell>X</cell><cell>English single language model</cell></row><row><cell>M5</cell><cell></cell><cell>X</cell><cell>X</cell><cell>English single language model</cell></row><row><cell>M6</cell><cell></cell><cell></cell><cell>X</cell><cell>Spanish single language model</cell></row><row><cell>M7</cell><cell></cell><cell>X</cell><cell>X</cell><cell>Spanish single language model</cell></row><row><cell>E1</cell><cell cols="4">Depends on the individual models used Majority vote, best models</cell></row><row><cell>E2</cell><cell cols="4">Depends on the individual models used Unstandardized value, best models</cell></row><row><cell>E3</cell><cell cols="4">Depends on the individual models used Standardized value, best models</cell></row><row><cell>E4</cell><cell cols="4">Depends on the individual models used Majority vote, all models</cell></row><row><cell>E5</cell><cell cols="4">Depends on the individual models used Unstandardized value, all models</cell></row><row><cell>E6</cell><cell cols="4">Depends on the individual models used Standardized value, all models</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Best hyperparameters values, accuracy, precision, recall and F1-binary for the different models on the validation subset of the training dataset for Task 1 -Sexism identification.</figDesc><table><row><cell cols="2">Lang. Model</cell><cell>Best hyperp. values</cell><cell>Acc. Prec. Rec. F1b</cell></row><row><cell cols="2">Multi M1</cell><cell cols="2">OB:pooler / Lr:0.00005 / Bs:32 / Ne:7 0.774 0.749 0.808 0.774</cell></row><row><cell>English</cell><cell cols="3">M2-Eng OB:hidden / Lr:0.00005 / Bs:32 / Ne:5 0.782 0.764 0.788 0.768 M3-Eng OB:pooler / Lr:0.00005 / Bs:64 / Ne:6 0.765 0.748 0.778 0.759</cell></row><row><cell>Spanish</cell><cell cols="3">M2-Sp OB:hidden / Lr:0.00005 / Bs:32 / Ne:8 0.790 0.780 0.795 0.783 M3-Sp OB:hidden / Lr:0.00005 / Bs:32 / Ne:6 0.775 0.756 0.795 0.771</cell></row><row><cell cols="4">Legend: Lang.: model language; OB: output BERT type; Lr: learning rate; Bs: batch</cell></row><row><cell cols="4">size; Ne: number of epochs; Acc.: accuracy; Prec.: precision; Rec.: recall; F1b: F1-binary.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Best hyperparameters values, accuracy, precision, recall and F1-macro for the different models on the validation subset of the training dataset for Task 2 -Sexism classification.</figDesc><table><row><cell cols="2">Lang. Model</cell><cell>Best hyperp. values</cell><cell>Acc. Prec. Rec. F1m</cell></row><row><cell cols="4">Multi M1-Multi OB:pooler / Lr:0.00005 / Bs:32 / Ne:8 0.636 0.632 0.624 0.604</cell></row><row><cell>English</cell><cell cols="3">M2-Eng OB:hidden / Lr:0.00005 / Bs:32 / Ne:8 0.661 0.647 0.633 0.610 M3-Eng OB:hidden / Lr:0.00005 / Bs:32 / Ne:5 0.661 0.660 0.652 0.632</cell></row><row><cell>Spanish</cell><cell cols="3">M2-Sp M3-Sp OB:hidden / Lr:0.00005 / Bs:64 / Ne:7 0.656 0.653 0.650 0.630 OB:hidden / Lr:0.00005 / Bs:32 / Ne:8 0.682 0.656 0.670 0.628</cell></row><row><cell cols="4">Legend: Lang.: model language; OB: output BERT type; Lr: learning rate; Bs: batch</cell></row><row><cell cols="4">size; Ne: number of epochs; Acc.: accuracy; Prec.: precision; Rec.: recall; F1m: F1-</cell></row><row><cell>macro.</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Percentage of best performing models for each hyperparameter value for both tasks and languages, considering the highest accuracy for task 1 and F1-macro for task 2 on the validation subset.</figDesc><table><row><cell></cell><cell cols="2">Hyperparameter values and percentage</cell></row><row><cell>Hyperparameter</cell><cell cols="2">of best performing models</cell></row><row><cell></cell><cell>Task 1</cell><cell>Task 2</cell></row><row><cell>Output BERT type</cell><cell>Hidden: 60% Pooler: 40%</cell><cell>Hidden: 80% Pooler: 20%</cell></row><row><cell></cell><cell>0.00002: 0%</cell><cell>0.00002: 0%</cell></row><row><cell>Learning rate</cell><cell>0.00003: 0%</cell><cell>0.00003: 0%</cell></row><row><cell></cell><cell>0.00005: 100%</cell><cell>0.00005: 100%</cell></row><row><cell>Batch size</cell><cell>32: 80% 64: 20%</cell><cell>32: 80% 64: 20%</cell></row><row><cell></cell><cell>&lt;= 6: 60%</cell><cell>&lt;= 6: 20%</cell></row><row><cell>Number epochs</cell><cell>7: 20%</cell><cell>7: 20%</cell></row><row><cell></cell><cell>8: 20%</cell><cell>8: 60%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>Accuracy, precision, recall and F1-score of the final models on the test subset for tasks 1 and 2.</figDesc><table><row><cell>Model</cell><cell cols="2">Task 1 -Sexism identification Task 2 -Sexism classification Acc. Prec. Rec. F1b Acc. Prec. Rec. F1m</cell></row><row><cell>M1</cell><cell>0.761 0.739 0.784 0.761</cell><cell>0.621 0.617 0.621 0.611</cell></row><row><cell>M2</cell><cell>0.774 0.749 0.803 0.775</cell><cell>0.688 0.677 0.674 0.675</cell></row><row><cell>M3</cell><cell>0.782 0.773 0.778 0.775</cell><cell>0.676 0.661 0.663 0.658</cell></row><row><cell>M4</cell><cell>0.732 0.737 0.693 0.715</cell><cell>0.612 0.607 0.597 0.595</cell></row><row><cell>M5</cell><cell>0.766 0.773 0.730 0.751</cell><cell>0.661 0.648 0.657 0.649</cell></row><row><cell>M6</cell><cell>0.735 0.739 0.699 0.718</cell><cell>0.639 0.629 0.621 0.624</cell></row><row><cell>M7</cell><cell>0.753 0.734 0.769 0.751</cell><cell>0.642 0.625 0.627 0.624</cell></row><row><cell>E1</cell><cell>0.784 0.797 0.744 0.769</cell><cell>0.669 0.683 0.666 0.653</cell></row><row><cell>E2</cell><cell>0.786 0.777 0.791 0.781</cell><cell>0.686 0.673 0.677 0.674</cell></row><row><cell>E3</cell><cell>0.784 0.773 0.784 0.779</cell><cell>0.682 0.669 0.674 0.670</cell></row><row><cell>E4</cell><cell>0.790 0.781 0.787 0.784</cell><cell>0.661 0.673 0.656 0.645</cell></row><row><cell>E5</cell><cell>0.785 0.767 0.799 0.782</cell><cell>0.701 0.687 0.690 0.687</cell></row><row><cell>E6</cell><cell>0.789 0.776 0.794 0.785</cell><cell>0.703 0.690 0.692 0.689</cell></row><row><cell cols="3">Legend: Acc.: accuracy; Prec.: precision; Rec.: recall; F1b: F1b-binary; F1m:</cell></row><row><cell></cell><cell cols="2">F1-macro. M1 is the baseline (BERT Multilingual).</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 .</head><label>6</label><figDesc>Comparison of best individual and ensemble models for tasks 1 and 2 considering accuracy and F1-score.</figDesc><table><row><cell>Model</cell><cell>Task 1 -Sexism identification Task 2 -Sexism classification Acc. Diff E6 F1b Diff E6 Acc. Diff E6 F1m Diff E6</cell></row><row><cell>M1</cell><cell>0.761 -3.55% 0.761 -3.06% 0.621 -11.66% 0.611 -11.32%</cell></row><row><cell>M2</cell><cell>0.774 -1.90% 0.775 -1.27% 0.688 -2.13% 0.675 -2.03%</cell></row><row><cell>M3</cell><cell>0.782 -0.89% 0.775 -1.27% 0.676 -3.84% 0.658 -4.50%</cell></row><row><cell>E4</cell><cell>0.790 0.13% 0.784 -0.13% 0.661 -5.97% 0.645 -6.39%</cell></row><row><cell>E6</cell><cell>0.789 0.00% 0.785 0.00% 0.703 0.00% 0.689 0.00%</cell></row><row><cell cols="2">Legend: Acc.: accuracy; Prec.: precision; Rec.: recall; F1b: F1b-binary; F1m:</cell></row><row><cell cols="2">F1-macro. M1 is the baseline (BERT Multilingual). Diff E6 is the difference between</cell></row><row><cell cols="2">that model's metric and the same metric for the E6 model (best overall model for</cell></row><row><cell cols="2">both tasks).</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Transformer models for textbased emotion detection: a review of bert-based approaches</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">A</forename><surname>Acheampong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Nunoo-Mensah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="page" from="1" to="41" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Evaluating multiple classifiers for stock price direction prediction</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ballings</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">V D</forename><surname>Poel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hespeels</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Gryp</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2015.05.013</idno>
		<ptr target="https://doi.org/10.1016/j.eswa.2015.05.013" />
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="issue">20</biblScope>
			<biblScope unit="page" from="7046" to="7056" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Hurtlex: A multilingual lexicon of words to hurt</title>
		<author>
			<persName><forename type="first">E</forename><surname>Bassignana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th Italian Conference on Computational Linguistics, CLiC-it</title>
				<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="volume">2253</biblScope>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Spanish pretrained bert model and evaluation data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cañete</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chaperon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fuentes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Ho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PML4DC at ICLR</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Hate speech review in the context of online social networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Chetty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Alathur</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Aggression and violent behavior</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="108" to="118" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">An annotated corpus for sexism detection in french tweets</title>
		<author>
			<persName><forename type="first">P</forename><surname>Chiril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Moriceau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Benamara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Origgi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Coulomb-Gully</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of The 12th Language Resources and Evaluation Conference</title>
				<meeting>The 12th Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1397" to="1403" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Tweet sentiment analysis with classifier ensembles</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">F</forename><surname>Da Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">R</forename><surname>Hruschka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">R</forename><surname>Hruschka</surname><genName>Jr</genName></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Decision Support Systems</title>
		<imprint>
			<biblScope unit="volume">66</biblScope>
			<biblScope unit="page" from="170" to="179" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Ensemble forecast modeling for the design of COVID-19 vaccine efficacy trials</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">E</forename><surname>Dean</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Piontti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">J</forename><surname>Madewell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A T</forename><surname>Cummings</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D T</forename><surname>Hitchings</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kahn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vespignani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Halloran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">M</forename><surname>Longini</surname><genName>Jr</genName></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Vaccine</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="issue">46</biblScope>
			<biblScope unit="page" from="7213" to="7216" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Large scale crowdsourcing and characterization of twitter abusive behavior</title>
		<author>
			<persName><forename type="first">A</forename><surname>Founta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Djouvas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chatzakou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Leontiadis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Blackburn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stringhini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vakali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sirivianos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kourtellis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International AAAI Conference on Web and Social Media</title>
				<meeting>the International AAAI Conference on Web and Social Media</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">12</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Online hate speech against women: Automatic identification of misogyny and sexism on twitter</title>
		<author>
			<persName><forename type="first">S</forename><surname>Frenda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ghanem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes-Y Gómez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Intelligent &amp; Fuzzy Systems</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="4743" to="4752" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Vader: A parsimonious rule-based model for sentiment analysis of social media text</title>
		<author>
			<persName><forename type="first">C</forename><surname>Hutto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gilbert</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International AAAI Conference on Web and Social Media</title>
				<meeting>the International AAAI Conference on Web and Social Media</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">8</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Racist and sexist hate speech detection: Literature review</title>
		<author>
			<persName><forename type="first">O</forename><surname>Istaiteh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Al-Omoush</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tedmori</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2020 International Conference on Intelligent Data Science Technologies and Applications (IDSTA)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="95" to="99" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Study on the impact of activity-based flexible office characteristics on the employees&apos; innovative behavioral intention</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Jang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">G</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">I</forename><surname>Hur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">J</forename><surname>An</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="87" to="103" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Predicting FTSE 100 returns and volatility using sentiment analysis</title>
		<author>
			<persName><forename type="first">M</forename><surname>Johnman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">J</forename><surname>Vanstone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gepp</surname></persName>
		</author>
		<idno type="DOI">10.1111/acfi.12373</idno>
		<ptr target="https://doi.org/10.1111/acfi.12373" />
	</analytic>
	<monogr>
		<title level="j">Accounting Finance</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="253" to="274" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Hurtbert: Incorporating lexical features with bert for the detection of abusive language</title>
		<author>
			<persName><forename type="first">A</forename><surname>Koufakou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">W</forename><surname>Pamungkas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fourth Workshop on Online Abuse and Harms</title>
				<meeting>the Fourth Workshop on Online Abuse and Harms</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="34" to="43" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">The impact of social media on social movements: The new opportunity and mobilizing structure</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Lopes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Political Science Research</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="1" to="23" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A comparison of machine learning approaches for detecting misogynistic speech in urban dictionary</title>
		<author>
			<persName><forename type="first">T</forename><surname>Lynn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">T</forename><surname>Endo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">L</forename><surname>Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ging</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2019 International Conference on Cyber Situational Awareness, Data Analytics And Assessment (Cyber SA)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Proceedings of the iberian languages evaluation forum (iberlef</title>
		<author>
			<persName><forename type="first">M</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Aragón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Agerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Álvarez Carmona</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mellado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Freitas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">G</forename><surname>Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutiérrez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M J</forename><surname>Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Plaza-De Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Taulé</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting>the Iberian Languages Evaluation Forum<address><addrLine>IberLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A bert-based transfer learning approach for hate speech detection in online social media</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mozafari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Farahbakhsh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Crespi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Complex Networks and Their Applications</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="928" to="940" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Hate speech detection and racial bias mitigation in social media based on bert model</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mozafari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Farahbakhsh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Crespi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PloS one</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page">e0237861</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Text mining for market prediction : a systematic review</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Nassirtoussi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Aghabozorgi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">Y</forename><surname>Wah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">C L</forename><surname>Ngo</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2014.06.009</idno>
		<ptr target="https://doi.org/10.1016/j.eswa.2014.06.009" />
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">16</biblScope>
			<biblScope unit="page" from="7653" to="7670" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Misogyny detection in twitter: a multilingual and cross-domain study</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">W</forename><surname>Pamungkas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">57</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">102360</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Automatic identification of misogyny in english and italian tweets at evalita 2018 with a multilingual hate lexicon</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">W</forename><surname>Pamungkas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Cignarella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Sixth Evaluation Campaign of Natural Language Processing and Speech Tools for Italian</title>
				<meeting><address><addrLine>EVALITA</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="volume">2263</biblScope>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Convai at semeval-2019 task 6: Offensive language identification and categorization with perspective and bert</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pavlopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Thain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dixon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Androutsopoulos</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th international Workshop on Semantic Evaluation</title>
				<meeting>the 13th international Workshop on Semantic Evaluation</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="571" to="576" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Resources and benchmark corpora for hate speech detection: a systematic review</title>
		<author>
			<persName><forename type="first">F</forename><surname>Poletto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sanguinetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bosco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Language Resources and Evaluation</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="47" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Automatic classification of sexism in social networks: An empirical study on twitter data</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rodríguez-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="219563" to="219576" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A primer in bertology: What we know about how bert works</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kovaleva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rumshisky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="842" to="866" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Learning and understanding different categories of sexism using convolutional neural network&apos;s filters</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sharifirad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jacovi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">B I</forename><surname>Univesity</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Matwin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings the 2019 Workshop on Widening NLP</title>
				<meeting>the 2019 Workshop on Widening NLP</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="21" to="23" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Big data : deep Learning for financial sentiment analysis</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sohangir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pomeranets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Khoshgoftaar</surname></persName>
		</author>
		<idno type="DOI">10.1186/s40537-017-0111-6</idno>
		<ptr target="https://doi.org/10.1186/s40537-017-0111-6" />
	</analytic>
	<monogr>
		<title level="j">Journal of Big Data</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1" to="25" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Mc-bert4hate: Hate speech detection using multi-channel bert for different languages and translations</title>
		<author>
			<persName><forename type="first">H</forename><surname>Sohn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2019 International Conference on Data Mining Workshops (ICDMW)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="551" to="559" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Impact of unreliable content on social media users during covid-19 and stance detection system</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Wani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bours</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">5</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Transformers: State-of-the-art natural language processing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Delangue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cistac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Funtowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Davison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shleifer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</title>
				<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="38" to="45" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dredze</surname></persName>
		</author>
		<author>
			<persName><surname>Beto</surname></persName>
		</author>
		<author>
			<persName><surname>Bentz</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1904.09077</idno>
		<title level="m">becas: The surprising cross-lingual effectiveness of bert</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b34">
	<monogr>
		<title level="m" type="main">Towards generalisable hate speech detection: a review on obstacles and solutions</title>
		<author>
			<persName><forename type="first">W</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2102.08886</idno>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
