<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Progress in the Modeling of Violent Messages in Spanish Social Networks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Beatriz</forename><surname>Botella-Gil</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dept. of Software and Computing Systems</orgName>
								<orgName type="institution">University of Alicante</orgName>
								<address>
									<addrLine>Apdo. de Correos 99</addrLine>
									<postCode>E-03080</postCode>
									<settlement>Alicante</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Progress in the Modeling of Violent Messages in Spanish Social Networks</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2849DD822903031ACFD68864B1D0663F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:05+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Processing</term>
					<term>Annotation Guideline</term>
					<term>Dataset Annotation</term>
					<term>Detection of Violent Messages</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Society advances loaded with new and highly accessible knowledge, that is published in the virtual world. It is a reality that ICTs have brought many benefits for our lives but we also see how year after year the use of violence in platforms increase. The application of PLN is fundamental in this type of research given the large volume of existing data, which facilitates a breakthrough in the investigation of the detection of this type of messageThe doctoral work focuses on the detection of violent messages in the social network twitter from different perspectives.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The Internet has become an indispensable part of our lives, being used in practically all of society's daily activities. Nowadays it is possible to have immediate contact with any person in the world through an electronic device. Society is moving forward with new and very accessible knowledge that is published in the virtual world. Personal relationships have also been affected, not only in the private sphere, but also in the workplace.</p><p>According to We are Social <ref type="bibr" target="#b0">[1]</ref>, almost 44 million people in Spain spend more than 6 hours a day on the Internet and around 41 million Spaniards are users of social networks. It is a reality that ICTs have brought many benefits to our lives, but also, thanks to the possibility of being an anonymous user and the absence of observing face to face the damage that our words can generate, problems still to be solved are created <ref type="bibr" target="#b1">[2]</ref>. In particular, many researchers call this type of violent action as hate speech, an offensive behavior through language towards people or groups and whose detection is being a problem for researchers, since, it is possible that violence is not used explicitly in a discourse, but as a single word or even implicitly through the use of emoticons <ref type="bibr" target="#b2">[3]</ref>, or by using humor, irony, sarcasm <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref> or stereotypes <ref type="bibr" target="#b5">[6]</ref>.</p><p>Given the number of users present in social networks, it is impossible to manually control the comments that are registered and their intention, creating impunity for people who use these networks in order to do harm. The identification of violent messages and control of hate speech on the Internet has been approached from different points of view, being essential the use of Natural Language Processing (NLP) to develop computational systems that help to interpret and process human language quickly and effectively. Doctoral Symposium on Natural Language Processing from the Proyecto ILENIA, 28 September 2023, Jaén, Spain.</p><p>beatriz.botella@dlsi.ua.es (B. Botella-Gil)</p><p>One barrier we encountered right at the start of the study is the collection of messages in social networks, since, as Bruns points out <ref type="bibr" target="#b6">[7]</ref>, the restriction of access to social network data makes it difficult to analyze issues of great importance such as abusive language, harassment, hate speech or disinformation campaigns. That is why in the present research the social network Twitter will be used, where as Ott <ref type="bibr" target="#b7">[8]</ref> defines: "Twitter discourse is disrespectful because its register is informal, and because it depersonalizes social interactions". This research aims to provide solutions to the existing problems in the detection of violent messages in social networks in a fast, automatic and effective way.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background and Related Work</head><p>Many studies have been carried out on the analysis of violent messages in social networks and media. In particular, much research can be found focused on discovering the characteristics of human behavior that promote the emission of such messages, as well as those that focus on discovering the characteristics of the messages themselves through PLN techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Language and behavior study</head><p>There is a wealth of research on human behavior in the face of violent messages and the language used. As McMenamin said <ref type="bibr" target="#b8">[9]</ref>, "hate speech is studied according to how it is defined, how it is interpreted, and what are the best practices to deal with it". That is why we find works such as Salado <ref type="bibr" target="#b9">[10]</ref>, who based their research on a syntactic analysis of language, and discovered that there are different linguistic elements to take into account that are present in the forms of violent speech such as, the linguistic category, the lexicon used or how the words are placed. Plaza-Del-Arco et al. <ref type="bibr" target="#b10">[11]</ref> carried out a study of the implicit and explicit linguistic phenomena of offensive language. Others such as Gitari <ref type="bibr" target="#b11">[12]</ref> focused on something as specific as the creation of a list of verbs that can be indicators of violent messages. On the other hand, there are works that focus on the roles present in these acts, such as, for example, Nielsen, who, through interviews and a study of the participants, observed the consequences for the victim, his or her harm and the possibility of crime in the messages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">PLN applied to the detection of violent messages</head><p>The application of PLN is fundamental in this type of research given the large volume of existing data, which facilitates a breakthrough in the investigation of the detection of this type of messages, thanks to the following techniques:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• Keyword-based classifiers</head><p>Part of the research in this field has focused on the development of lists of insults that help automatic detection. In this sense, lexicons and dictionaries have been developed in order to observe whether the presence of these terms determines the violence in the message <ref type="bibr" target="#b12">[13]</ref>. Although such lists have aided detection, they have fallen short of being the sole tool for determining violence. Violent language is constantly evolving, language varies from place to place varies depending on where it occurs and there may be terms that in some geographic areas are insults and in others are not <ref type="bibr" target="#b13">[14]</ref>.</p><p>• Machine learning Most of the work related to the detection of violent messages addresses this problem using classical machine learning (ML) algorithms. Works such as Xu et al. <ref type="bibr" target="#b14">[15]</ref> and Dadvar et al. <ref type="bibr" target="#b15">[16]</ref> have used support vector machines (SVM) in their research and obtained satisfactory results, proving to be very effective with large training samples. SVM is not the only classical algorithm used in research in this field; works such as <ref type="bibr" target="#b16">[17]</ref>, used other algorithms, showing in their results that the one that offered the best performance is logistic regression, followed by Naive Bayes and SVMs.</p><p>Most ML-based classifiers use traditional text representations such as bag-of-words (BOW), n-grams, term frequency (TF), among others. In Burnap and Williams <ref type="bibr" target="#b17">[18]</ref> all of the above techniques are used. This research compares the results obtained individually by the classifiers with the use of a set of classifiers (ensemble) that integrates them all, demonstrating greater accuracy in the latter. Sentiment analysis is another of the most widely used tools in this field. With it we can extract the polarity of the message and use this indicator along with other tasks to determine more accurately whether we are facing a violent or non-violent message <ref type="bibr" target="#b18">[19]</ref>.</p><p>Corpus development, have an important role in offensive language research when ML techniques are applied. In recent years we have observed a large volume of work by PLN researchers to generate these resources <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b23">24,</ref><ref type="bibr" target="#b24">25]</ref>. These authors created English-language resources, with SOLID <ref type="bibr" target="#b24">[25]</ref> being the resource containing more than nine million English-language tweets tagged in a semi-supervised manner.</p><p>On the other hand, HurtLex <ref type="bibr" target="#b25">[26]</ref> is a multilingual lexicon of hate speech spanning multiple languages and hatebase3<ref type="foot" target="#foot_0">1</ref> is a collaborative repository of hate speech that is also multilingual. The main drawback of these resources is their paucity of English terms, and those that are present have been compiled using a semi-automatic translation from another language, neglecting the importance of cultural and linguistic factors in each country. However, despite the fact that Spanish is one of the most widely spoken languages in the world, we found a shortage of resources in this language to carry out the task of detection. There are resources in Spanish for offensive words such as Plaza-Del-Arco et al. <ref type="bibr" target="#b10">[11]</ref> for misogynistic and xenophobic terms; and Share <ref type="bibr" target="#b26">[27]</ref> that label them as offensive and not. After the study carried out on the literature, it is considered necessary to elaborate another corpus to collect more characteristics present in violent messages, which can help in the explicability and detail of the detection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>• Deep learning</head><p>Within AI there are other more complex techniques that have also been used in this task. We refer to deep learning (DL), as is the case of the research by Arcila-Calderón et al.</p><p>[17] that after using ML tools and neural networks, the latter improved the evaluation metrics against the models generated with traditional ML algorithms. To the same end Badjatiya et al. <ref type="bibr" target="#b27">[28]</ref> uses DL models to train different word embeddings validating that, using these representations, obtains better results than traditional representations such as term frequency -inverse document frequency (TF-IDF) or BoW. Models based on transformer architecture, such as BERT, RoBERTa and ALBERT, show the best state-of-the-art results in the detection of violent messages in tasks recognized as OffensEval or HatEval <ref type="bibr" target="#b28">[29]</ref>. In Sarkar et al. <ref type="bibr" target="#b28">[29]</ref> fine-tuning (fine-tuning) to BERT is performed using SOLID, the largest English offensive language identification corpus, improving the results obtained with BERT in the tasks mentioned above. In Song et al. <ref type="bibr" target="#b29">[30]</ref> an ensemble of classifiers (ensemble) based on RoBERTa and BERT is used which obtains the best results in the shared task "SemEval-2021 Task 7: HaHackathon, Detecting and Rating Humor and Offense"<ref type="foot" target="#foot_1">2</ref> which includes a subtask of detecting offensive messages. This work consists of fine tuning these models to create a classifier and clustering them into a set of classifiers based on stacking.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Main Hypothesis and Objectives</head><p>After literature reviews, future work was found to improve the detection of violent messages in social networks, more specifically focused not only on the language itself, but also on determining the existence of patterns of behavior and user profiles. This problem is being studied taking as references data such as user role, type of violence, message form (implicit/explicit), isolated message vs. thread of messages, number of followers and activity in the social network (number of # or mentions of the user).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Objectives</head><p>In our research the following objectives will be pursued:</p><p>1. Expanding the corpus VIL <ref type="bibr" target="#b30">[31]</ref>.</p><p>2. Define behavioral patterns for violent identification.</p><p>3. Achieve a system that is able to give a user violence score, both to know if the user is violent and to know if is receiving violence. 4. Identify the phases of the violence process (beginning, in process, completed). 5. Creation of a program for automatic detection of violence in social networks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Hypothesis</head><p>For the study of violence detection through messages on social networks with the help of NLP, the following hypotheses are pursued:</p><p>1. Is it possible to detect violence on social networks through language analysis?. Certain words and phrases could serve as indicators that increase the likelihood of a message being considered violent. Therefore, it is relevant to investigate how language can be used offensively in online communication, using linguistic cues as potential markers of violence in messages. 2. Do patterns exist that aid in the detection of violence on social networks?. The study of the characteristics surrounding the issue of violence on social networks, including user behavior and their use of language, could provide essential insights that bring us closer to identifying effective techniques to address this problem.</p><p>3. Is it possible to detect violent messages using NLP tools?. We believe that ML models trained with labeled data significantly enhance the ability to detect violent messages compared to rule-based approaches alone.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head><p>We have conducted 3 experiments during this thesis, a FIERO chatbot, an annotation guide and a Corpus of violent messages through the social network Twitter.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Fiero</head><p>Fiero, a virtual assistant that maintains a conversation with the user encouraging him to express expletives through Telegram. This application is popularly known in the environment of the use of digital tools by the population. The collected dialogue will be used to generate linguistic resources that can be used in automatic artificial intelligence systems to combat social problems such as cyberbullying or hate speech <ref type="bibr" target="#b31">[32]</ref>. In Figure <ref type="figure" target="#fig_0">1</ref>, you can observe what a conversation with the chatbot looks like to collect insults from users. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Annotation Guide</head><p>Pursuing the goal of creating a resource to aid in the detection of violent messages, we decided to generate a fine-grained annotation guide for messages, with a certain degree of semantic complexity, which not only marks whether a message is violent or not, but also certain important elements regarding the content of the message. In this annotation guide shown in figure <ref type="figure" target="#fig_1">2</ref> we collect information about: Insults, Violent vs No-Violent, Level of violence, Role and Type of Violence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">VIL Corpus</head><p>Having studied the literature on the task and the importance of the application of PLN and ML and DL techniques, and because these techniques are fed by training data, we conclude the need to create a resource in Spanish that can be used in the effective detection of violent messages, with a level of detail that goes beyond simple binary detection, marking features, which we detail in our annotation guide, such as the degree, the role or type of violence, since we consider that if detection in the messages that can help future explainability in the decisions taken <ref type="bibr" target="#b30">[31]</ref>.</p><p>In the figure 3 you can see an example of how the labeled tweets would look in our corpus. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Research issues to discuss</head><p>Following our research and the creation of an annotation guide and corpus presented in the previous sections, we have questions that could be addressed in the future:</p><p>1. Is possible to detect violent messages through Machine Learning to curb the problem of the use of violence in social networks? Due to the ambiguity and personal subjectivity in how users understand violence, we may encounter difficulties in reaching an agreement on what constitutes a violent message and what does not.</p><p>2. Different phases of violence can be defined in order to know what level of severity we find or in what phase of violence we are in? We could determine the severity level of violence based on the message content to take appropriate actions with the violent user, imposing different consequences based on severity. 3. An automatic alert could be created to warn us that violence is being generated? A prompt response in a case of violence would be beneficial for the virtual community. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Acknowledgments</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Fiero user interface.</figDesc><graphic coords="5,229.72,323.98,133.33,229.39" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Annotation guide</figDesc><graphic coords="6,150.55,267.80,291.68,218.76" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Examples of tweets annotated in the VIL corpus.</figDesc><graphic coords="7,108.88,188.54,375.02,173.37" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">urlhttps://hatebase.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://bit.ly/3J8uHOX</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This research work is part of the Spanish research project Cleartext «TED2021-130707B-I00&gt;. This work is funded by MCIN/AEI/10.13039/501100011033 and by the European Union NextGen-erationEU/ PRTR. I would also like to thank my thesis directors Patricio Martinez-Barco and Estela Saquete Boro for their help and collaboration with the project Coolang (proyecto de I+D+i «PID2021-122263OB-C22», funded por MCIN/ AEI/10.13039/501100011033/ and "FEDER Una manera de hacer Europa").</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Hootsuite</forename><surname>Wearesocial</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Digital</forename><surname>Report España</surname></persName>
		</author>
		<ptr target="https://encr.pw/8avSe" />
		<imprint>
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Guía rápida para la prevención del acoso por medio de las nuevas tecnologías</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Flores</forename><surname>Fernandez</surname></persName>
		</author>
		<ptr target="https://www.pantallasamigas.net/ciberbullying-guia-rapida" />
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Alonso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">J</forename><surname>Vázquez</surname></persName>
		</author>
		<title level="m">Sobre la libertad de expresión y el discurso del odio: Textos críticos, Athenaica ediciones universitarias</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Killing me softly: Creative and cognitive aspects of implicitness in abusive language online</title>
		<author>
			<persName><forename type="first">S</forename><surname>Frenda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Natural Language Engineering</title>
		<imprint>
			<biblScope unit="page" from="1" to="22" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The unbearable hurtfulness of sarcasm</title>
		<author>
			<persName><forename type="first">S</forename><surname>Frenda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Cignarella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bosco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">193</biblScope>
			<biblScope unit="page">116398</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Masking and bert-based models for stereotype identication</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sánchez-Junquera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="83" to="94" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">After the &apos;apicalypse&apos;: Social media platforms and their fight against critical scholarly research</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bruns</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information, Communication &amp; Society</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="1544" to="1566" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The age of twitter: Donald j. trump and the politics of debasement</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">L</forename><surname>Ott</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Critical studies in media communication</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="59" to="68" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">R</forename><surname>Mcmenamin</surname></persName>
		</author>
		<title level="m">Introducción a la lingüística forense: un libro de curso</title>
				<meeting><address><addrLine>Fresno</addrLine></address></meeting>
		<imprint>
			<publisher>Press at</publisher>
			<date type="published" when="2017">2017</date>
		</imprint>
		<respStmt>
			<orgName>California State University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Análisis lingüístico del discurso de odio en redes sociales, VISUAL REVIEW</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Salado</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Visual Culture Review/Revista Internacional de Cultura Visual</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1" to="11" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Detecting misogyny and xenophobia in spanish tweets using language technologies</title>
		<author>
			<persName><forename type="first">F.-M</forename><surname>Plaza-Del-Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Molina-González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ureña-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Internet Technology (TOIT)</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="1" to="19" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A lexicon-based approach for hate speech detection</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">D</forename><surname>Gitari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zuping</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Damien</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Long</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Multimedia and Ubiquitous Engineering</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="215" to="230" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Automatic identification of personal insults on social news sites</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">O</forename><surname>Sood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">F</forename><surname>Churchill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Antin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Society for Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">63</biblScope>
			<biblScope unit="page" from="270" to="285" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Abusive language detection in online user content</title>
		<author>
			<persName><forename type="first">C</forename><surname>Nobata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tetreault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Thomas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Mehdad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th international conference on world wide web</title>
				<meeting>the 25th international conference on world wide web</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="145" to="153" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Learning from bullying traces in social media</title>
		<author>
			<persName><forename type="first">J.-M</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-S</forename><surname>Jun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bellmore</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2012 conference of the North American chapter of the association for computational linguistics: Human language technologies</title>
				<meeting>the 2012 conference of the North American chapter of the association for computational linguistics: Human language technologies</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="656" to="666" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Improving cyberbullying detection with user context</title>
		<author>
			<persName><forename type="first">M</forename><surname>Dadvar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Trieschnigg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ordelman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">D</forename><surname>Jong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="693" to="696" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Using shallow and deep learning to automatically detect hate motivated by gender and sexual orientation on twitter in spanish</title>
		<author>
			<persName><forename type="first">C</forename><surname>Arcila-Calderón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Amores</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sánchez-Holgado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Blanco-Herrero</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Multimodal technologies and interaction</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page">63</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">for policy decision making</title>
		<author>
			<persName><forename type="first">P</forename><surname>Burnap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Williams</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Hate speech, machine classification and statistical modelling of information flows on twitter: Interpretation and communication</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Hate speech classification in social media using emotional analysis</title>
		<author>
			<persName><forename type="first">R</forename><surname>Martins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gomes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Novais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Henriques</surname></persName>
		</author>
		<idno type="DOI">10.1109/BRACIS.2018.00019</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings -2018 Brazilian Conference on Intelligent Systems</title>
				<meeting>-2018 Brazilian Conference on Intelligent Systems<address><addrLine>BRACIS</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="61" to="66" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Inducing a lexicon of abusive words-a feature-based approach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ruppenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Greenberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1046" to="1056" />
		</imprint>
	</monogr>
	<note>Long Papers</note>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Elsherief</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Belding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Wang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1904.02418</idno>
		<title level="m">Learning to decipher hate symbols</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">The effect of extremist violence on hateful speech online</title>
		<author>
			<persName><forename type="first">A</forename><surname>Olteanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Castillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Boy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Varshney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the international AAAI conference on web and social media</title>
				<meeting>the international AAAI conference on web and social media</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">12</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">A survey on automatic detection of hate speech in text</title>
		<author>
			<persName><forename type="first">P</forename><surname>Fortuna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nunes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys (CSUR)</title>
		<imprint>
			<biblScope unit="volume">51</biblScope>
			<biblScope unit="page" from="1" to="30" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Resources and benchmark corpora for hate speech detection: a systematic review</title>
		<author>
			<persName><forename type="first">F</forename><surname>Poletto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sanguinetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bosco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Language Resources and Evaluation</title>
		<imprint>
			<biblScope unit="volume">55</biblScope>
			<biblScope unit="page" from="477" to="523" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Rosenthal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Atanasova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Karadzhov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zampieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2004.14454</idno>
		<title level="m">A large-scale semisupervised dataset for offensive language identification</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Hurtlex: A multilingual lexicon of words to hurt</title>
		<author>
			<persName><forename type="first">E</forename><surname>Bassignana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">5th Italian Conference on Computational Linguistics, CLiC-it 2018</title>
				<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">2253</biblScope>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Share: A lexicon of harmful expressions by spanish speakers</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Plaza-Del Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B P</forename><surname>Portillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">L</forename><surname>Úbeda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Gil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference</title>
				<meeting>the Thirteenth Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1307" to="1316" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Deep learning for hate speech detection in tweets</title>
		<author>
			<persName><forename type="first">P</forename><surname>Badjatiya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Varma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 26th international conference on World Wide Web companion</title>
				<meeting>the 26th international conference on World Wide Web companion</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="759" to="760" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Sarkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zampieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ranasinghe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ororbia</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2109.05074</idno>
		<title level="m">Fbert: A neural transformer for identifying offensive content</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Deepblueai at semeval-2021 task 7: Detecting and rating humor and offense with stacking diverse language model-based methods</title>
		<author>
			<persName><forename type="first">B</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Luo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th international workshop on semantic evaluation</title>
				<meeting>the 15th international workshop on semantic evaluation<address><addrLine>SemEval-</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="1130" to="1134" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Violencia identificada en el lenguaje (vil). creación de recurso para mensajes violentos</title>
		<author>
			<persName><forename type="first">B</forename><surname>Botella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sepúlveda-Torres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">M</forename><surname>Barco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Saquete</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">70</biblScope>
			<biblScope unit="page" from="187" to="198" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Fiero: Asistente virtual para la captación de insultos</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">B</forename><surname>Gil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M P</forename><surname>Del Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B P</forename><surname>Portillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutiérrez</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2968/paper8.pdf" />
	</analytic>
	<monogr>
		<title level="s">CEUR Workshop Proceedings</title>
		<imprint>
			<date type="published" when="2021">2968. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
