<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Explainable and Debiased Misogyny Identification In Code-Mixed Hinglish using Artificial Intelligence Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Sargam</forename><surname>Yadav</surname></persName>
							<email>sargam.yadav@dkit.ie</email>
							<affiliation key="aff0">
								<orgName type="institution">Dundalk Institute of Technology</orgName>
								<address>
									<postCode>A91 K584</postCode>
									<settlement>Dundalk</settlement>
									<country key="IE">Ireland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Explainable and Debiased Misogyny Identification In Code-Mixed Hinglish using Artificial Intelligence Models</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">7427C21EDB21FBB8E04878CE82397690</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural language processing</term>
					<term>Code-mixed languages</term>
					<term>Hinglish</term>
					<term>Hate speech detection</term>
					<term>Explainable artificial intelligence</term>
					<term>Misogyny detection</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Hate speech in online platforms has become a significant problem. Women generally face more online harassment, and it is crucial to investigate automated approaches to filter large volumes of data to detect hate speech. In recent years, artificial intelligence models have shown great potential in automatic filtering of hate comments. In this project, a novel dataset for misogyny detection in code-mixed Hinglish will be created from YouTube and Twitter (X) and labelled for coarse-grained and fine-grained misogyny detection. Explainable Artificial Intelligence (XAI) tools will be applied to the resulting models to ensure that they are using logical features for classification. The resulting models will also be evaluated on fairness metrics, and efforts will be made to measure and mitigate unintended biases from all steps of the training process. To date, the literature review for the project has been completed and the data collection and annotation process has been started.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the era of digital communication, social media platforms have become the center of social interaction <ref type="bibr">[1,</ref><ref type="bibr">2,</ref><ref type="bibr">3,</ref><ref type="bibr">4]</ref>, business, and communication. However, this ease of communication along with the perceived anonymity of the internet has allowed hate, misinformation, and cyberbullying to spread even more rapidly than in the real world. Hate speech is any derogatory speech towards a group of people based on identity factors such as race, gender, nationality, etc. Misogyny is defined as an ingrained prejudice against women. Women are disproportionately targeted on social media platforms, and this harassment has unfortunately led to several realworld consequences <ref type="bibr" target="#b8">[5]</ref>. In recent years, machine learning and deep learning models have shown great potential in automatically filtering toxic and hateful content <ref type="bibr" target="#b9">[6]</ref>  <ref type="bibr" target="#b10">[7]</ref>. Studies have delved into using simple machine learning models such as Support Vector Machine (SVM) <ref type="bibr" target="#b11">[8]</ref> and Logistic Regression (LR) <ref type="bibr" target="#b10">[7]</ref> as well as more complex deep learning models such as Bidirectional Encoder Representations from Transformers (BERT) <ref type="bibr" target="#b12">[9]</ref>, Recurrent Neural Network (RNN), and Long short-term memory (LSTMs) to automatically detect hate speech in languages such as English <ref type="bibr" target="#b13">[10]</ref>, Italian <ref type="bibr" target="#b10">[7]</ref>, and more. However, the research effort is scarce for under-resourced and code-mixed languages. Code-mixed languages combine the features of two languages and are a product of multilingual societies. India, with over 467 million active users <ref type="bibr" target="#b14">[11]</ref>, is home to code-mixed languages such as Hinglish. Hinglish is a portmanteau of Hindi and English, and combines Romanized Hindi with English words.</p><p>The decision making process of deep learning models cannot be analysed directly due to its black-box nature. To make models interpretable and trustworthy, XAI techniques can be implemented at the global and local level. Feature attribution methods such as Local Interpretable Model-Agnostic Explanations (LIME) <ref type="bibr" target="#b15">[12]</ref> and Shapely Additive explanations (SHAP) <ref type="bibr" target="#b16">[13]</ref> can be used to understand which parts of the input affect the final prediction. Artificial intelligence models are trained on real-world datasets, which can unfortunately reflect societal biases <ref type="bibr">[14] [15]</ref>. The model can then amplify this bias and further propagate harmful stereotypes. In this project, a novel dataset of Hinglish comments collected from YouTube and Twitter will be annotated for misogyny classification at two levels: binary classification into misogyny and not-misogyny, and multi-class classification of the misogynistic comments into additional labels. Feature vectors will be obtained from the dataset using techniques such as TF-IDF, Word2Vec, etc. Additional features such as length of comments, URLs, etc., will also be considered. Machine learning and deep learning models will be trained on these features and evaluated using standard metrics such as accuracy, F1-score, precision, recall, etc. The best models will be then evaluated using XAI techniques to ensure that the model behaviour aligns with human rationale. Additionally, unintended bias will be measured from the word embeddings and the resulting model, and mitigation strategies will be applied accordingly. A toolkit will be developed from the resulting debiased and explainable model. The rest of the article is structured as follows: Section 2 describes the motivation behind conducting the project as well as the research hypothesis, questions, and objectives. Section 3 covers the relevant literature in the areas of hate speech detection, code-mixed and under-resourced languages, explainability in hate speech classification, and fairness and bias in hate speech classifiers. Section 4 details the methodology used in the project. Section 5 outlines the progress of the project to date. Section 6 discusses the plan for progression of the project, and Section 7 concludes the study.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Motivation</head><p>Social media platforms have become the epicenter of communication, business, and political discourse. Individuals from different demographic groups can connect with like-minded individuals and share their opinions. However, reports have demonstrated that women politicians and journalists face a significant amount of online gendered trolling and harassment, which can cause them to withdraw from online discussions, effectively silencing their voices <ref type="bibr" target="#b8">[5]</ref>. The continued and targeted harassment facilitated by social media can also adversely impact the mental health of victims. Unfortunately, a number of cases of online harassment have resulted in real-life harassment and murders <ref type="bibr" target="#b19">[16]</ref>. Therefore, a proactive approach to automatically filtering misogyny in online comments is essential to ensure a safe and inclusive digital space. Previous studies and shared tasks have demonstrated the potential of Natural Language Processing (NLP) techniques in monitoring online content. Therefore, the project will attempt to build on these findings and provide a novel dataset and a debiased and explainable toolkit to automatically filter misogyny in code-mixed Hinglish comments.</p><p>The research objectives, hypothesis, and questions are highlighted in the next sections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Research Objectives</head><p>The objectives of the project are listed as follows:</p><p>1. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Hypothesis and Research Questions</head><p>After a comprehensive literature review and preliminary studies, the following research hypothesis has been formulated:</p><p>Hypothesis: Artificial intelligence models can be used to detect misogyny in code-mixed Hinglish comments in a fair and explainable manner.</p><p>The research questions formulated to explore the hypothesis are as follows:</p><p>1 </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Related Work</head><p>Hate speech is a complex issue that needs to be tackled without infringing on free speech. In recent years, several studies have been conducted to counter hate speech from several perspectives in various languages. The Automatic Misogyny Identification (AMI) @ Evalita 2018 shared task focused on English and Italian datasets of 10000 tweets each and consisted of two subtasks: misogyny identification, and categorization of misogynistic behaviour and identification of the target of such behaviour <ref type="bibr" target="#b10">[7]</ref>. The Explainable Detection of Online Sexism task at SemEval-2023 <ref type="bibr" target="#b13">[10]</ref> focused on three subtasks: classification into sexist and non-sexist, classification of sexist posts into 4 categories: 1: threats, plans to harm and incitement, 2: derogation, 3: animosity, and 4: prejudiced discussion, and classification into 11 fine-grained vectors. The best scores were achieved by models such as DeBERTa-v3 and TwHINBERT. Hate speech detection in low-resource languages is a challenging task as social and cultural perspectives are required to understand the nuances of a language. Identification of Conversational Hate-Speech in Code-Mixed Languages ICHCL -2021 <ref type="bibr" target="#b9">[6]</ref> was a subtask at Hate Speech and Offensive Content Identification (HASOC) 2021 shared task that evaluated contextual hate classification in codemixed Hinglish. The best macro-F1 score obtained was 0.7253 by using a hard voting-based ensemble of 3 transformer-based models: IndicBERT, Multilingual-BERT, and XML-RoBERTa. Bhattacharya et al. <ref type="bibr" target="#b11">[8]</ref> presented a dataset of over 25,000 YouTube comments in Indian English, Hindi, and Indian Bangla for aggression and misogyny identification. An SVM classifier using 3-, 5-, and 2-character n-grams yielded the best F1 scores of 0.87, 0.89, and 0.93 for Hindi, Bangla, and English, respectively. Deep learning models are considered black boxes because their decision-making process cannot be analysed to ensure that they are behaving logically. XAI techniques allow for the exploration and examination of model behaviour to ensure responsible and fair use. DeepHate-Explainer <ref type="bibr" target="#b20">[17]</ref> is an explainable approach to hate speech classification in the Indian language 'Bengali'. Sensitivity analysis and layer-wise relevant propagation are used for explainability at a local level, and the faithfulness with respect to comprehensiveness and sufficiency are used at the model level. Artificial intelligence tools are modelled from real-world data which contain biases. Bolukbasi et al. <ref type="bibr" target="#b17">[14]</ref> demonstrated that Word2Vec embeddings exhibit gender bias through stereotypes, such as associating 'man' with 'computer programmer' and 'woman' with 'homemaker'. The study introduces two mitigation approaches: hard debiasing and soft debiasing. Nozza et al. <ref type="bibr" target="#b18">[15]</ref> evaluate unintended bias from neural network for misogyny classification trained on the dataset in <ref type="bibr" target="#b10">[7]</ref> using a synthetic identity test set and AUC-based metrics. Data augmentation has been performed by sampling data from an external corpus to balance the class distribution of identity terms. The literature review suggests that deep learning models perform better at the task than traditional machine learning models due to their complex architectures. Models that score higher on empirical metrics such as accuracy and F1-score, may not be the most explainable or interpretable. Additionally, models can become biased towards identity terms such as 'women', 'wife', etc. Therefore, it is important to explore explainability, and measure and mitigated unintended bias from the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Research Approach, Methods, Rationale</head><p>Figure <ref type="figure" target="#fig_0">1</ref> displays the methodology of the project. A novel dataset for misogyny identification in code-mixed Hinglish will be compiled from YouTube and Twitter (X), and annotated by two annotators at two level: binary classification into Misogyny (MGY) or Not Misogyny (NOT), and a multi-label classification of tweets from the MGY class into labels such as the following: 'Derailing', 'Sexual Harassment and Threats of Violence', 'Stereotyping and Objectification', 'Minimization and Trivialization', 'Whataboutism', 'Religion-based', 'Moral Policing', 'Shaming', and 'Victim Blaming'. Inter-annotator agreement will be calculated using the Cohen's kappa, and the final labels will be decided after deliberation. Where necessary, cleaning and pre- processing will be performed by removing punctuation, stopwords, hyperlinks, URLs, etc. Feature vectors will be extracted from the annotated and cleaned dataset using techniques such as Term Frequency -Inverse Document Frequency (TF-IDF) vectorizer, count vectorizer, hashing vectorizer, Word2Vec, GloVe, fastText, etc. Parametric (LR, Naive Bayes, Linear SVM, Perceptron, and more) and non-parametric (K-nearest neighbours, Decision Trees, Random Forests, etc.) machine learning models will be trained and compared. Deep learning models such as RNN, LSTM, Convolutional Neural Networks (CNN), BERT, Generative Pre-Trained Transformer (GPT), Robustly Optimized BERT Pre-training Approach (RoBERTa), etc., will also be implemented. The models will be evaluated on standard evaluation criteria such as F1 score, accuracy, precision, recall, and more. After the training and evaluation step, the best models will be examined using XAI techniques such as LIME and SHAP. To compare all the XAI techniques, a custom evaluation framework will be developed from empirical and human-readable metrics. To measure bias in the feature vectorization step, metrics such as the Word Embedding Association Test (WEAT) will be used to calculate biases in the word embeddings. After the training is completed, the best models will be evaluated for gender bias on synthetic identity test sets. Bias mitigation strategies such as hard debiasing, soft debiasing, data augmentation, etc., will be utilised, and the performance of the debiased models will be compared. After obtaining the final debiased explainable models, a toolkit will be developed to detect misogyny from comments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Preliminary Results and Contributions to date</head><p>Since the commencement of the project, considerable progress has been made. The specifics are as follows:</p><p>1. Literature Review: The preliminary literature review for the project has been completed.</p><p>The </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Next Steps</head><p>The project has been divided into the following 4 phases:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion and Future Work</head><p>Hate speech detection is a highly challenging and nuanced task due to its contextual nature. Previous approaches have achieved a reasonable level of success in differentiating between hate and non-hate content, but still struggle in differentiating between types of hateful behaviour. The project will provide a novel dataset for misogyny detection in code-mixed languages, and evaluate the performance of various machine learning and deep learning models. The models will also be evaluated using XAI techniques such as LIME and SHAP. Additionally, unintended bias will be measured from various stages of the training process and debiasing methods will be implemented. The preliminary literature review for the project has been completed. So far, approximately 2600 comments have been labeled and EDA has been performed on them. Moving forward, the data collection and annotation process will be completed, and model training and evaluation will be performed.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Flowchart of Methodology</figDesc><graphic coords="5,235.13,84.19,125.01,230.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Review current state-of-the-art in hate speech and misogyny detection, explainable hate speech classifiers, and fairness and bias in NLP to identify knowledge gaps and compare approaches. 2. Implement and compare different statistical and contextual feature embedding techniques. 3. Investigate the explainability of the classifiers through XAI tools and analyse performance on explainability metrics. 4. Analyse model fairness by measuring performance disparity for different demographics.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>paper 'Comprehensive Analysis of the Artificial Intelligence Approaches for Detecting Misogynistic Mixed-Code Online Content in South Asian Countries: A Review'<ref type="bibr" target="#b21">[18]</ref> was published in the Cyberfeminism and Gender Violence in Social Media journal. An additional survey paper titled 'Exploring Hate Speech Classification in Low-Resource Languages: A Comprehensive Review' is currently under review. 2. Dataset Collection and Annotation: Ethical permission for the data collection and annotation process has been obtained from the host institute. The data collection and annotation process has been started. Approximately 2600 YouTube comments have been collected and annotated by me and my primary supervisor for misogyny identification and categorization of misogynistic behaviour. The paper titled 'Exploratory Data Analysis on Code-mixed Misogynistic Comments. ' [19] performs exploratory analysis on the current dataset using techniques such as word cloud visualisations, sentiment polarity, PCA, etc. The Hinglish comments were clustered together. Comments belonging to the class 'MGY' are longer in length, as concluded by previous studies [15]. 3. Model Training and Evaluation: The paper 'Hate Speech is not Free Speech: Explainable Machine Learning for Hate Speech Detection in Code-Mixed Languages. ' [20] was published and presented at IEEE-ISTAS. The paper compared the performance of machine learning models and 3 feature vectorization techniques: TF-IDF, count vectorizer, and Word2Vec, on the combination of datasets from HASOC-ICHCL 2020 and 2021. LIME was also applied to the best performing model to interpret its results. The paper 'Leveraging Weakly Annotated Data for Hate Speech Detection in Code-Mixed Hinglish: A Feasibility-Driven Transfer Learning Approach with Large Language Models. ' explores zero-shot, one-shot, and few-shot learning techniques on a subset of the novel dataset (100 comments). 4. Explainability: A brief exploration of LIME, SHAP, and attention scores was performed in the 'Understanding Interpretability: Explainable AI Approaches for Hate Speech Classifiers.' paper which was accepted and presented at the XAI World Conference 2023. Models were trained on the combined HASOC-ICHCL dataset, and the XAI techniques were applied and analysed. 5. Fairness and Bias: A survey paper titled 'Unveiling the Layers: A Comprehensive Review of Fairness and Bias in Natural Language Processing' has been submitted to a relevant journal and is currently under review. The paper reviews the various sources of biases in NLP systems, fairness criteria and metrics, and bias measurement and mitigation strategies.</figDesc><table /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work has been supported by the Regulated Software Research Center at Dundalk Institute of Technology.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m">It consists of the following steps: data scraping (tweets and YouTube comments), data cleaning, and manual annotation</title>
				<imprint>
			<date type="published" when="2024-07">April -July 2024</date>
		</imprint>
	</monogr>
	<note>Data Gathering and Foundational Research. Literature review will be performed regularly to keep track of the current state-of-the-art</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m">It consists of the following tasks: pre-processing and feature extraction, model training, and model evaluation and comparative analysis</title>
				<imprint>
			<date type="published" when="2024-07">August -December 2024. July 2024</date>
		</imprint>
	</monogr>
	<note>Model Implementation and Evaluation. A large dataset consisting of comments from YouTube and Twitter (X) will be compiled and annotated by the end of. A DOI will be obtained for the dataset, and it will be made publicly available in accordance with the host institute&apos;s open access policy. 1 conference paper will be targeted for EACL 2025</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">It consists of the following tasks: investigate XAI techniques, evaluate model fairness, evaluate and mitigate unintended biases, analyse results</title>
	</analytic>
	<monogr>
		<title level="m">Model Explainability</title>
				<imprint>
			<date type="published" when="2025-10">January -October 2025. 2025</date>
		</imprint>
	</monogr>
	<note>XAI-World Conference</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Writeup, and Completion</title>
	</analytic>
	<monogr>
		<title level="m">It consists of the following steps: finalize toolkit</title>
				<imprint>
			<date type="published" when="2025-11">November 2025 -March 2026. 2026</date>
		</imprint>
	</monogr>
	<note>PhD writeup, and PhD examination. 1 conference paper will be targeted for SIGIR</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Cooking is creating emotion: A study on hinglish sentiments of youtube cookery channels using semi-supervised approach</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kaur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Big Data and Cognitive Computing</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page">37</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Sentiment analysis on google play store data using deep learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Venkatakrishnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">K</forename><surname>Verma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Applications of Machine Learning</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="15" to="30" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Opinion-mining on marglish and devanagari comments of youtube cookery channels using parametric and non-parametric learning models</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R</forename><surname>Shah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shah</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Big Data and Cognitive Computing</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page">3</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Disruptive technologies in smart farming: an expanded view with sentiment analysis</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AgriEngineering</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="424" to="460" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">A</forename></persName>
		</author>
		<ptr target="https://www.amnesty.org/en/latest/research/2018/03/online-violence-against-women-chapter-1-1/" />
		<title level="m">Toxic Twitter -A Toxic Place for Women</title>
				<imprint>
			<date type="published" when="2024-04-18">2024. 2024-04-18</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<author>
			<persName><forename type="first">S</forename><surname>Satapara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Modha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mandl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Madhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Majumder</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Overview of the hasoc subtrack at fire 2021: Conversational hate speech detection in code-mixed language</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Overview of the evalita 2018 task on automatic misogyny identification (ami), EVALITA Evaluation of NLP and Speech Tools for Italian</title>
		<author>
			<persName><forename type="first">E</forename><surname>Fersini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nozza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">59</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Bhattacharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bansal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bhagat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Dawer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lahiri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2003.07428</idno>
		<title level="m">Developing a multilingual annotated corpus of misogyny and aggression</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">R</forename><surname>Kirk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Vidgen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Röttger</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2303.04222</idno>
		<title level="m">SemEval-2023 Task 10: Explainable Detection of Online Sexism</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">YouTube Users by Country</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">P</forename><surname>Review</surname></persName>
		</author>
		<ptr target="https://worldpopulationreview.com/country-rankings/youtube-users-by-country" />
		<imprint>
			<date type="published" when="2024-04-18">2024. 2024. 2024-04-18</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Why should i trust you?&quot; Explaining the predictions of any classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</title>
				<meeting>the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1135" to="1144" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Man is to computer programmer as woman is to homemaker? debiasing word embeddings</title>
		<author>
			<persName><forename type="first">T</forename><surname>Bolukbasi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Y</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Saligrama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Kalai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Unintended bias in misogyny detection</title>
		<author>
			<persName><forename type="first">D</forename><surname>Nozza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Volpetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Fersini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Ieee/wic/acm international conference on web intelligence</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="149" to="155" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Right-Wing Extremist Convicted of Murdering Jo Cox</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chan</surname></persName>
		</author>
		<ptr target="https://www.nytimes.com/2016/11/23/world/europe/thomas-mair-convicted-murder-jo-cox.html" />
		<editor>U.K</editor>
		<imprint>
			<date type="published" when="2016">2016. 2023-02-01</date>
			<pubPlace>New York Times</pubPlace>
		</imprint>
	</monogr>
	<note>. Lawmaker</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Deephateexplainer: Explainable hate speech detection in under-resourced bengali language</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Karim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Islam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sarker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">H</forename><surname>Menon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hossain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Hossain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Decker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 8th International Conference on Data Science and Advanced Analytics (DSAA), IEEE</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="1" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Comprehensive Analysis of the Artificial Intelligence Approaches for Detecting Misogynistic Mixed-Code Online Content in South Asian Countries: A Review</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Cyberfeminism and Gender Violence in Social Media</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="350" to="368" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mcdaid</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2403.09709</idno>
		<title level="m">Exploratory Data Analysis on Code-mixed Misogynistic Comments</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Hate Speech is not Free Speech: Explainable Machine Learning for Hate Speech Detection in Code-Mixed Languages</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yadav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaushik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Mcdaid</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Symposium on Technology and Society (ISTAS), IEEE</title>
				<imprint>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
