<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Nullpointer at CheckThat! 2024: Identifying Subjectivity from Multilingual Text Sequence</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Md</forename><forename type="middle">Rafiul</forename><surname>Biswas</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Hamad Bin Khalifa University</orgName>
								<address>
									<settlement>Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Abrar</forename><surname>Tasneem Abir</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Education</orgName>
								<orgName type="institution">Carnegie Mellon University in Qatar</orgName>
								<address>
									<settlement>City, Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Wajdi</forename><surname>Zaghouani</surname></persName>
							<email>wajdi.zaghouani@northwestern.edu</email>
							<affiliation key="aff2">
								<orgName type="department">Education</orgName>
								<orgName type="institution">Northwestern University in Qatar</orgName>
								<address>
									<settlement>City, Doha</settlement>
									<country key="QA">Qatar</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Nullpointer at CheckThat! 2024: Identifying Subjectivity from Multilingual Text Sequence</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">ED92825808FFB7408119DA18B6E65E0C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>subjectivity</term>
					<term>natural language processing</term>
					<term>sentiment</term>
					<term>fact checking</term>
					<term>news articles</term>
					<term>text sequence</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This study addresses a binary classification task to determine whether a text sequence, either a sentence or paragraph, is subjective or objective. The task spans five languages-Arabic, Bulgarian, English, German, and Italian-along with a multilingual category. Our approach involved several key techniques. Initially, we preprocessed the data through parts of speech (POS) tagging, identification of question marks, and application of attention masks. We fine-tuned the sentiment-based Transformer model 'MarieAngeA13/Sentiment-Analysis-BERT' on our dataset. Given the imbalance with more objective data, we implemented a custom classifier that assigned greater weight to objective data. Additionally, we translated non-English data into English to maintain consistency across the dataset. Our model achieved notable results, scoring top marks for the multilingual dataset (Macro F1-0.7121) and German (Macro F1-0.7908). It ranked second for Arabic (Macro F1-0.4908) and Bulgarian (Macro F1-0.7169), third for Italian (Macro F1-0.7430), and ninth for English (Macro F1-0.6893).</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The concepts of objectivity and subjectivity are crucial in shaping methodologies, interpretations, and the perceived validity of findings in many natural language processing (NLP) applications, such as sentiment analysis and information extraction <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. Objectivity analysis relies on data that can be measured, observed, and verified by others and is achieved through careful experimental designs, standard procedures, and statistical analysis. In an ideal sense, objective analysis is supposed to be free from individual biases, emotions, and personal judgments, thereby ensuring that the results are universally valid and replicable <ref type="bibr" target="#b2">[3]</ref>.</p><p>Subjectivity, on the other hand, refers to perspectives, interpretations, or analyses that are influenced by personal experiences, feelings, beliefs, or biases <ref type="bibr" target="#b3">[4]</ref>. Subjective analysis is inherently shaped by the individual's background, cultural context, and personal viewpoints. While often perceived as less reliable or credible in scientific contexts, subjectivity is an unavoidable aspect of human cognition and can provide valuable insights, particularly in fields such as humanities, social sciences, and qualitative research where personal interpretation and contextual understanding are essential <ref type="bibr" target="#b4">[5]</ref>.</p><p>Identifying whether a text sequence expresses personal opinions, emotions, or factual information is essential for enhancing the accuracy and relevance of automated systems in diverse fields such as social media monitoring, customer feedback analysis, and news content categorization. In data analysis, the tension that arises from the interaction of objectivity and subjectivity frequently affects decision-making procedures and the dissemination of findings. The challenge lies in creating systems that can accurately classify text sequences-whether sentences or paragraphs-as either subjective, reflecting personal opinions or sentiments, or objective, presenting factual information devoid of personal bias <ref type="bibr" target="#b5">[6]</ref>. In an effort to improve the acceptability and credibility of work, researchers may strive for objectivity, occasionally avoiding or hiding choices that might be viewed as subjective. Subjective opinions can, for example, slightly skew the analysis's ostensibly objective results in the data selection, analytical method selection, and result interpretation processes. Thus, there is a high chance that the dataset contains a relatively higher number of objective values compared to subjective values.</p><p>Task 2 in CheckThat Lab at CLEF 2024 <ref type="bibr" target="#b6">[7]</ref> classifies text as either subjective or objective. This binary classification task requires systems to accurately identify the nature of a text sequence. The task is offered in multiple languages: Arabic, Bulgarian, English, German, and Italian, providing a comprehensive multilingual evaluation of the systems' capabilities. The challenge of multilingual and cross-linguistic text classification is compounded by the inherent linguistic and cultural differences that influence the expression of subjectivity and objectivity.</p><p>This study presents an approach to a binary classification task aimed at discerning subjective from objective text across multiple languages. By leveraging advanced NLP techniques and Transformer models, we aim to enhance the accuracy and robustness of subjective-objective text classification. The implications of this research extend to improving automated news analysis, enhancing content recommendation systems, and promoting a comprehension understanding in various languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>The task of classifying text as subjective or objective has been studied extensively in natural language processing. Early work by Wiebe et al. <ref type="bibr" target="#b7">[8]</ref> laid the foundations for subjectivity analysis, proposing a scheme for annotating subjective elements in text. They developed a system called OpinionFinder <ref type="bibr" target="#b0">[1]</ref> which performed subjectivity analysis using various lexical and syntactic features. More recently, deep learning approaches have been applied to this task with great success. Nakov et al. <ref type="bibr" target="#b8">[9]</ref> provide a thorough overview of modern approaches to sentiment analysis, including detecting subjectivity. They highlight the effectiveness of leveraging pre-trained language models like BERT <ref type="bibr" target="#b9">[10]</ref> and fine-tuning them for the target task. Several studies have specifically examined subjectivity classification in a multilingual setting. Balahur et al. <ref type="bibr" target="#b10">[11]</ref> constructed a multilingual dataset for subjectivity classification in English, Spanish, French and German. They experimented with various machine translation approaches to make the problem cross-lingual. Similarly, Mihalcea et al. <ref type="bibr" target="#b11">[12]</ref> generated subjectivity datasets for English and Romanian, using English tools and manually translating the subjective sentences into Romanian. The CLEF <ref type="bibr" target="#b12">[13]</ref>(Conference and Labs of the Evaluation Forum) has run workshops on automatic identification and verification of claims in political debates, speeches, and news articles since 2018 <ref type="bibr" target="#b13">[14]</ref>. The CheckThat! shared task at CLEF focuses on detecting checkworthy claims across various languages including Arabic <ref type="bibr" target="#b14">[15]</ref>, which is one of the languages in the current study. In terms of methodology, fine-tuning pre-trained Transformer models has proven very effective for subjectivity and sentiment tasks. Xu et al. <ref type="bibr" target="#b15">[16]</ref> fine-tuned BERT for sentiment classification and demonstrated its strong performance on multiple benchmarks. Exploring multi-task learning, Yu and Jiang <ref type="bibr" target="#b16">[17]</ref> showed that jointly learning sentiment and subjectivity through a shared BERT encoder led to improvements on both tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">System Overview</head><p>This works system for subjectivity classification comprises several key components, including data preprocessing, model selection, and training strategies. This section provides an overview of each component and the techniques employed (see Figure <ref type="figure" target="#fig_0">1</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Data Preprocessing</head><p>The first step in the pipeline is data preprocessing, which involves cleaning and transforming the raw text data into a suitable format for model. The preprocessing steps include: We also experiment with additional preprocessing techniques such as part-of-speech (POS) tagging and attention masking, but find that they do not significantly improve the performance of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Model Selection</head><p>For the subjectivity classification task, we choose to fine-tune pre-trained Transformer models that have been previously trained on sentiment analysis tasks. Specifically, we use the 'MarieAngeA13/Sentiment-Analysis-BERT' model, which is a BERT-based model fine-tuned for sentiment analysis. We find that this approach of using a model already fine-tuned for a related task (i.e., multi-task learning) yields better results compared to fine-tuning a pre-trained model from scratch. The code and data can be found in the GitHub repository https://github.com/Abrar-Abir/CLEF2024task02.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Training Strategies</head><p>The training was conducted on a remote Dell server running the latest Ubuntu 22 OS with 512 GB RAM and 24-core CPU. The server was equipped with NVIDIA A100 GPU with 80 GB GPU memory. We employ several training strategies to improve the performance of the model listed below.</p><p>• Label mapping: The pre-trained sentiment analysis model is designed for three-class prediction (positive, neutral, negative), while our subjectivity classification task requires only two classes (subjective and objective). We experimented with different label mappings and found that mapping subjective to negative sentiment and objective to positive sentiment yielded the best results. • Confidence weighting: For the English dataset, we incorporate the confidence level information provided in the dataset (in the 'solved_conflict' column where 1[true] means conflict was resolved i.e., higher annotation confidence and vice versa). We assign 20% higher weight for the training losses-coming from the annotations with higher confidence (i.e, 1.2 weight)-before passing the losses to the loss function so that backpropagation prioritizes minimizing loss for higher confidence annotation compared to their counterparts. • Hyperparameter tuning: We experiment with different hyperparameter settings and find that a batch size of 16, learning rate of 2e-5, and training for 20 epochs yields the best performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Language Adaptation</head><p>To handle the multilingual nature of the task, we employ machine translation to convert non-English data into English. We use the Google Translator API through the deep translator library for this purpose.</p><p>While we also experiment with fine-tuning language-specific pre-trained models for non-English languages, we find that translating the training and test datasets to English and using the English model yields better performance. These preprocessing, model selection, and training strategies form the core of the subjectivity classification system. In the following sections, we detail our experimental setup and present the results of the approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>This section presents the results of subjectivity classification system across various languages and datasets. We first describe the dataset characteristics and then provide a detailed analysis of the model's performance using different evaluation metrics. Finally, we compare our results with those of other participating teams in the CheckThat! Lab at CLEF 2024.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Dataset Description</head><p>The dataset for the Subjectivity Subtask consists of sentences from news articles in five languages: Arabic, Bulgarian, English, German, and Italian. Additionally, there is a multilingual dataset that combines all five languages. Table <ref type="table" target="#tab_1">1</ref> shows the distribution of objective and subjective sentences in the training and test sets for each language. Across all languages, the percentage of objective sentences is higher than that of subjective sentences, with the imbalance being more pronounced in the training sets. This imbalance poses a challenge for subjectivity classification systems, as they need to learn from skewed data distributions. The test set contains 500 sentences, evenly split with 250 objective sentences (50%) and 250 subjective sentences (50%). This comprehensive dataset provides a robust foundation for developing and evaluating systems that distinguish between subjective and objective statements in news articles across multiple languages. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Performance Metrics</head><p>We evaluate our subjectivity classification model using various performance metrics, including macroaveraged F1-score, precision, recall, and accuracy. Table <ref type="table" target="#tab_2">2</ref> presents the results for each language and the multilingual dataset. Our model achieves the best performance on the German dataset, with an F1 Macro score of 0.79 and an accuracy of 0.81, indicating high prediction correctness. The multilingual dataset obtains good performance, with an F1 Macro score of 0.71, an F1 SUBJ of 0.69, and an accuracy of 0.71. The model also shows good performance for the Italian language with an F1 Macro score of 0.74 and strong subjective class metrics, with an F1 SUBJ of 0.64. On the other hand, the model struggles the most with the Arabic dataset, obtaining an F1 Macro score of 0.49 and an accuracy of 0.52. The performance is relatively lower than in other languages, which shows the difficulty in identifying subjective data. The model performs well in Bulgarian, achieving an F1 Macro score of 0.72 and high subjective class performance with an F1 SUBJ of 0.69. For English, the performance is moderate to good, with an F1 Macro score of 0.68. The model handles subjective data in English relatively better, with an F1 SUBJ of 0.54, precision (P SUBJ) of 0.52, and recall (R SUBJ) of 0.64. The overall accuracy for English is 0.64. In summary, the model shows the highest performance in German, followed by Italian and Bulgarian, with Arabic being the most challenging language for the model. The performance in English is moderate, and the overall multilingual performance is strong, suggesting the model's effectiveness across multiple languages but with some variability in specific language performance. • F1 Macro: The macro-averaged F1 score, which is the harmonic mean of precision and recall across all classes. • P Macro: The macro-averaged precision.</p><p>• R Macro: The macro-averaged recall.</p><p>• F1 SUBJ: The F1 score for subjective classification.</p><p>• P SUBJ: The precision for subjective classification.</p><p>• R SUBJ: The recall for subjective classification.</p><p>• Accuracy: The overall accuracy of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Comparison with Other Teams</head><p>We compare the performance of our subjectivity classification model with that of other participating teams in the CheckThat! Lab at CLEF 2024. Table <ref type="table" target="#tab_3">3</ref> shows the official results for each language and the multilingual dataset. Our team achieves the highest rank in the German and multilingual categories, with Macro F1 scores of 0.7908 and 0.7121, respectively. We also secure the second position in Arabic and Bulgarian. For Arabic, our model achieved second place with a Macro F1 score of 0.4908 and a SUBJ F1 score of 0.37. In Bulgarian, our model also secured second place with a Macro F1 score of 0.7169 and a SUBJ F1 score of 0.69. In Italian, our model ranks third with a Macro F1 score of 0.7430 and a SUBJ F1 score of 0.64. In the English category, our model ranks ninth with a Macro F1 score of 0.6893 and a SUBJ F1 score of 0.54.</p><p>These results showcase the competitiveness of our approach in the shared task, especially in the German and multilingual categories. They also indicate areas for improvement, particularly in English, where our model's performance is lower than other teams. Overall, our team's participation in the ArAIEval shared task demonstrated strong performance across multiple languages, securing top ranks in several categories and showcasing our model's capabilities in multilingual data and subjective data evaluation. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussion</head><p>Our system leveraged state-of-the-art pre-trained language models, specifically BERT, which we finetuned for subjectivity classification task. Through extensive experiments, we demonstrated the ef-fectiveness of our approach, achieving competitive performance in various languages. Our system ranked first in the German and multilingual categories, second in Arabic and Bulgarian, and third in Italian. These results highlight the robustness of our model and its ability to generalize across different languages. We also investigated the impact of various preprocessing techniques, such as part-of-speech tagging and attention masking, on the performance of our system. Furthermore, our analysis of the dataset characteristics revealed the challenges posed by the imbalance between objective and subjective sentences across all languages. This imbalance underscores the need for developing strategies to handle skewed data distributions effectively.</p><p>Our work contributes to the growing body of research on subjectivity classification and multilingual natural language processing. The insights gained from our experiments can inform future research directions and help develop more robust and accurate systems for subjectivity analysis across diverse languages.</p><p>However, our study also has some limitations. The performance of our system in English was relatively lower compared to other languages, indicating room for improvement. Future work could explore more advanced techniques, such as domain adaptation and transfer learning, to enhance the model's performance in English and other languages. Moreover, the scope of our study was limited to the dataset provided by the CheckThat! Lab. Further research could investigate the generalizability of our approach to other datasets and domains, such as social media and customer reviews.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In conclusion, our subjectivity classification system, Nullpointer, demonstrates the potential of leveraging pre-trained language models and multilingual approaches for identifying subjective and objective statements in news articles. As the volume of online content continues to grow, the ability to automatically distinguish between subjective and objective information becomes increasingly crucial. Our work contributes to this important research area and paves the way for more advanced and reliable subjectivity analysis systems in the future.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Diagram for classification of subjectivity in text sequence</figDesc><graphic coords="3,128.41,65.61,338.45,323.48" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Training and Test Data Distribution</figDesc><table><row><cell>Language</cell><cell>Dataset</cell><cell cols="2">OBJ (N) (%) SUBJ (N) (%)</cell></row><row><cell>Arabic</cell><cell cols="2">Train (1185) 905 (76.37)</cell><cell>280 (23.63)</cell></row><row><cell></cell><cell>Test (748)</cell><cell>425 (56.81)</cell><cell>323 (43.18)</cell></row><row><cell>Bulgarian</cell><cell>Train (729)</cell><cell>406 (55.69)</cell><cell>323 (44.23)</cell></row><row><cell></cell><cell>Test (250)</cell><cell>143 (57.2)</cell><cell>107 (42.8)</cell></row><row><cell>English</cell><cell>Train (830)</cell><cell>532 (64.09)</cell><cell>298 (35.9)</cell></row><row><cell></cell><cell>Test (484)</cell><cell>362 (74.79)</cell><cell>122 (25.2)</cell></row><row><cell>German</cell><cell>Train (800)</cell><cell>492 (61.5)</cell><cell>308 (38.5)</cell></row><row><cell></cell><cell>Test (337)</cell><cell>226 (67.07)</cell><cell>111 (32.93)</cell></row><row><cell>Italian</cell><cell cols="2">Train (1613) 1231 (76.31)</cell><cell>382 (23.68)</cell></row><row><cell></cell><cell>Test (513)</cell><cell>377 (73.4)</cell><cell>136 (26.5)</cell></row><row><cell cols="4">Multilingual Train (5159) 3568 (69.16) 1591 (30.83)</cell></row><row><cell></cell><cell>Test (500)</cell><cell>250 (50)</cell><cell>250 (50)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Performance metrics across different languages</figDesc><table><row><cell>Language</cell><cell cols="7">F1 Macro P Macro R Macro F1 SUBJ P SUBJ R SUBJ Accuracy</cell></row><row><cell>Arabic</cell><cell>0.49</cell><cell>0.49</cell><cell>0.50</cell><cell>0.37</cell><cell>0.43</cell><cell>0.33</cell><cell>0.52</cell></row><row><cell>Bulgarian</cell><cell>0.72</cell><cell>0.72</cell><cell>0.72</cell><cell>0.69</cell><cell>0.66</cell><cell>0.72</cell><cell>0.72</cell></row><row><cell>English</cell><cell>0.68</cell><cell>0.43</cell><cell>0.50</cell><cell>0.54</cell><cell>0.52</cell><cell>0.64</cell><cell>0.64</cell></row><row><cell>German</cell><cell>0.79</cell><cell>0.78</cell><cell>0.81</cell><cell>0.73</cell><cell>0.67</cell><cell>0.80</cell><cell>0.81</cell></row><row><cell>Italian</cell><cell>0.74</cell><cell>0.73</cell><cell>0.77</cell><cell>0.64</cell><cell>0.57</cell><cell>0.73</cell><cell>0.78</cell></row><row><cell>Multilingual</cell><cell>0.71</cell><cell>0.72</cell><cell>0.71</cell><cell>0.69</cell><cell>0.76</cell><cell>0.63</cell><cell>0.71</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Official results for six test languages in Subtask2 CheckThat! Lab at CLEF 2024</figDesc><table><row><cell>Language</cell><cell>Team</cell><cell cols="3">Rank Macro F1 SUBJ F1</cell></row><row><cell>Arabic</cell><cell>IAI Group</cell><cell>1</cell><cell>0.4947</cell><cell>0.46</cell></row><row><cell></cell><cell>Nullpointer</cell><cell>2</cell><cell>0.4908</cell><cell>0.37</cell></row><row><cell></cell><cell>Baseline</cell><cell>3</cell><cell>0.4852</cell><cell>0.40</cell></row><row><cell></cell><cell>JUNLP (last)</cell><cell>7</cell><cell>0.3623</cell><cell>0.00</cell></row><row><cell>Bulgarian</cell><cell>Baseline</cell><cell>1</cell><cell>0.7531</cell><cell>0.73</cell></row><row><cell></cell><cell>Nullpointer</cell><cell>2</cell><cell>0.7169</cell><cell>0.69</cell></row><row><cell></cell><cell>Hybrinfox</cell><cell>3</cell><cell>0.7147</cell><cell>0.65</cell></row><row><cell></cell><cell>JUNLP (last)</cell><cell>5</cell><cell>0.3639</cell><cell>0.00</cell></row><row><cell>English</cell><cell>Hybrinfox</cell><cell>1</cell><cell>0.7442</cell><cell>0.60</cell></row><row><cell></cell><cell>Nullpointer</cell><cell>9</cell><cell>0.6893</cell><cell>0.54</cell></row><row><cell></cell><cell>Baseline</cell><cell>11</cell><cell>0.6346</cell><cell>0.45</cell></row><row><cell></cell><cell>IAI Group (last)</cell><cell>15</cell><cell>0.4491</cell><cell>0.39</cell></row><row><cell>German</cell><cell>Nullpointer</cell><cell>1</cell><cell>0.7908</cell><cell>0.73</cell></row><row><cell></cell><cell>IAI Group</cell><cell>2</cell><cell>0.7302</cell><cell>0.66</cell></row><row><cell></cell><cell>Baseline</cell><cell>3</cell><cell>0.6994</cell><cell>0.63</cell></row><row><cell></cell><cell>Hybrinfox (last)</cell><cell>4</cell><cell>0.6968</cell><cell>0.57</cell></row><row><cell>Italian</cell><cell>JK_PCIC_UNAM</cell><cell>1</cell><cell>0.7917</cell><cell>0.69</cell></row><row><cell></cell><cell>Nullpointer</cell><cell>3</cell><cell>0.7430</cell><cell>0.64</cell></row><row><cell></cell><cell>Baseline</cell><cell>4</cell><cell>0.6503</cell><cell>0.52</cell></row><row><cell></cell><cell>IAI Group (last)</cell><cell>5</cell><cell>0.5862</cell><cell>0.49</cell></row><row><cell>Multilingual</cell><cell>Nullpointer</cell><cell>1</cell><cell>0.7121</cell><cell>0.69</cell></row><row><cell></cell><cell>Hybrinfox</cell><cell>2</cell><cell>0.6849</cell><cell>0.63</cell></row><row><cell></cell><cell>Baseline</cell><cell>3</cell><cell>0.6697</cell><cell>0.66</cell></row><row><cell></cell><cell>IAI Group (last)</cell><cell>4</cell><cell>0.6292</cell><cell>0.67</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://carpedm20.github.io/emoji/docs/api.html#emoji.demojize</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Acknowledgments</head><p>We acknowledge Qatar National Research Fund grant NPRP14C0916-210015 from the Qatar Research Development and Innovation Council (QRDI) for funding this research.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Opinionfinder: A system for subjectivity analysis</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wilson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hoffmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Somasundaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kessler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wiebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cardie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Riloff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Patwardhan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of HLT/EMNLP 2005 Interactive Demonstrations</title>
				<meeting>HLT/EMNLP 2005 Interactive Demonstrations</meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="34" to="35" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Beyond subjective and objective in statistics</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gelman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hennig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the Royal Statistical Society Series A: Statistics in Society</title>
		<imprint>
			<biblScope unit="volume">180</biblScope>
			<biblScope unit="page" from="967" to="1033" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Decline of a paradigm? bias and objectivity in news media studies</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Hackett</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Critical Studies in Media Communication</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="229" to="259" />
			<date type="published" when="1984">1984</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Learning personal human biases and representations for subjective tasks in natural language processing</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kocoń</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gruza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bielaniewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Grimling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kanclerz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Miłkowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kazienko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Data Mining (ICDM), IEEE</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="1168" to="1173" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Subjectivity, identification and differentiation: Key issues in early social development</title>
		<author>
			<persName><forename type="first">U</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carpendale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bibok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Racine</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Monographs of the Society for Research in Child Development</title>
		<imprint>
			<biblScope unit="page" from="167" to="179" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Using nlp approach for opinion types classifier</title>
		<author>
			<persName><forename type="first">M</forename><surname>Othman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Moawad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Idrees</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dimitrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Galassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Siegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<title level="m">Overview of the CLEF-2024 CheckThat! lab task 2 on subjectivity in news articles</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Development and use of a gold-standard data set for subjectivity classifications</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wiebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bruce</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>O'hara</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 37th annual meeting of the Association for Computational Linguistics</title>
				<meeting>the 37th annual meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="246" to="253" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Semeval-2016 task 4: Sentiment analysis in twitter</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ritter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rosenthal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Sebastiani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th international workshop on semantic evaluation</title>
				<meeting>the 10th international workshop on semantic evaluation<address><addrLine>SemEval-</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016</date>
			<biblScope unit="page" from="1" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Opinion mining on newspaper quotations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Balahur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Steinberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Van Der Goot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Pouliquen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kabadjov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology-Volume 03</title>
				<meeting>the 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology-Volume 03</meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="523" to="526" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Learning multilingual subjective language via cross-lingual projections</title>
		<author>
			<persName><forename type="first">R</forename><surname>Mihalcea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Banea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wiebe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics</title>
				<meeting>the 45th Annual Meeting of the Association of Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="976" to="983" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">The clef-2024 checkthat! lab: Check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Przybyła</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Goharian</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Tonellotto</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Lipani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Mcdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</editor>
		<meeting><address><addrLine>Nature Switzerland, Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="449" to="458" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">The clef-2022 checkthat! lab on fighting the covid-19 infodemic and fake news detection</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barr'on-Cede No</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>M'ıguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kutlu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="416" to="428" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Fighting the covid-19 infodemic in social media: a holistic perspective and a call to arms</title>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Dalvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Durrani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Sajjad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International AAAI Conference on Web and Social Media</title>
				<meeting>the International AAAI Conference on Web and Social Media</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="913" to="922" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Bert post-training for review reading comprehension and aspectbased sentiment analysis</title>
		<author>
			<persName><forename type="first">H</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Shu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="2324" to="2335" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Adapting bert for target-oriented multimodal sentiment classification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th International Joint Conference on Artificial Intelligence</title>
				<meeting>the 28th International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="5408" to="5414" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
