<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Indigo at CheckThat! 2024: Using Setfit: A Resource Efficient Technique for Subjectivity Detection in News Article Notebook for the CheckThat! Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Soumyadeep</forename><surname>Sar</surname></persName>
							<email>soumyadeepsar26@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Indian Institute Of Science Education and Research Kolkata</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dwaipayan</forename><surname>Roy</surname></persName>
							<email>dwaipayan.roy@iiserkol.ac.in</email>
							<affiliation key="aff0">
								<orgName type="institution">Indian Institute Of Science Education and Research Kolkata</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Indigo at CheckThat! 2024: Using Setfit: A Resource Efficient Technique for Subjectivity Detection in News Article Notebook for the CheckThat! Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">9561AA0A9AAC982B2DBBEE84481A7D91</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:04+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>SetFit</term>
					<term>Sentence Transformers</term>
					<term>Few shot text classification</term>
					<term>Deep Learning</term>
					<term>Fine-tuning</term>
					<term>Natural Language processing</term>
					<term>LLMs</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The spread of misinformation and biased news across various reliable news outlets has led to serious consequences in our society. It has become crucial to understand the patterns in such misleading news articles, identify key evidence, and learn how to recognize false information. Subjectivity can play a pivotal role in identifying misleading news. In this work, we employed a resource-efficient method called SetFit on the English sub-task for Task-2 (subjectivity detection) at CheckThat!. This technique uses a few shot examples from training data and aims to produce results comparable to those of fully fine-tuned models like BERT on the entire dataset. For the selected sample data, we filter out conflict-resolved instances from the dataset and combine them with some other chosen data-points, then train our desired models on this dataset.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the realm of natural language processing, accurately detecting subjectivity in text is a crucial task for many reasons. It allows us to distinguish between objective information and content skewed by personal biases and opinions. This is particularly important in the digital age, where opinions and biases spread rapidly through media, potentially influencing public perception. Traditional methods for subjectivity detection often rely on large, labelled datasets, which can be expensive and time-consuming to create. Additionally, achieving high performance often involves fine-tuning large language models (LLMs), further increasing computational costs.</p><p>So we explore a resource-efficient approach at CLEF 2024 Check that! <ref type="bibr" target="#b0">[1]</ref> for the Task-2 (Subjectivity Detection in News Articles) <ref type="bibr" target="#b1">[2]</ref>. We only employed our methods on the English language sub-task of the challenge. We leverage SetFit <ref type="bibr" target="#b2">[3]</ref>, a few-shot learning algorithm that utilizes sentence embeddings and contrastive learning to efficiently fine-tune a model even with limited data. This technique enables the model to learn quickly with minimal labelled examples and requires fewer training epochs compared to traditional fine-tuning. Additionally, SetFit can be run on CPUs, eliminating the need for expensive GPUs. This study investigates the effectiveness of SetFit in distinguishing subjective and objective sentences within news articles. Our goal is to provide a robust and scalable solution for subjectivity detection, paving the way for more efficient and accurate identification of subjective content.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>The ability to distinguish between objective and subjective information in text is crucial for various natural language processing tasks. This distinction is particularly important in the digital age, where opinions and biases spread rapidly through media, potentially influencing public perception. There are several works that deal with the problem. The work by Chaturvedi et al. <ref type="bibr" target="#b3">[4]</ref> provide a comprehensive review of subjectivity detection methods, categorizing them into three types: hand-crafted, automatic, and multi-modal. Hand-crafted methods, while effective for identifying strong sentiments, struggle with weakly subjective sentences. Automatic methods, such as deep learning, overcome this limitation by creating meta-level feature representations that generalize well across domains and languages. Multi-modal methods further enhance accuracy by incorporating audio and video data with text using multiple kernels. This review highlights the advantages and limitations of each approach, emphasizing the challenges of high-dimensional n-gram features and the temporal nature of sentiments in long texts. In another work, Pang and Lee <ref type="bibr" target="#b4">[5]</ref> introduced a groundbreaking approach to subjectivity detection, focusing on identifying and categorizing subjective portions of a document to determine sentiment polarity. Their method utilizes text-categorization techniques and efficient graph-based algorithms to extract subjective text segments, allowing for the incorporation of cross-sentence contextual constraints. This approach significantly improves sentiment classification accuracy by targeting the relevant subjective content within the text.</p><p>Recent advancements have led to the development of resource-efficient methods for several textprocessing tasks. Abdedaiem et al. <ref type="bibr" target="#b5">[6]</ref> demonstrated the effectiveness of SetFit, that offers a highly efficient and prompt-free approach to fine-tuning Sentence Transformers (ST) for few-shot learning scenarios. Its two-stage process begins with contrastively fine-tuning a pre-trained ST model on a limited set of text pairs. This step leverages a Siamese architecture, where the model learns to distinguish similar and dissimilar sentences. The resulting fine-tuned ST then generates rich text embeddings, which are subsequently used to train a separate classification head for the specific task at hand. This elegant framework eliminates the need for handcrafted prompts, enabling accurate classification with minimal labeled data. In our work, we utilize SetFit to tackle the challenge of fake news detection in low-resource languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Background</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Baseline solution</head><p>The baseline solution for subjectivity detection employed a powerful combination of Sentence-BERT <ref type="bibr" target="#b6">[7]</ref> and Logistic Regression. Sentence-BERT, a pre-trained sentence encoder, was used to transform each statement into a high-dimensional vector representation, capturing its semantic meaning. This step provided a rich and informative representation of the sentence's content. Subsequently, a Logistic Regression classifier was trained on these sentence embeddings. This classifier learned to distinguish between objective and subjective statements based on the patterns and features present in the embeddings.</p><p>The complete details and code for this baseline solution are available on the CLEF 2024 CheckThat! Lab GitLab repository 1 , providing a valuable resource for researchers and practitioners interested in replicating or building upon this approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Dataset Distribution</head><p>We conducted a thorough analysis of the English dataset provided for the subjectivity detection task.</p><p>The dataset is structured in a tab-separated format, conveniently pre-split into training, development, and development-test sets.</p><p>A noteworthy observation is the class imbalance within the training data: subjective sentences constitute only 35.9% (298) of the training set, while objective sentences account for 64.1% (532). This imbalance is not present in the development and development-test sets, which exhibit a more balanced distribution. While development and development-test sets show balanced distributions, the final test set where model performance is measured is skewed towards objective sentences (74.8%, 362) with only 25.2% (122) subjective sentences. This imbalance in the test set should be considered when interpreting and analyzing the model's performance. Figure <ref type="figure" target="#fig_0">1</ref> visually depicts the distribution of subjective and objective sentences across the four dataset splits. This figure provides a clear overview of the class distribution within each set and highlights the potential challenges associated with the imbalanced test set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Schema for the Refined Dataset in Our Approach</head><p>As SetFit leverages contrastive learning, selecting a set of high-quality examples is crucial for its success. These examples will define the decision boundaries the model learns for the classification task. Given the limited number of examples used in few-shot learning, careful selection is essential.</p><p>In this work, we took an innovative approach by focusing on instances from the training set where annotator conflicts were resolved. This marks the first time in the competition that the 'solved-conflict' feature of the dataset has been utilized. We specifically chose 69 statements from the training split where solved-conflict had a value of True. These statements represent cases where human annotators initially disagreed on the subjectivity label, but ultimately reached a consensus. This consensus can be viewed as a strong indicator of the statement's true subjectivity or objectivity, making them valuable examples for contrastive learning. By focusing on these resolved-conflict instances, we aim to provide the model with clear and unambiguous examples, potentially leading to more robust and accurate decision boundaries for subjectivity classification. Some examples of such instances are shown in Table <ref type="table" target="#tab_0">1</ref>.</p><p>While resolved-conflict instances offer valuable training examples, relying solely on them could potentially lead to overfitting, as they represent a specific subset of the data. To address this, we incorporated a filtered selection of additional sentences from the training set. We identified the average number of words in each sentence across the entire training dataset as 22.84. This served as a baseline for filtering sentences based on word count. We aimed to include sentences with a moderate length, avoiding excessively long or short ones. Therefore, we implemented a filtering process where we gradually increased the word count threshold, starting from 24 words. We observed that including sentences with more than 32 words resulted in a significant performance drop due to the limited number of available examples. Hence, we created our sampled dataset accordingly and removed those data points for whom the conflict was resolved since we are already considering them in the conflict-resolved sampled data. This sampled data contains 145 statements. Ultimately, combining both the sampled datasets (69 and 145), we finally get 214 sentences from the training data for contrastive learning. This represents a substantial reduction compared to the full training set, highlighting the efficiency of the approach. Notably, the filtered dataset maintained a balanced distribution with 114 subjective and 100 objective sentences, ensuring a representative sample for model training. Figure <ref type="figure">2</ref> visually depicts the distribution of subjective and objective sentences within the filtered dataset.</p><p>In Table <ref type="table" target="#tab_1">2</ref> we demonstrate the improvement in performance of a Setfit Model with SVM classifier head when it is trained on a randomly sampled dataset (using default seed value of 42 for reproducibility) and on our specially designed sampled dataset. In both cases, they were trained for 1 epoch with the same hyperparameters. We clearly observe an improvement in performance of the Setfit Model when it is being trained on our dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Working of SetFit</head><p>SetFit leverages a streamlined two-stage procedure to enhance sentence transformers for classification tasks. This approach offers significant efficiency gains while maintaining accuracy, particularly in scenarios with limited labelled data. The stages of SetFit are described below: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.1.">Sentence Transformer Fine-tuning</head><p>The The core objective of contrastive learning in this stage is to:</p><p>• Minimize the distance: between the embeddings generated for positive pairs, ensuring that sentences with similar meanings have closely aligned representations. • Maximize the distance: between the embeddings generated for negative pairs, creating a clear differentiation between sentences with contrasting meaning and sentiment.</p><p>To achieve this, an appropriate loss function, such as the cosine similarity function, is employed to measure the semantic similarity between the sentences within each pair. The model is then fine-tuned to generate embeddings that effectively capture these semantic relationships, laying the foundation for accurate classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.2.">Classification Head Training</head><p>The fine-tuned sentence transformer from the first stage plays a crucial role in the second stage. It encodes each sentence, capturing the essence of the text data in a format suitable for classification. These sentence embeddings can then be utilized to train a variety of classical machine learning models for subjectivity classification such as: Support Vector Machines (SVMs) or Logistic Regression. Alternatively, Differentiable Linear Neural Layer, a neural network layer can be employed for classification. This layer can be trained with its own hyperparameters, potentially different from those used in the first stage. This flexibility allows for customization and fine-tuning of the classification head without retraining the entire model as a single end-to-end system.</p><p>By adopting this two-stage approach, SetFit effectively leverages the power of contrastive learning to enhance sentence transformers for subjectivity classification. It achieves this with remarkable efficiency, requiring significantly less data and computational resources compared to traditional fine-tuning methods. A simplified diagram for the entire process is described in the Fig. <ref type="figure" target="#fig_1">3</ref> 4. Experiment and Evaluation</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Model and Hyperparameters used</head><p>The field of pre-trained sentence transformers offers a wealth of options for various tasks. To select the most suitable model for our subjectivity classification task, we explored the extensive collection available on the sentence bert (SBERT) leaderboard<ref type="foot" target="#foot_0">2</ref> . These pre-trained models have been trained on a massive dataset exceeding one billion training pairs, making them well-equipped for general-purpose use. While all models offer robust performance, the following two models stand out to be significantly effective: 𝑖) all-mpnet-base-v2 model and 𝑖𝑖) all-MiniLM-L6-v2 model. all-MiniLM-L6-v2 model is known for its faster speed and good quality results. On the other hand all-mpnet-base-v2 model is the best model in the entire leaderboard of SBERT, producing the best quality embeddings for a variety of NLP tasks. Since all-mpnet-base-v2 model occupies the topmost position in the leaderboard of SBERT, we consider choosing it for our experiments over the all-MiniLM-L6-v2 model. For the fine-tuning process, we leveraged the filtered dataset of 214 sentences. We specifically trained our model for 1 epoch rather than trying for higher epochs in order to develop resource efficient models, which could perform as good as fully-finetuned LLMs. The cosine similarity was employed to evaluate the semantic similarity between the generated embeddings to provide a reliable measure of their alignment. The seed value was set to 42 to ensure the reproducibility of the results.</p><p>For the classification head, multiple scikit-learn based ML classifiers were tried like SVM, Random Forest, etc. We tested their performance using the same hyper-parameters and on the same developmenttest dataset available before the evaluation cycle of the competition. The performances are listed in the Table <ref type="table" target="#tab_2">3</ref>. The best performing classifier was the linear differentiable layer, which is being trained for classifying the sentence embeddings for 1 epoch. All the hyperparameters used for finetuning the sentence bert and the train the classification head is mentioned in Table <ref type="table">4</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Comparative Performance Analysis</head><p>We fully fine-tuned other BERT-based large language models (LLMs) on the entire training dataset of 830 sentences. The models used for comparison are-BERT <ref type="bibr" target="#b7">[8]</ref> and RoBERTa <ref type="bibr" target="#b8">[9]</ref> . The transfomers <ref type="bibr" target="#b9">[10]</ref> library from HuggingFace was being used to get pre-trained models of BERT and RoBERTa. The hyperparameters for training these models were kept the same as those for Setfit for fair comparison. We fine-tuned both the LLMs for 1 epoch and for 4 epochs, respectively. The 1 epoch fine-tuned model's performance will demonstrate the effectiveness of contrastive learning employed in Setfit, which uses nearly 4 times smaller data, yet produces a better result. A comparison is made between their performance on the development test (dev-test.tsv) and the test datasets provided during the evaluation period for the competition. The results are listed in Table <ref type="table" target="#tab_3">5</ref>. We discuss the results obtained in the Result section below. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Result</head><p>The performance in Table <ref type="table" target="#tab_3">5</ref> shows that Setfit performed much better when compared to fine-tuned LLMs, which were trained for 1 epoch. This demonstrates the effectiveness of contrastive learning that allows our model to get well-trained on a 4 times smaller sampled dataset within just 1 epoch. We also notice that the performance of fully fine-tuned models is better than our SetFit approach when they are trained for 4 epochs. This is an anticipated result since they are seeing the entire training dataset and getting a lot of time (epochs) to train. But training on the entire dataset and for a large number of epochs is clearly resource intensive, which from the beginning was our goal to avoid. So considering our approach as a resource-efficient technique the performance was fairly good.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this article, we report the performance of our proposed method for the CLEF 2024 task CheckThat.</p><p>Our approach has successfully outperformed the baseline solution, demonstrating its potential for subjectivity detection. The proposed approach was specifically focused on developing a computationally efficient technique that delivers competitive results compared to existing state-of-the-art models. Large models like GPT, while powerful, contribute significantly to the carbon footprint, posing a threat to our environment. This research delves into the development of lighter and more efficient NLP solutions, potentially paving the way for replacing massive LLMs in various applications in the near future.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Entire Datasets distribution.</figDesc><graphic coords="2,138.45,199.49,315.89,203.28" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: 2 step working phases in SetFit (https://huggingface.co/blog/assets/103_setfit/setfit_diagram_process. png).</figDesc><graphic coords="5,72.00,65.61,451.27,101.32" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>first stage commences with the provided few-shot training data. SetFit constructs sentence pairs from this training data, enabling the model to grasp the relationships and context within the text. Specifically the following pairs are strategically formed: • Positive pairs: Sentences belonging to the same class are coupled together, representing examples of similar meaning and sentiment. • Negative pairs: Sentences from different classes are paired together, showcasing contrasting meanings and sentiments.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="4,139.69,157.74,315.89,203.28" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Conflict resolved Statements in our dataset. Gone are the days when they led the world in recession-busting SUBJ True Whether this would be all for the best, or otherwise, is not yet the point. SUBJ True Is this a warning of what's to come? OBJ True "That is to say, we have not consciously intended it. " OBJ True</figDesc><table><row><cell>sentence</cell><cell>label solved conflict</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Improvement in performance using our technique of data sampling.</figDesc><table><row><cell>Sampled Training dataset</cell><cell cols="2">Macro-avg F1 SUBJ F1</cell></row><row><cell>Randomly sampled dataset</cell><cell>0.73</cell><cell>0.72</cell></row><row><cell>Dataset designed according to our schema</cell><cell>0.75</cell><cell>0.78</cell></row></table><note>Figure 2: Filtered Dataset distribution.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Performance of different Classification heads tried for SetFit.</figDesc><table><row><cell>Classification head</cell><cell cols="2">Macro-avg F1 score SUBJ F1 score</cell></row><row><cell>Random Forest head</cell><cell>0.72</cell><cell>0.78</cell></row><row><cell>SVM Classifier head</cell><cell>0.74</cell><cell>0.78</cell></row><row><cell>Linear Differentiable head</cell><cell>0.74</cell><cell>0.79</cell></row><row><cell>Table 4</cell><cell></cell><cell></cell></row><row><cell>Hyperparameters used for SetFit.</cell><cell></cell><cell></cell></row><row><cell>Hyperparameter</cell><cell></cell><cell>Values</cell></row><row><cell>batch size</cell><cell></cell><cell>8</cell></row><row><cell cols="2">number of epochs (for both the training phases)</cell><cell>1</cell></row><row><cell cols="2">learning rate (head and body training)</cell><cell>2e-05</cell></row><row><cell>seed</cell><cell></cell><cell>42</cell></row><row><cell>loss function</cell><cell></cell><cell>Cosine Similarity function</cell></row><row><cell>number of iterations</cell><cell></cell><cell>20</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 5</head><label>5</label><figDesc>Comparative study of performances of different models.</figDesc><table><row><cell></cell><cell></cell><cell>Dev-test</cell><cell></cell><cell>Test</cell></row><row><cell>Models</cell><cell cols="4">SUBJ F1 Macro-avg F1 SUBJ F1 Macro-avg F1</cell></row><row><cell>BERT(1 epoch)</cell><cell>0.69</cell><cell>0.63</cell><cell>0.35</cell><cell>0.61</cell></row><row><cell>RoBERTa (1 epoch)</cell><cell>0.56</cell><cell>0.64</cell><cell>0.23</cell><cell>0.55</cell></row><row><cell>Setfit (1 epoch)</cell><cell>0.79</cell><cell>0.74</cell><cell>0.47</cell><cell>0.64</cell></row><row><cell>BERT (4 epochs)</cell><cell>0.74</cell><cell>0.75</cell><cell>0.57</cell><cell>0.73</cell></row><row><cell cols="2">RoBERTa (4 epochs) 0.77</cell><cell>0.78</cell><cell>0.54</cell><cell>0.71</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://sbert.net/docs/sentence_transformer/pretrained_models.html</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The clef-2024 checkthat! lab: Check-worthiness, subjectivity, persuasion, roles, authorities, and adversarial robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Przybyła</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Goharian</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Tonellotto</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Lipani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Mcdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</editor>
		<meeting><address><addrLine>Nature Switzerland, Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="449" to="458" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dimitrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Galassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Siegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<title level="m">Overview of the CLEF-2024 CheckThat! lab task 2 on subjectivity in news articles</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Efficient few-shot learning without prompts</title>
		<author>
			<persName><forename type="first">L</forename><surname>Tunstall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><forename type="middle">E S</forename><surname>Jo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bates</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wasserblat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Pereg</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2209.11055</idno>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Distinguishing between facts and opinions for sentiment analysis: Survey and challenges</title>
		<author>
			<persName><forename type="first">I</forename><surname>Chaturvedi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Cambria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">E</forename><surname>Welsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Herrera</surname></persName>
		</author>
		<idno type="DOI">10.1016/J.INFFUS.2017.12.006</idno>
		<ptr target="https://doi.org/10.1016/j.inffus.2017.12.006.doi:10.1016/J.INFFUS.2017.12.006" />
	</analytic>
	<monogr>
		<title level="j">Inf. Fusion</title>
		<imprint>
			<biblScope unit="volume">44</biblScope>
			<biblScope unit="page" from="65" to="77" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts</title>
		<author>
			<persName><forename type="first">B</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lee</surname></persName>
		</author>
		<idno type="DOI">10.3115/1218955.1218990</idno>
		<ptr target="https://aclanthology.org/P04-1035/.doi:10.3115/1218955.1218990" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Scott</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><surname>Daelemans</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Walker</surname></persName>
		</editor>
		<meeting>the 42nd Annual Meeting of the Association for Computational Linguistics<address><addrLine>Barcelona, Spain, ACL</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004-07-26">21-26 July, 2004. 2004</date>
			<biblScope unit="page" from="271" to="278" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Fake news detection in low resource languages using setfit framework</title>
		<author>
			<persName><forename type="first">A</forename><surname>Abdedaiem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Dahou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Chéragui</surname></persName>
		</author>
		<idno type="DOI">10.4114/INTARTIF.VOL26ISS72PP178-201</idno>
		<ptr target="https://doi.org/10.4114/intartif.vol26iss72pp178-201.doi:10.4114/INTARTIF.VOL26ISS72PP178-201" />
	</analytic>
	<monogr>
		<title level="j">Inteligencia Artif</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="178" to="201" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1908.10084" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">BERT: pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/V1/N19-1423</idno>
		<ptr target="https://doi.org/10.18653/v1/n19-1423.doi:10.18653/V1/N19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Burstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Doran</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Solorio</surname></persName>
		</editor>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019<address><addrLine>Minneapolis, MN, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">June 2-7, 2019. 2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Roberta: A robustly optimized bert pretraining approach</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.11692</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Transformers: State-of-the-art natural language processing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Delangue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cistac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Louf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Funtowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Davison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shleifer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Von Platen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jernite</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Plu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gugger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Drame</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lhoest</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Rush</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.emnlp-demos.6" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</title>
				<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="38" to="45" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
