<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Pixel Phantoms at Touché: Ideology and Power Identification in Parliamentary Debates using Linear SVC</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Janani Hariharakrishnan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jithu Morrison S</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>P Mirunalini</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, Sri Sivasubramaniya Nadar College of Engineering</institution>
          ,
          <addr-line>Chennai, Tamil Nadu</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>The research addresses two fundamental tasks in parliamentary discourse analysis: the first one is identifying the political ideology of speakers, and the other one is categorizing their party afiliation as either governing or opposition. Analyzing parliamentary speeches involves discerning the ideological leanings of speakers' parties based on their articulated beliefs and determining whether a party is in a governing or opposition role within legislative contexts, which is crucial for understanding their political dynamics. These tasks are formulated as binary classification problems. Pixel Phantoms at Touché presents a framework for classifying speakers' political ideology and party afiliation in parliamentary debates by using a diverse set of models including Linear SVC, Logistic Regression, and DistilBERT. We proposed three diferent models, In the first model, DistilBERT was used for extracting the embeddings, and then Logistic Regression was applied for classification. In the other two models, Linear SVC and Logistic Regression models were built and trained for classification based on the extracted TF-IDF vectorization features. The study utilizes a multilingual corpus derived from ParlaMint, encompassing parliamentary speeches across various languages and jurisdictions. The proposed system was evaluated using F1 score and found Linear SVC emerging as the top-performing model. It achieved F1 scores of 0.5921 and 0.66 in the orientation and power test data, respectively, showcasing its robustness in handling complex political discourse data.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Multilingual Text Classification</kwd>
        <kwd>Grid Search</kwd>
        <kwd>Neural Network Architectures</kwd>
        <kwd>Natural Language Processing (NLP)</kwd>
        <kwd>TF-IDF (Term Frequency-Inverse Document Frequency)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Understanding the intricate landscape of political discourse within national parliaments necessitates a
thorough analysis of two critical variables: political ideology and party alignment. This research aims to
elucidate the ideological leanings of parliamentary speakers and their afiliations with either governing
or opposition parties by systematically analyzing parliamentary speeches. Through this study, the
multifaceted nature of political communication within legislative bodies is explored, contributing to a
deeper comprehension of the forces shaping public policy and governance.</p>
      <p>Traditionally, the task of categorizing parliamentary speeches based on political ideology and
party alignment has been approached through manual methods, relying on human expertise and
subjective judgment. However, this manual process is not without its drawbacks. It is time-consuming,
labor-intensive, and susceptible to biases inherent in human interpretation. Moreover, the sheer volume
of parliamentary data poses significant challenges in maintaining consistency and accuracy across
analyses. These limitations underscore the need for innovative automated solutions capable of handling
the complexity and scale of political discourse data.</p>
      <p>In recent years, advancements in machine learning and deep learning have revolutionized the
ifeld of natural language processing, ofering promising avenues for automated analysis of textual
data. Techniques such as neural networks, recurrent neural networks (RNNs), and transformers have
demonstrated remarkable capabilities in tasks ranging from text classification to sentiment analysis.
Leveraging these techniques, researchers have begun exploring automated approaches to political
discourse analysis, aiming to overcome the limitations of traditional manual methods.</p>
      <p>In this context, the research work represents a significant innovation in the field of political discourse
analysis. By harnessing state-of-the-art deep learning models and leveraging advanced computational
techniques, the team aims to develop a robust framework for classifying parliamentary speeches based
on political ideology and party alignment. The approach seeks to address the shortcomings of traditional
methods by providing a scalable, eficient, and objective means of analyzing political discourse. Through
the innovative solution, the team endeavors to contribute to the advancement of computational methods
in political science and facilitate deeper insights into the dynamics of parliamentary debates.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        The dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] derived from ParlaMint, which comprises corpora of transcribed parliamentary
speeches from 29 national and regional parliaments. Statistics on the dataset are presented, along
with baseline results obtained using a simple classifier. These results pertain to predicting political
orientation on the left-to-right axis and distinguishing between speeches delivered by governing
coalition party members and those from opposition party members. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], authors analysis neoliberal,
traditionalist, and social justice perspectives, provided crucial insights for the power identification
task. The research work in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] focuses on linguistic hierarchies and the marginalization of non-oficial
languages informed the approach to detecting power dynamics in parliamentary debates. The study also
aided in developing a more nuanced understanding of how linguistic diversity impacts political discourse.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], provides crucial insights which enhances feature engineering and data annotation for accurate
classification of political ideology and party afiliation. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Wang and Banko describe practical
methods for implementing and optimizing transformer models to handle text classification tasks across
multiple languages. In [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], authors reviewed text vectorization methods for sentiment classification,
covering Tf-Idf, Word2vec, and Doc2vec and the study emphasizes the significance of selecting
suitable vectorization techniques tailored to the characteristics and objectives of sentiment analysis tasks
      </p>
      <p>
        The research work [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] provides a review of political sentiment analysis techniques for tweets,
covering lexicon-based and machine learning methods. In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], the research ofers insights of
innovative methodologies in enhancing information access and retrieval across a range of
linguistic and contextual settings, including experimental approaches that integrate multilinguality,
multimodality, and interaction within information retrieval. In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], Owais Ahmad et al. hybridize
Sentence Transformer embeddings with Multi KNN to enhance the accuracy and eficiency of
multi-label text classification tasks, specifically in biomedical literature analysis. In [ 10], Varun
Dogra et al. analyze DistilBERT for sentiment classification of banking financial news. The
research provides insights into leveraging modern NLP models techniques for accurate sentiment analysis.
      </p>
      <p>In [11], Hardik Jadia conducts a comparative analysis of sentiment analysis techniques, including
Support Vector Machines (SVM), Logistic Regression, and TF-IDF feature extraction. This research
provides insights into the efectiveness of these methods for text classification tasks. The findings ofer
guidance for optimizing models to accurately identify political ideologies and party afiliations in
parliamentary speeches, contributing to more reliable computational political analysis. In [12], the
comparison of supervised classification models on textual data highlights the efectiveness of diferent
algorithms for text classification tasks. This research work underscores the importance of selecting and
optimizing classification models for accurately identifying political ideologies and party afiliations in
parliamentary speeches, thereby improving the overall performance and reliability of the analysis.</p>
      <p>The literature review emphasizes the application of advanced NLP techniques and computational
methods to enhance the accuracy of identifying political ideologies and party afiliations in parliamentary
discourse. Insights from referenced studies inform the integration of transformer models like DistilBERT,
logistic regression, and LinearSVC, ofering a structured approach for computational political analysis.</p>
    </sec>
    <sec id="sec-3">
      <title>3. System Overview</title>
      <p>This research work is structured around two primary objectives within parliamentary discourse
analysis: distinguishing speakers’ political ideologies and classifying their party afiliations into
governing or opposition categories. To achieve these goals, separate binary classification models were
employed for both orientation (political ideology) and power classification. Models such as Linear SVC
and Logistic Regression were utilized for both tasks, alongside a DistilBERT-based model for generating
text embeddings, followed by Logistic Regression for prediction.</p>
      <p>In this task, all training was conducted on the English translations of the parliamentary speeches.
The dataset includes both the original texts and their English translations (where available), and we
focused on the English texts to maintain consistency and leverage the capabilities of pre-trained
language models like DistilBERT, which are primarily trained on English data. By doing so, it is ensured
that the feature extraction using TF-IDF and sentence embeddings was uniform across the dataset,
allowing for efective training and evaluation of our models.</p>
      <p>Separate models were developed for both distinguishing political orientation and categorizing power
dynamics within parliamentary discourse using approaches like Linear SVC, Logistic Regression, and
DistilBERT combined with Logistic Regression.</p>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>The dataset for this task is derived from ParlaMint, a multilingual comparable corpus of parliamentary
debates. The data is provided in tab-separated text files, containing the following fields: id, speaker,
sex, text, text_en, and label. Each entry represents a speech, with information on the speaker and
automatic translations where available. The dataset covers parliamentary speeches from various
national and regional parliaments, and is designed to minimize confounding variables like speaker identity.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Data Preprocessing</title>
        <p>During the data preprocessing phase, missing values were removed and the columns with
null values were replaced with empty strings. TF-IDF features were extracted from the
English text using Tfidf Vectorizer with specified parameters, and were used in Linear SVC
and Logistic Regression. The text data (X_train and X_val) was transformed into sentence
embeddings using the pre-trained DistilBERT model from the Sentence Transformers library
(SentenceTransformer(’distilbert-base-nli-mean-tokens’)). This step converted the
text data into numerical vectors suitable for machine learning algorithms, which were used in
DistilBERT.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Feature Extraction using DistilBERT for Text Classification</title>
        <p>DistilBERT excels in capturing semantic meaning and relationships within sentences, making it highly
versatile for various natural language processing tasks like sentiment analysis, question answering,
document classification, and named entity recognition. Its embedding process converts text into
numerical representations that encode deep contextual information, enabling precise model predictions
across nuanced linguistic contexts. This capability to extract meaningful embeddings from input text is
essential for a wide array of NLP applications.</p>
        <p>After preprocessing, the model initiates by utilizing the "distilbert-base-nli-mean-tokens" pre-trained
transformer from the Sentence Transformers library to convert parliamentary speeches into dense
semantic embeddings. These embeddings serve as high-dimensional features fed directly into a Logistic
Regression classifier.</p>
        <p>Following embedding generation, the Logistic Regression classifier is directly trained on these
embeddings. This training phase involves optimizing the Logistic Regression’s parameters, including
regularization strength (’C’) and penalties (’l1’ and ’l2’), using GridSearchCV. This hyperparameter
tuning ensures that the model maximizes its predictive accuracy based on the embeddings’ semantic
content.This methodology guarantees that the model efectively leverages the semantic information
distilled by DistilBERT to make precise predictions regarding the political ideologies and party
afiliations of speakers based solely on their parliamentary speeches.</p>
        <p>The adoption of an alternative model from considerations of computational complexity and resource
requirements, particularly given the size of the dataset and the robust nature of the classification
task. Moreover, the interpretability ofered by Logistic Regression’s feature coeficients seemed more
congruent.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Text Classification Using Logistic Regression</title>
        <p>Logistic Regression is fundamental in supervised learning for its interpretability and eficacy across
various domains, prominently in natural language processing (NLP). In tasks like sentiment analysis
and political discourse classification, Logistic Regression harnesses TF-IDF vectorization to convert
text into numerical features. This process optimizes parameters such as the maximum feature count
(10,000), inclusion of both unigrams and bigrams (-grams of length 1 and 2), and thresholds for
document frequencies (min_df =2 and max_df =0.9). Additionally, common English stop words are
removed during the vectorization process.</p>
        <p>GridSearchCV refines these parameters to maximize predictive accuracy, ensuring the model performs
optimally. Post-vectorization, Logistic Regression incorporates these optimized features, adjusting
class weights to manage dataset imbalances and ensure balanced learning outcomes. Utilizing a fixed
random state guarantees consistent results across diferent runs. Leveraging the Scikit-learn library,
these methodologies provide robust and scalable solutions for diverse NLP applications. This integrated
approach underscores Logistic Regression’s versatility and pivotal role in converting raw textual data
into actionable insights through parameter optimization and efective model training.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Text Classification Using LinearSVC</title>
        <p>LinearSVC is renowned for its eficiency and efectiveness in text classification tasks, particularly
suited for high-dimensional data and widely applied in areas such as spam detection and sentiment
analysis. Like Logistic Regression, LinearSVC starts with TF-IDF vectorization to convert textual data
into numerical features. It employs similar TF-IDF settings, including the limitation of features,
exclusion of rare terms, and removal of common terms and stop words to enhance the quality of the feature set.</p>
        <p>Despite using similar TF-IDF settings, there is a distinction between Logistic Regression, which
predicts class probabilities using a logistic function, and LinearSVC, which separates classes with
a hyperplane based on maximum margin criteria, emphasizing clear class distinction rather than
probability estimation.</p>
        <p>During model training, LinearSVC adjusts class weights to manage imbalanced data distributions
efectively. Hyperparameter optimization via GridSearchCV focuses on parameters like the
regularization parameter  (tuned across values of [0.1, 1]) and maximum iterations (explored with [1000, 1700,
2500]), aiming to enhance model performance without overstating capabilities. Evaluation metrics,
including the F1 score, provide rigorous assessment of the model’s classification accuracy, ensuring
reliable performance validation in practical applications.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>We have experimented with three diferent models and performance of each model was analysed using
diferent measures such as F1 score, Recall and Precision. The results were tabulated in the Table 1 and
Table 2.</p>
      <p>Comparing the performance of three distinct models—Linear SVC, Logistic Regression, and
DistilBERT approach—reveals their respective strengths and nuances in handling the tasks of
identifying political ideology and categorizing party afiliation in parliamentary debates. These models
were evaluated for both tasks, assessing their capabilities and limitations in computational political
analysis.</p>
      <p>The Linear SVC model achieves the highest training and testing F1 scores of approximately 0.74 and
0.66, respectively, for categorizing party afiliation (Power) and 0.77 and 0.5921 for identifying political
ideology (Orientation). This showcases its robust capability in accurately classifying whether speakers’
parties are governing or in opposition and their political beliefs. Known for its efectiveness in handling
linearly separable data, Linear SVC constructs hyperplanes to delineate data points, ensuring high
precision and recall in this context.</p>
      <p>The Logistic Regression model achieves the training and testing F1 scores of approximately
0.73 and 0.657, respectively, for categorizing party afiliation (Power) and 0.77 and 0.5926 for
identifying political ideology (Orientation). Enhanced by Tfidf Vectorizer for feature extraction
and optimized hyperparameters, Logistic Regression efectively manages collinear features and
provides precise class probability estimates. This probabilistic framework makes it adept at
scenarios where the relationship between features and target variables is probabilistic rather than strictly linear.</p>
      <p>The DistilBERT model, achieves the training and testing F1 scores of approximately 0.74 and
0.453, respectively, for categorizing party afiliation (Power) and 0.74 and 0.563 for identifying
political ideology (Orientation). DistilBERT’s strength lies in its ability to encode complex, non-linear
relationships within the feature space, leveraging its capability as a powerful feature extraction tool.
DistilBERT is employed primarily for feature extraction, harnessing its capacity to uncover intricate
data patterns relevant to political discourse analysis.</p>
      <p>Overall, while Linear SVC demonstrates superior performance in this study for the task of categorizing
party afiliation and Logistic Regression for identifying political ideology, each model presents
tradeofs between complexity, interpretability, and performance. These insights underscore the nuanced
considerations necessary when applying machine learning to computational political analysis.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In conclusion, the Linear SVC model exhibits the highest validation F1 score, showcasing its robustness
in handling linearly separable data. Logistic Regression closely follows, demonstrating competitive
performance by eficiently managing collinear features and providing precise class probability estimates.
Meanwhile, the DistilBERT model ofers flexibility in capturing complex relationships but may be prone
to overfitting.
[10] V. Dogra, A. Singh, S. Verma, Kavita, N. Jhanjhi, M. Talib, Analyzing distilbert for sentiment
classification of banking financial news, in: Intelligent Computing and Innovation on Data Science:
Proceedings of ICTIDS 2021, Springer, 2021, pp. 501–510.
[11] H. Jadia, Comparative analysis of sentiment analysis techniques: Svm, logistic regression, and
tf-idf feature extraction (2023).
[12] B.-M. Hsu, Comparison of supervised classification models on textual data, Mathematics 8 (2020)
851.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Ç.</given-names>
            <surname>Çöltekin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kopp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Meden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Morkevicius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ljubešić</surname>
          </string-name>
          , T. Erjavec,
          <article-title>Multilingual power and ideology identification in the parliament: a reference dataset and simple baselines</article-title>
          ,
          <source>arXiv preprint arXiv:2405.07363</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Becker</surname>
          </string-name>
          , Identity, power, and
          <article-title>prestige in Switzerland's multilingual education</article-title>
          , transcript Verlag,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Erjavec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ogrodniczuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Osenova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ljubešić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Simov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pančur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rudolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kopp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barkarson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Steingrímsson</surname>
          </string-name>
          , Çöltekin, J. de Does,
          <string-name>
            <given-names>K.</given-names>
            <surname>Depuydt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Agnoloni</surname>
          </string-name>
          , G. Venturi,
          <string-name>
            <given-names>M. Calzada</given-names>
            <surname>Pérez</surname>
          </string-name>
          , L. D. de Macedo, C. Navarretta, G. Luxardo,
          <string-name>
            <given-names>M.</given-names>
            <surname>Coole</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rayson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Morkevičius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Krilavičius</surname>
          </string-name>
          , R. Darg´is,
          <string-name>
            <given-names>O.</given-names>
            <surname>Ring</surname>
          </string-name>
          , R. van Heusden,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marx</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fišer</surname>
          </string-name>
          ,
          <article-title>The parlamint corpora of parliamentary proceedings</article-title>
          ,
          <source>Language resources and evaluation 57</source>
          (
          <year>2022</year>
          )
          <fpage>415</fpage>
          -
          <lpage>448</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10579-021-09574-0.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cieciuch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vecchione</surname>
          </string-name>
          , E. Davidov,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Beierlein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Verkasalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-E.</given-names>
            <surname>Lönnqvist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Demirutku</surname>
          </string-name>
          , et al.,
          <source>Refining the Theory of Basic Individual Values, Journal of personality and social psychology 103</source>
          (
          <year>2012</year>
          ). doi:
          <volume>10</volume>
          .1037/a0029393.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Banko</surname>
          </string-name>
          ,
          <article-title>Practical transformer-based multilingual text classification</article-title>
          ,
          <source>in: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Papers</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>121</fpage>
          -
          <lpage>129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H. D.</given-names>
            <surname>Abubakar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Umar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Bakale</surname>
          </string-name>
          ,
          <article-title>Sentiment classification: Review of text vectorization methods: Bag of words, tf-idf, word2vec and doc2vec</article-title>
          ,
          <source>SLU Journal of Science and Technology</source>
          <volume>4</volume>
          (
          <year>2022</year>
          )
          <fpage>27</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Britzolakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kondylakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Papadakis</surname>
          </string-name>
          ,
          <article-title>A review on lexicon-based and machine learning political sentiment analysis using tweets</article-title>
          ,
          <source>International Journal of Semantic Computing</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>517</fpage>
          -
          <lpage>563</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kiesel</surname>
          </string-name>
          , Ç. Çöltekin,
          <string-name>
            <given-names>M.</given-names>
            <surname>Heinrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alshomary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. D.</given-names>
            <surname>Longueville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Erjavec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Handke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kopp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ljubešić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Meden</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mirzakhmedova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Morkevičius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Reitis-Munstermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Scharfbillig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stefanovitch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wachsmuth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          , Overview of Touché 2024:
          <article-title>Argumentation Systems</article-title>
          , in: L.
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mulhem</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Quénot</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Soulier</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. M. D. Nunzio</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Galuščáková</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF</source>
          <year>2024</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O.</given-names>
            <surname>Ahmad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Azim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharan</surname>
          </string-name>
          ,
          <article-title>Hybridizing sentence transformer embeddings with multi knn for improved multi-label text classification for pubmed (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>