<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Forum for Information Retrieval Evaluation, December</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Deciphering Vaccine Sentiments: Transformer Models in Social Media Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>M.S.Shriram</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. Karthika</string-name>
          <email>skarthika@ssn.edu.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Technology, Sri Sivasubramaniyanadar College of Engineering</institution>
          ,
          <addr-line>Kalavakkam, Chennai</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>1</volume>
      <fpage>5</fpage>
      <lpage>18</lpage>
      <abstract>
        <p>The Forum for Information Retrieval Evaluation, a platform emphasizing on data evaluation and analysis, has conducted a novel challenge in the NLP domain titled Artificial Intelligence on Social Media. The main idea behind this task is to formulate an efective multi-label classifier that labels social media posts according to specific concerns towards vaccines expressed by authors of the posts. The critical point here is that each post has multiple unlikely characters and multiple labels, this would ideally require models with better attention mechanisms for accurate classifications. Upon appropriate pre-processing, our team performed three diferent transformer based approaches that provided optimal results compared to other models with lesser attention span. The models used were BERT, RoBERTa and XLNet. The BERT and RoBERTa models gave the same macro-f1 score of 0.57 but had a mild diference with the Jaccard score with RoBERTa giving 0.57 and BERT giving 0.56. The XLNet model gave out a 0.48 macro-f1 score and 0.49 Jaccard score. Our research primarily aims to throw light on why these state-of-the-art transformer approaches prove to be deal breakers in modern day NLP.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>In more recent times there is no greater illustration of how crucial vaccinations are as a vital
weapon in fighting of diseases than how this has played out and the impact vaccines have had
on controlling the COVID-19 pandemic. While there are numerous advantages to vaccination,
there is still an extensive group of individuals who don’t trust in them. Such skepticism has
been fed by a number of issues including doubts over the safety and efectiveness of vaccines,
conspiracy theories about pharmaceutical companies and governments having hidden agendas.
These issues need to be understood and addressed in order to encourage mass vaccination and
secure the public health. In this paper, we propose an exploratory method to analyze social
media sentiment regarding vaccines in general. They aim at building an eficient multi-label
classification system for the task of identifying and labeling diferent types of issues raised
in tweets about vaccines. Our typology includes many issues such as concerns over vaccine
CEUR
Workshop
Proceedings
(S. Karthika)
safety and eficacy, concerns about government influence or manipulation, various forms of
conspiracies, and more. To accommodate the possibility that one tweet could address more than
one issue, we have built our classification framework in such a way that multiple labels can be
assigned to each tweet. We utilize cutting edge NLP models namely BERT, RoBERTa, XLNET
for the task, which can learn subtle linguistic features and context information with great
efectiveness. These tools allow us to get a pretty accurate reading of what people are saying
in tweet-form; it’s quite useful for understanding feelings and topics people are discussing
on social media.The goal of these studies is to advance knowledge about public attitudes and
receptivity towards immunizations to better develop public health information campaign and
communication strategies. With the power of social media data analytics and sophisticated
advanced NLP models at our disposal, we are attempting to illuminate what may contribute to
anti-vaccine attitudes so as to increase awareness and immunization rates across the globe</p>
    </sec>
    <sec id="sec-3">
      <title>2. Related Work</title>
      <p>
        This section provides an overview of research in developing architectures for comment
classification in the context of the COVID-19 pandemic’s impact on social media information flow.
The urgency of reliable classification methods for the surge in COVID-19-related tweets has
led to various approaches. One study [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] achieved remarkable F1 scores of 0.93 and 0.92 with
two machine learning approaches for classifying tweets into three categories. Another work
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] focused on sentiment analysis of COVID-19-related Twitter posts from March to mid-April
2020, utilizing seven deep learning models with F1 scores consistently exceeding 0.90. In [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
advanced machine learning techniques resulted in 95.35% accuracy for sentiment analysis and
91.49% for topic classification. COVID-19 vaccine sentiment on Twitter was analyzed in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ],
employing diverse data collection and preprocessing methods, along with classification using
BERT Transformers. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] introduced four deep learning models combining BERT with BiLSTM
and BiGRU algorithms, excelling over classical machine learning models. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] harnessed
Bidirectional Long Short-Term Memory (Bi-LSTM) to analyze COVID-19 sentiments from Twitter and
Reddit, surpassing conventional LSTM. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] introduced a deep learning sentiment analysis model
that achieved an accuracy of 78.062% on COVID-19-related tweets by focusing on sentence
embeddings. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] analyzed COVID-19 sentiment on Twitter, with BERT emerging as the most
accurate model. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] conducted sentiment analysis on Twitter discussions about COVID-19
vaccines, while [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] employed various sentiment analysis models, with BERT outperforming all.
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] conducted a comprehensive analysis of COVID-19 vaccine-related tweets, highlighting a
shift from initial negativity to a more positive outlook. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] investigated public perceptions of
COVID-19 vaccine adverse efects through social media data, with LSTM achieving the highest
accuracy. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] analyzed #sideefects social media conversations during early COVID-19 vaccine
rollout, with BERT outperforming other methods. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] examined over 200,000 COVID-19
vaccination tweets, achieving up to 90% accuracy in classifying misleading content. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] focused on
classifying COVID-19 vaccine-related tweets into negative, neutral, and positive sentiments,
with Bidirectional LSTM (BiLSTM) standing out with a 94.12% accuracy rate.A variety of
machine learning and deep learning models have been successful in capturing the subtleties of
public opinion during the pandemic, as shown by the most current work in this field, which
shows the extensive eforts made to assess sentiments in tweets related to COVID-19.
      </p>
    </sec>
    <sec id="sec-4">
      <title>3. Dataset</title>
      <p>
        The training dataset[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] provided by AiSoMe FIRE 2023[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] contains 9921 records about
antivaccine tweets about COVID vaccines that were posted during 2020-21. The dataset contains
tweet id tweet and labels.The important pattern to look into is that each tweet can be categorized
under multiple labels. This dataset encompasses 12 unique categories, each explicitly defined by
distinct concerns and sentiments regarding vaccines. These categories include ”Unnecessary,”
where tweets suggest vaccines are unnecessary or that alternate cures are superior; ”Mandatory,”
which pertains to objections against mandatory vaccination; ”Pharma,” focusing on skepticism
towards pharmaceutical companies and their profit motives; ”Conspiracy,” delving into deeper
vaccine-related conspiracies; ”Political,” addressing concerns about government or political
influences on vaccination; ”Country,” involving objections based on the country of origin of
vaccines; ”Rushed,” which expresses concerns about rushed or inadequately tested vaccine
processes; ”Ingredients,” centering around worries regarding vaccine components and technology;
”Side-efect,” focusing on concerns about vaccine side efects and associated deaths; ”Inefective,”
which questions vaccine efectiveness; ”Religious,” indicating objections to vaccines for religious
reasons; and finally, ”None,” representing tweets with no specified reason or concerns outside
of the defined categories. A test dataset[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] was also provided which contained 486 records
which were used for prediction by the models.
      </p>
    </sec>
    <sec id="sec-5">
      <title>4. Methodology</title>
      <p>Upon receiving the training dataset for the AISoMe track as part of the FIRE 2023 conference we
ifrst analyzed a few inconsistencies that the textual data had. To overcome such inconsistencies
and to use a clean form of data for the multi-label classifier model, we first used certain methods
to preprocess the data. In the provided dataset, each “post” started of with the usernames
being tagged. In social media tagging a user usually involves using the “@” symbol. So the
ifrst task involved removing the “@” symbol. Post this, we removed html tags, brackets and
emojis and converted the characters to lower cases. This was followed by stop word removal
and lemmatization. This preprocessing technique provided us with clean data that can be used
for further processing. Since there are multiple labels for a single post, this task can be
considered multi label classification. The processing of these labels involved splitting them up and
encoding the labels, using multi-label binarizer technique to the appropriate numerical format.
Upon understanding the attention mechanism of transformer based models, we proceeded to
incorporate three such models namely BERT, RoBERTa and XLNET for the classification task.
These models were chosen because of the self-attention mechanism that they possess which
enables it to weigh the significance of diferent elements within a sequence simultaneously, this
gives an edge over traditional models that relied on sequential data processing.
4.1. BERT
The first model devised for this task is the Bidirectional Encoder Representation from
Transformers, shortly known as BERT. It is a state of the art model used in a variety of NLP tasks but in
this particular case, we are formulating one for multi-label classification. Any text classification
model requires in depth context extraction/learning. BERT has this intricate feature called MLM
that helps it learn this concept more eficiently. MLM refers to a masked language model which
means, certain tokens in a text are intentionally masked or hidden, and the model’s task is to
predict or fill in the missing tokens. BERT is bidirectional which means it takes into account
both left and right contexts of the masked tokens in general. While using BERT for The input
text is tokenized and fed into the BERT model, and the final hidden state corresponding to the
[CLS] token (a special token added to the input) is used as the representation for the entire input
sequence.classification, we append a classification layer upon the pre-trained BERT model.The
output of the BERT model’s [CLS] token representation is passed through this classification
layer. The number of output units in the classification layer corresponds to the number of
classes in the classification task, for this task it is 12 output units representing 12 unique classes.
4.2. RoBERTa
RoBERTa, an extension of the BERT model, excels in various NLP tasks, including multi-label
classification. It utilizes a Masked Language Model (MLM) during pre-training, similar to BERT,
where certain tokens in the input text are concealed, and the model’s objective is to predict these
hidden tokens, taking both left and right contexts into account. The key diference between
BERT and RoBERTa here is in BERT architecture, the masking is performed once during data
preprocessing, resulting in a single static mask. To avoid using the single static mask, training
data is duplicated and masked 10 times, each time with a diferent mask strategy over 40 epochs
thus having 4 epochs with the same mask. This strategy is compared with dynamic masking in
which diferent masking is generated every time we pass data into the model.. In multi-label
classification, RoBERTa adapts by adding a classification layer atop its pre-trained layers. Input
text is tokenized, and the final hidden state associated with the [CLS] token is used as the
sequence representation. The number of output units in the classification layer matches the
number of classes in the classification task. During inference a sigmoid activation was used and
a threshold is applied to determine the assigned labels.
4.3. XLNET
Much similar to BERT and RoBERTa, XLNET also has a bidirectional approach for concept
learning but what sets XLNET apart is the autoregressive method that it employs. This helps
XLNET to capture contextual dependencies .Much like RoBERTa, XLNet employs a Masked
Language Model (MLM) during pre-training, akin to BERT, where specific tokens within the
input text are concealed, and the model’s primary objective is to predict these concealed tokens,
factoring in both left and right contexts. The pivotal distinction between BERT and XLNet lies
in the methodology of masking. While BERT employs a static mask, applied once during data
preprocessing, XLNet opts for a more dynamic approach. During training, the data is duplicated
and masked tenfold, with each instance employing a distinct masking strategy. This dynamic
masking strategy evolves with each pass through the model, enriching context comprehension.
In the realm of multi-label classification, XLNet adapts seamlessly by introducing a classification
layer atop its pre-trained layers. Text inputs are tokenized, and the ultimate hidden state
corresponding to the [CLS] token functions as the sequence representation. The number of
output units in our case is 12 mirroring the number of classes in the classification task. During
inference, sigmoid activation function was used, with a threshold applied to establish the
assigned labels similar to the other 2 models.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Evaluation and Results</title>
      <p>The training phase of our models was a meticulous process that involved fine-tuning several
critical hyperparameters, each with its unique role in shaping model performance. Weight
decay, represented by λ and set at 0.001, acted as a vital regularizer. It controlled the influence
of model weights on the loss function, efectively curbing overfitting by penalizing large weight
values. The learning rate, denoted as η and thoughtfully configured at 0.0001 or 0.00001,
played a pivotal role in determining the step size for weight updates during training. This
parameter significantly impacted convergence, with an appropriate learning rate ensuring that
the model efectively learned from the data without overshooting optimal parameter values.
Dropout rates, which ranged between 0.4 and 0.5, were judiciously chosen to regulate neuron
activations. Dropout is a regularization technique that prevents over-reliance on specific neurons
by randomly deactivating a fraction during each training iteration, promoting generalization.
Furthermore, the ”epochs” parameter governed the number of complete iterations through
the training data, allowing models to progressively learn and adapt over time. Finally, the
”maximum sequence length” parameter was set at 300, ensuring that input sequences were
appropriately processed during training, thereby enhancing the model’s overall ability to
achieve peak accuracy. These hyperparameters collectively contributed to the fine-tuning
process, facilitating the models’ optimization and ensuring their robust performance.Table 1
specifies the various hyperparameters and their corresponding values that gave maximum
training accuracy</p>
      <p>In the evaluation phase, a dedicated test dataset was provided to rigorously assess the
performance of the trained models. Each model underwent a single run, generating predictions
that were then submitted for evaluation. The evaluation process hinged on two pivotal metrics:
the Macro F1 Score and Jaccard Score.Together, these two reliable assessment criteria provided
reliable benchmarks for carefully assessing and contrasting the performance of the models
 1  = 2 ∗ (   ∗  )/(   +  )</p>
      <p>The Macro F1-Score is a metric used to evaluate the performance of a classification model. It
is calculated by taking the harmonic mean of precision and recall. The Macro F1-Score is useful
in multiclass classification problems where the classes are imbalanced. It gives equal weight to
each class, making it a good metric to use when the data is imbalanced.</p>
      <p>∑ F1 scores
Macro F1 Score = (2)</p>
      <p>Number of classes</p>
      <p>The Jaccard Index, also known as the Jaccard similarity coeficient, is a statistic used in
understanding the similarities between sample sets. The Jaccard Score is particularly valuable
for evaluating models dealing with binary or multiclass classification, as it doesn’t take into
account true negatives, focusing solely on the overlap between predicted and actual positive
instances.</p>
      <p>Jaccard( ,  ) =
| ∩  |
| ∪  |
(1)
(3)</p>
    </sec>
    <sec id="sec-7">
      <title>6. Conclusion</title>
      <p>This study harnessed advanced transformer-based models, BERT, RoBERTa, and XLNET, to
address the complex task of multi-label classification in the context of tweets expressing concerns
about vaccines, primarily comprising anti-vaccine sentiments. Notably, RoBERTa demonstrated
superior performance, followed closely by BERT and XLNET. In this predominantly anti-vaccine
dataset, these models showcased their prowess in discerning and categorizing various concerns.
To further enhance our model’s performance, future investigations could explore data
augmentation techniques, given the data-hungry nature of transformer-based models. Additionally,
fortifying model robustness through methods like adversarial training presents a promising
avenue. These eforts align with our mission of accurately categorizing concerns within the
domain of vaccine-related discussions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Anupam</given-names>
            <surname>Mondal</surname>
          </string-name>
          , Sainik Kumar Mahata , Monalisa Dey ,
          <string-name>
            <surname>Dipankar Das</surname>
          </string-name>
          , ”
          <article-title>Classification of COVID19 tweets using Machine Learning Approaches”,SMM4H 1 June 2021</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Vernikou</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lyras</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Kanavos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <article-title>Multiclass sentiment analysis on COVID-19-related tweets using deep learning models</article-title>
          .
          <source>Neural Comput &amp; Applic</source>
          <volume>34</volume>
          ,
          <fpage>19615</fpage>
          -
          <lpage>19627</lpage>
          (
          <year>2022</year>
          ). https://doi.org/10.1007/s00521-022-07650-2
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Gurkan</surname>
            , Caglar , Kozalioglu, Sude , Akdag, Ismail , Göçen, Cem , Palandoken,
            <given-names>Merih.</given-names>
          </string-name>
          (
          <year>2022</year>
          ). COVID-19
          <source>Related Tweets Classification and Sentiment Analysis Based on Machine Learning Approaches and Deep Learning Architecture Designs: A Comprehensive Analysis. 10.13140/RG.2.2.18200.47365/1.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Agustiningsih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Utami</surname>
          </string-name>
          and H. Al Fatta,
          <source>”Sentiment Analysis of COVID-19 Vaccine on Twitter Social Media: Systematic Literature Review,” 2021 IEEE 5th International Conference on Information Technology, Information Systems and Electrical Engineering (ICITISEE)</source>
          , Purwokerto, Indonesia,
          <year>2021</year>
          , pp.
          <fpage>121</fpage>
          -
          <lpage>126</lpage>
          , doi: 10.1109/ICITISEE53823.
          <year>2021</year>
          .
          <volume>9655960</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Talaat</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          <article-title>Sentiment analysis classification system using hybrid BERT models</article-title>
          .
          <source>J Big Data</source>
          <volume>10</volume>
          ,
          <issue>110</issue>
          (
          <year>2023</year>
          ). https://doi.org/10.1186/s40537-023-00781-w
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Arbane</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benlamri</surname>
            <given-names>R</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brik</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alahmar</surname>
            <given-names>AD</given-names>
          </string-name>
          .
          <article-title>Social media-based COVID-19 sentiment classification model using Bi-LSTM</article-title>
          .
          <source>Expert Syst Appl</source>
          .
          <year>2023</year>
          Feb;
          <volume>212</volume>
          :
          <fpage>118710</fpage>
          . doi:
          <volume>10</volume>
          .1016/j.eswa.
          <year>2022</year>
          .
          <volume>118710</volume>
          . Epub 2022 Aug 30. PMID: 36060151; PMCID:
          <fpage>PMC9425711</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Fattoh</surname>
            <given-names>IE</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kamal Alsheref</surname>
            <given-names>F</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ead</surname>
            <given-names>WM</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Youssef</surname>
            <given-names>AM</given-names>
          </string-name>
          .
          <article-title>Semantic Sentiment Classification for COVID-19 Tweets Using Universal Sentence Encoder</article-title>
          .
          <source>Comput Intell Neurosci. 2022 Oct</source>
          <volume>5</volume>
          ;
          <year>2022</year>
          :6354543. doi:
          <volume>10</volume>
          .1155/
          <year>2022</year>
          /6354543. PMID: 36248924; PMCID:
          <fpage>PMC9556213</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>U. N.</given-names>
            <surname>Wisesty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rismala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Munggana</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Purwarianti</surname>
          </string-name>
          , ”Comparative Study of Covid-19
          <source>Tweets Sentiment Classification Methods,” 2021 9th International Conference on Information and Communication Technology (ICoICT)</source>
          , Yogyakarta, Indonesia,
          <year>2021</year>
          , pp.
          <fpage>588</fpage>
          -
          <lpage>593</lpage>
          , doi: 10.1109/ICoICT52021.
          <year>2021</year>
          .
          <volume>9527533</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Philip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Thatha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Harshini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. V. S. L.</given-names>
            <surname>Haritha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Patil</surname>
          </string-name>
          and
          <string-name>
            <given-names>B. Veerasekhar</given-names>
            <surname>Reddy</surname>
          </string-name>
          , ”
          <article-title>Classification of Covid-19 Vaccines tweets using Naïve Bayes Classification</article-title>
          ,”
          <year>2022</year>
          6th International Conference on Electronics, Communication and
          <string-name>
            <given-names>Aerospace</given-names>
            <surname>Technology</surname>
          </string-name>
          , Coimbatore, India,
          <year>2022</year>
          , pp.
          <fpage>1384</fpage>
          -
          <lpage>1387</lpage>
          , doi: 10.1109/ICECA55336.
          <year>2022</year>
          .
          <volume>10009511</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>V.</given-names>
            <surname>Battula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. G.</given-names>
            <surname>Goli</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Nasigari</surname>
          </string-name>
          , ”
          <article-title>Identification of Optimal Model for Multi-Class Classification of COVID Tweets</article-title>
          ,”
          <source>2022 9th International Conference on Computing for Sustainable Global Development (INDIACom)</source>
          , New Delhi, India,
          <year>2022</year>
          , pp.
          <fpage>495</fpage>
          -
          <lpage>499</lpage>
          , doi: 10.23919/INDIACom54597.
          <year>2022</year>
          .
          <volume>9763291</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gottipati</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Guha</surname>
          </string-name>
          , ”Analysing Tweets on COVID-19
          <string-name>
            <surname>Vaccine: A Text Mining</surname>
            <given-names>Approach</given-names>
          </string-name>
          ,”
          <source>2022 IEEE 12th Annual Computing and Communication Workshop and Conference (CCWC)</source>
          ,
          <source>Las Vegas</source>
          ,
          <string-name>
            <surname>NV</surname>
          </string-name>
          , USA,
          <year>2022</year>
          , pp.
          <fpage>0467</fpage>
          -
          <lpage>0474</lpage>
          , doi: 10.1109/CCWC54503.
          <year>2022</year>
          .
          <volume>9720793</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K. R. S. N.</given-names>
            <surname>Kariyapperuma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Banujan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. M. A. K.</given-names>
            <surname>Wijeratna</surname>
          </string-name>
          and
          <string-name>
            <given-names>B. T. G. S.</given-names>
            <surname>Kumara</surname>
          </string-name>
          , ”
          <article-title>Classification of Covid19 Vaccine-Related Tweets Using Deep Learning,” 2022 International Conference on Data Analytics for Business and Industry (ICDABI), Sakhir</article-title>
          , Bahrain,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          , doi: 10.1109/ICDABI56818.
          <year>2022</year>
          .
          <volume>10041615</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Manjrekar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>McConnell</surname>
          </string-name>
          and
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Gokhale</surname>
          </string-name>
          , ”
          <source>Analyzing Twitter Conversations on Side Efects of Covid-19 Vaccine,” 2022 2nd International Conference on Intelligent Technologies (CONIT)</source>
          , Hubli, India,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          , doi: 10.1109/CONIT55038.
          <year>2022</year>
          .
          <volume>9848134</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sharma</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Datta</surname>
          </string-name>
          , ”
          <article-title>(Mis)leading the COVID-19 Vaccination Discourse on Twitter: An Exploratory Study of Infodemic Around the Pandemic,” in IEEE Transactions on Computational Social Systems</article-title>
          , doi: 10.1109/TCSS.
          <year>2022</year>
          .
          <volume>3225216</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mansouri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Soui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Alhassan</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Abed</surname>
          </string-name>
          , ”
          <article-title>TextBlob and BiLSTM for Sentiment analysis toward COVID-19 vaccines</article-title>
          ,”
          <source>2022 7th International Conference on Data Science and Machine Learning Applications (CDMA)</source>
          , Riyadh, Saudi Arabia,
          <year>2022</year>
          , pp.
          <fpage>73</fpage>
          -
          <lpage>78</lpage>
          , doi: 10.1109/CDMA54072.
          <year>2022</year>
          .
          <volume>00017</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Soham</surname>
            <given-names>Poddar</given-names>
          </string-name>
          , Azlaan Mustafa Samad, Rajdeep Mukherjee, Niloy Ganguly, and
          <string-name>
            <given-names>Saptarshi</given-names>
            <surname>Ghosh</surname>
          </string-name>
          .
          <year>2022</year>
          .
          <article-title>CAVES: A Dataset to facilitate Explainable Classification and Summarization of Concerns towards COVID Vaccines</article-title>
          .
          <source>In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '22)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>3154</fpage>
          -
          <lpage>3164</lpage>
          . https://- doi.org/10.1145/3477495.3531745
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Soham</surname>
            <given-names>Poddar</given-names>
          </string-name>
          , Moumita Basu, Kripabandhu, and Saptarshi Ghosh. ”
          <source>Overview of the FIRE 2023 Track: Artificial Intelligence on Social Media (AISoMe).” In Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>