<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging LLMs for Identifying types of Misinformation on Reddit</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bhavana Ramesh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Durwankur Gursale</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abram Jopaul</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marina Ernst</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universität Koblenz</institution>
          ,
          <addr-line>Universitätsstr. 1, 56070 Koblenz</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The recent development of LLMs has demonstrated their ability to generate coherent, contextually relevant responses to a variety of tasks. However, they also pose significant risks, including disinformation, factual inaccuracies and the propagation of bias. This research aims to access the efect of prompting techniques on the detection of diferent types of misinformation using data collected from Reddit. Our findings suggest that the Few Shot prompting method performs well across diferent LLMs. However, the efect of positional bias found in our experiments indicates that prompt engineering needs to be further investigated for such a sensitive task.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Misinformation</kwd>
        <kwd>Fake news</kwd>
        <kwd>Misinfrmation detection</kwd>
        <kwd>LLMs</kwd>
        <kwd>generative AI</kwd>
        <kwd>Reddit</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        When analyzing most available datasets, a common pattern emerges: they often use broad classifications
such as "true" and "false". Although these datasets often contain millions of rows and are well-developed,
they struggle to capture the nuanced nature of misinformation. However, some datasets, such as
NELA-GT-2018 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], LIAR [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and FakeNewsCorpus [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], ofer more detailed classifications. While some
datasets, such as NELA-GT-2018 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], consider for heterogeneous sources of information, others focus
on specific areas, such as politics and gossip. Thus, these datasets provide limited context due to their
restricted sources and classification types.
      </p>
      <p>The ability of LLMs to generate both reliable information and disinformation has been studied
extensively in recent years. Studies such as Jiang et al. [8] and Leite et al. [9] have investigated how
the LLMs perform under diferent prompting techniques to classify disinformation. These studies are
important for understanding the potential of LLMs to detect disinformation. However, these studies
focused on the broad classification of disinformation as true or false. Furthermore, as mentioned above,
these analyses used datasets that focused on a specific domain.</p>
      <p>
        Positional bias in LLMs refers to the tendency to rely on the positions of the tokens in the input, which
can adversely afect them, especially in the classification tasks [ 10], was a path breaking study that
introduced the transformer architecture. Transformer based LLMs like GPT, Llama etc. use positional
encoding which leads to the positional bias. Yu et al. [11] and Hsieh et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] explained the tendency of
LLMs to give priority to the tokens at the beginning and end of the input.
      </p>
      <p>Even though LLMs are thought to be powerful, Hu et al. [12] reveals that they fail to outperform
ifne-tuned small language models(SLM), when appropriate prompting techniques are used. Rather than
relying entirely on LLMs to detect disinformation, the authors propose a model in which LLMs act as
advisors to the fine-tuned SLMs. Like the studies mentioned above, this analysis also classifies data into
binary and focuses on GossipCop [13].</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>Most available datasets are built for standard binary classification rather than fine-grained classification,
and our bespoke dataset focuses on distinguishing each new instance into 4 categories. We briefly
describe each category:
• True: Indicates that the news item is true.
• Satire/Parody: Indicates if the content is twisted or misinterpreted in a satirical or humorous way.
• Misleading Content: Corresponds to the news where the information is intentionally manipulated
to fool the audience.</p>
        <p>• Imposter Content: Represents content generated by bots.</p>
        <p>While Wardle et al. [14] outline additional disinformation types (e.g., Fabricated Content, False
Connection), we excluded these as they rely on multimodal data. Our dataset was sourced from Reddit, a
platform known for its threaded discussions and popularity, ranking among the top 20 websites globally.
Using subreddits identified in Fakeddit [15], we obtained posts from July 2012 to December 2024. The
collected data is published at Zenodo1.</p>
        <p>Data collection utilized the pullpush.io API, extracting submission titles, text, metadata (e.g., score,
upvote ratio), and using tools such as Newspaper3k and BeautifulSoup4 for structured content scraping.
To ensure quality, subreddit moderators enforced relevance to themes, performing an initial filter.
We further refined the dataset by excluding posts with a score below 1, assuming that of-topic or
inappropriate posts would either be deleted or downvoted. This multi-level processing ensures the
credibility and relevance of our dataset. The final dataset contains 2343 posts with title, content, time,
misinformation category, subreddit, score and upvote ratio. The data is fairly balanced in terms of
misinformation category (Figure 1)</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Prompting</title>
        <p>Zero-shot prompting is used to classify the given content into predefined categories without any prior
training on specific examples. The model is provided with instructions to perform the classification
task based solely on its existing knowledge. An example of prompt is shown in Figure 2
Content categories are as follows: True, Satire, Imposter Content, Misleading
Content
Analyze the type of content and return the corresponding label.</p>
        <p>The chain-of-thought prompting was used to not only to classify the content, but also to explain the
reasoning process, for a possible manual evaluation. The model is guided to systematically evaluate
the given content by considering multiple factors before reaching a conclusion. An example of a CoT
prompt is shown in Figure 3.</p>
        <p>Analyze the type of the content enclosed in square brackets, and determine if it is
true content, satire, misleading content, or imposter content. Explain your
reasoning step by step and then return the answer as the corresponding content
label "true content" or "satire" or "misleading content" or "imposter content".</p>
        <p>Step-by-step reasoning:
1. Identify the primary purpose of the content (inform, entertain, deceive, etc
.).
2. Check for factual accuracy and sources.
3. Determine if there is any exaggeration, humor, or irony.
4. Check for any signs of manipulation or alteration.</p>
        <p>5. Determine if the content has been entirely fabricated.</p>
        <p>To address the challenge of positional bias, a binary classification was conducted, and each category
was labelled separately. The example of the prompt for True information case is shown on Figure 4.
Each class was isolated and classified individually to verify whether focusing on individual categories
could yield better accuracy than the combined classification.This approach helps to determine whether
the model favours certain types of content when labels are presented in diferent orders or contexts.
Given below the content category and definition:
True: True information.</p>
        <p>Analyze the content within square brackets and determine if it is true or not.
Return "True" if the content is true, otherwise return "None".</p>
        <p>To investigate positional bias in the Large Language Model (LLM), we explored all possible positional
arrangements of the four content classes: True, Satire, Misleading Content, and Imposter Content. This
involved changing the order of the classes in all 4! (24) possible ways.</p>
        <p>By classifying the content using these 24 diferent permutations, we aimed to determine if the LLM’s
performance varied based on the positional order of the categories, thereby revealing any potential
biases in their responses.
3.3. SLM vs LLM
This approach evaluates the performance of SLMs and LLMs over similar dataset. RoBERTa was
chosen as the SLM, and LlaMA (Llama-3-8B-Instruct), Gemma (Gemma-2-9b-it), and Mistral
(Mistral7B-Instruct-v0.3) were employed as the LLMs.</p>
        <p>There are many reasons why RoBERTa was chosen over other SLMs. To name a few, RoBERTa has
fewer restrictions on the data to be trained compared to the original BERT. RoBERTa is more focused on
the goal of modelling disguised speech. Since RoBERTa uses a dynamic masking strategy, it is renowned
for its robust and eficient generalization. Not to forget that RoBERTa is exceptionally flexible and can
be fine-tuned to specific tasks.</p>
        <p>Finally, after considering the pros and cons of using diferent models, we tested the dataset on both
SLM and LLM models without any fine-tuning to witness their out-of-the-box capabilities. Later,
finetuned RoBERTa was examined to see if larger datasets could improve its performance, and a comparison
with LLMs was performed to confirm if it could outperform.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In this section, we present the results obtained for each model. We report the recall, precision, and
F1 scores obtained by all the models for diferent prompting techniques. Comparing the accuracy of
various prompting techniques helps us find the best prompting approach. In addition, we are also
interested in the variety of biases exhibited by LLMs.</p>
      <p>Among several prompting strategies, Zero shot, Few Shot and Chain-of-Thoughts are selected to
test their performance. For each prompt accuracy, recall, precision, and F1 scores were recorded across
models. The results are presented in Table 1.</p>
      <p>Llama achieves an accuracy of 53% with Few Shot, significantly outperforming the other two
prompting techniques. Not only did Few Shot give the highest accuracy, but it also took all positions into
account when labelling. Overall, this made it the most eficient prompting technique.</p>
      <p>From an SLM perspective, the fine-tuned RoBERTa is run over diferent dataset sizes to record the
eficiency of how well it can classify the labels compared to LLMs. The graph on Figure 6 shows the
performance of RoBERTa with 500 and 1000 rows of data. It managed to achieve an accuracy of 23%
with the 500 row dataset and 30% with the 1000 row dataset. Surprisingly, there is a spike in accuracy
as the dataset size increases.</p>
      <sec id="sec-4-1">
        <title>4.1. Positional Bias</title>
        <p>When experimenting with the Zero Shot prompting technique by placing labels at diferent positions, it
is noticeable that not all positions receive a similar weight. This is due to the fact that LLMs are built
using a sequence-based architecture, which assigns disproportionate weights to certain tokens based
on their sequence in the input.</p>
        <p>The heat maps shown above in Figure 5 demonstrate how the accuracy varies with respect to position
when labels are called in sequence using the Zero Shot prompting technique. It is noteworthy that
Mistral outperforms all other models by showing less positional bias.</p>
        <p>To cross-validate the previous observation on positional bias, a binary classification was performed
using the prompt shown in Figure 4 was performed. The results of the binary classification, presented
in Table 2, show that the accuracy of all three models is generally higher when the prompt is focused
on detecting only one type of content. However, in this case, Imposter content was not identified at all,
which requires further investigation.</p>
        <p>The observed pattern, known as attention bias, occurs when either the beginning, the end, or the
middle of a sequence receives greater emphasis.The graph in Figure 7 demonstrates 2 shapes typical for
attention bias. Llama and Gemma depicts a U-shaped curve indicating the importance given to start
and end of the sequence, whereas Mistral presents an inverted U-shaped curve, indicating the attention
given to the middle of the sequence. In the case of Llama and Gemma, the attention bias is called as</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Performance by category</title>
        <p>Considering the F1-scores and accuracy of each label, Misleading content has an accuracy of 0.12 and an
F1 score of 0.21 for Llama (Table 3), accuracy of 0.03 and an F1 score of 0.06 for Mistral (Table 4). Thus,
by looking at the low F1 scores, we can say that the content does not cover all information needed by
LLM to categorize it as Misleading, making it the most deceptive disinformation category in this setting.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This research focuses on the classification of misinformation with LLMs and how prompting techniques
can influence the model’s decision making. The contribution of this study to the field starts with the
novel dataset of textual misinformation. The results of the experiments suggest that the structure
of the prompt has a significant impact on the performance of the models in detecting the type of
misinformation. While a Few shot prompt yields the best results in the context of this study, it is evident
that a correct prompt technique alone cannot ensure efective classification. Further investigation
revealed the presence of the positional bias, which skews the classification and prevalence of some types
over others. From the misinformation types perspective, Misleading and Imposter content appeared to
be the most dificult to identify. These findings open up many possibilities for further work. The first
possible direction is to run the experiment on a larger corpus of data and to include state-of-the-art
solutions, both commercial and open source. Another approach is to further investigate diferent
prompting methods.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT, DeepL Write in order to: Grammar
and spelling check, Paraphrase and reword, Improve writing style. After using this tool/service, the
author(s) reviewed and edited the content as needed and take(s) full responsibility for the publication’s
content.
Computational Linguistics, Florence, Italy, 2019, pp. 357–362. URL: https://aclanthology.org/
P19-2050/. doi:10.18653/v1/P19-2050.
[8] B. Jiang, Z. Tan, A. Nirmal, H. Liu, Disinformation detection: An evolving challenge in the age of
llms, in: Proceedings of the 2024 SIAM International Conference on Data Mining (SDM), SIAM,
2024, pp. 427–435.
[9] J. A. Leite, O. Razuvayevskaya, K. Bontcheva, C. Scarton, Weakly supervised veracity
classification with llm-predicted credibility signals, 2024. URL: https://arxiv.org/abs/2309.07601.
arXiv:2309.07601.
[10] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin,
Attention is all you need, in: Proceedings of the 31st International Conference on Neural Information
Processing Systems, NIPS’17, Curran Associates Inc., Red Hook, NY, USA, 2017, p. 6000–6010.
[11] Y. Yu, H. Jiang, X. Luo, Q. Wu, C.-Y. Lin, D. Li, Y. Yang, Y. Huang, L. Qiu, Mitigate position bias in
large language models via scaling a single dimension (2024). doi:10.48550/arXiv.2406.02536.
[12] B. Hu, Q. Sheng, J. Cao, Y. Shi, Y. Li, D. Wang, P. Qi, Bad actor, good advisor: exploring the role of
large language models in fake news detection, AAAI’24/IAAI’24/EAAI’24, AAAI Press, 2024. URL:
https://doi.org/10.1609/aaai.v38i20.30214. doi:10.1609/aaai.v38i20.30214.
[13] K. Shu, D. Mahudeswaran, S. Wang, D. Lee, H. Liu, Fakenewsnet: A data repository with news
content, social context, and spatiotemporal information for studying fake news on social media,
Big Data 8 (2020) 171–188. doi:10.1089/big.2020.0062.
[14] C. Wardle, H. Derakhshan, Information disorder: Toward an interdisciplinary framework for
research and policymaking, volume 27, Council of Europe Strasbourg, 2017.
[15] K. Nakamura, S. Levy, W. Y. Wang, Fakeddit: A new multimodal benchmark dataset for fine-grained
fake news detection, in: N. Calzolari, F. Béchet, P. Blache, K. Choukri, C. Cieri, T. Declerck, S. Goggi,
H. Isahara, B. Maegaard, J. Mariani, H. Mazo, A. Moreno, J. Odijk, S. Piperidis (Eds.), Proceedings
of the Twelfth Language Resources and Evaluation Conference, European Language Resources
Association, Marseille, France, 2020, pp. 6149–6157. URL: https://aclanthology.org/2020.lrec-1.755/.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>On the risk of misinformation pollution with large language models</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Singapore,
          <year>2023</year>
          , pp.
          <fpage>1389</fpage>
          -
          <lpage>1403</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-emnlp.
          <volume>97</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-emnlp.
          <volume>97</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.-Y.</given-names>
            <surname>Hsieh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-S.</given-names>
            <surname>Chuang</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-L. Li</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Glass</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ratner</surname>
            ,
            <given-names>C.-Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Krishna</surname>
          </string-name>
          , T. Pfister,
          <article-title>Found in the middle: Calibrating positional attention bias improves long context utilization</article-title>
          ,
          <year>2024</year>
          , pp.
          <fpage>14982</fpage>
          -
          <lpage>14995</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .findings-acl.
          <volume>890</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>M. G.</surname>
          </string-name>
          <article-title>Buchholz, Assessing the efectiveness of gpt-3 in detecting false political statements: A case study on the liar dataset</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2306.08190. arXiv:
          <volume>2306</volume>
          .
          <fpage>08190</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pelrine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Imouza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Thibault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reksoprodjo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Christoph</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Godbout</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rabbany</surname>
          </string-name>
          ,
          <article-title>Towards reliable misinformation mitigation: Generalization, uncertainty, and GPT-4</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>6399</fpage>
          -
          <lpage>6429</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .emnlp-main.
          <volume>395</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          . emnlp-main.
          <volume>395</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gruppi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. D.</given-names>
            <surname>Horne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Adalı</surname>
          </string-name>
          , Nela-gt
          <article-title>-2020: A large multi-labelled news dataset for the study of misinformation in news articles</article-title>
          ,
          <source>arXiv preprint arXiv:2102.04567</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          , “
          <article-title>liar, liar pants on fire”: A new benchmark dataset for fake news detection</article-title>
          , in: R. Barzilay, M.-Y. Kan (Eds.),
          <source>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>2</volume>
          :
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Vancouver, Canada,
          <year>2017</year>
          , pp.
          <fpage>422</fpage>
          -
          <lpage>426</lpage>
          . URL: https://aclanthology.org/P17-2067/. doi:
          <volume>10</volume>
          .18653/ v1/
          <fpage>P17</fpage>
          -2067.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Pathak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Srihari</surname>
          </string-name>
          ,
          <article-title>BREAKING! presenting fake news corpus for automated fact checking</article-title>
          , in: F.
          <string-name>
            <surname>Alva-Manchego</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Choi</surname>
          </string-name>
          , D. Khashabi (Eds.),
          <source>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop</source>
          , Association for
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>