<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Information System for the Fact-checker Support</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yurii Baryshev</string-name>
          <email>yuriy.baryshev@vntu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Leonid Kupershtein</string-name>
          <email>kupershtein.lm@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Maidanovych</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olesia Voitovych</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Semen Kuznets Kharkiv National University of Economics</institution>
          ,
          <addr-line>Nauky Ave., 9-A, Kharkiv, 61166</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Vinnytsia National Technical University</institution>
          ,
          <addr-line>Khmelnytske shoes str., 95, Vinnytsia, 21021</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>127</fpage>
      <lpage>138</lpage>
      <abstract>
        <p>The research results of fact-checking automation are presented. The main types of fact-checking were described and the methods and techniques used by fact-checkers were analyzed. Particular attention was paid to the development of the system architecture, which consists of various modules, including a visualization module, database search, internet search, analysis and verification of news on websites and search in telegram channels. Algorithms were developed to work with these modules, allowing for efficient fact-checking. The software architecture was designed and experimental application was developed.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Fact-checking</kwd>
        <kwd>automation</kwd>
        <kwd>fake news</kwd>
        <kwd>information attack</kwd>
        <kwd>cybersecurity</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In a world where inaccurate news can spread quickly and have harmful impact, the process of
factchecking is becoming vital and increasingly topical. It is a key tool in the fight against so-called fake
news and disinformation. Fact-checking is not only a way to identify false statements but also a means
of restoring the trust to the media. However, the process of fact-checking can be labor-intensive and
time-consuming, especially when dealing with large amounts of information. Fact-checkers often have
to spend hours, and sometimes days, searching for evidence to confirm or refute certain facts. In
addition, the verification process itself can be challenging, as it requires not only extensive knowledge
but also the ability to think critically and analyze information. In this context, the development of an
automated fact-checking tool can be not only a significant contribution to the field but also a real asset
for journalists, fact-checkers, and anyone working in the informational and media sectors. It is believed
that these kind of tools can greatly simplify the fact-checking process, reduce the time required for it,
and increase the efficiency and accuracy of the process.</p>
      <p>The goal of our work is to increase the efficiency of the fact-checking process by automating the
search, collection and analysis of information using modern information intelligent technologies.</p>
      <p>In the modern world, inundated with vast amounts of information, it is imperative that we have
confidence in the accuracy of the information we consume. In this context, this work is highly relevant,
as it aims to develop effective tools to assist in verifying the authenticity of information. Future research
in this field could greatly enhance capacity to identify and counteract fake news and disinformation.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Analysis of Fact-Checking</title>
      <p>
        Fact-checking is a modern tool for journalistic investigations with a unique concept and technical
features aimed at exposing populism, manipulations, and false information in the rhetoric of speakers
at various levels and in media publications [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This procedure is crucial in the media industry, where
journalists, editors, and other professionals are required to effectively verify information before
disclosing it. Fact-checking encompasses the verification of aspects such as factual data, quotes, sources
of information, and other content.
      </p>
      <p>
        It is important to note that fact-checking is not only about identifying false information, but also
about the ability to recognize and identify subjectivity and hidden aspects, as well as to assess the quality
of information in the context. Professional journalism is distinguished by the usage of reliable data and
facts verified by all possible means [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        This process also plays a significant role in cybersecurity. Disinformation and information
manipulation are increasingly prevalent in the modern world, posing threats in various contexts,
including wars and conflicts. Military conflicts involve not only physical combat but also the usage of
information warfare, such as disinformation and manipulation, to achieve objectives and influence
public opinion [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Fact-checking, as an investigative trend, swiftly entered Ukrainian media landscape.
This was primarily in response to the wave of anti-Ukrainian propaganda that rapidly flooded
information channels with fake and blatantly false reports [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>Ukraine serves as an example of information warfare as an integral component of a military conflict.
Such disinformation campaigns and information manipulation aim to sow discord, foster distrust, and
create social tensions. Their consequences can be severe, impacting political stability, public safety,
and the economic situation in the country. Furthermore, these information attacks can have an
international impact, altering perceptions and support from allied countries and influencing political
relations. To ensure cybersecurity and defend against such information threats, it is crucial to actively
implement fact-checking methods and counter the disinformation. The development and utilization of
tools for detecting fake news, promoting critical thinking and information literacy, and facilitating
access to reliable information can mitigate the impact of disinformation campaigns and enhance
cybersecurity in the context of military conflicts and other challenges in the modern world.</p>
      <p>
        For instance, in the initial days of the full-scale invasion of Ukraine by the russian federation, the
Gvara Media team launched the fact-checking telegram-bot "PEREVIRKA". This bot assists in
distinguishing fake news from real news and recognizing manipulation and propaganda. Upon receiving
news submissions, the information is forwarded to project volunteers who manually verify its
truthfulness and accuracy. In just the first month of the invasion, they processed approximately 30000
requests [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        In Ukraine, the concept of fact-checking is currently represented by several resources:
 VoxUkraine: This is a fact-checking project of the independent analytical platform Vox Ukraine.
The team exposes lies, manipulations, and russian propaganda, both in Ukraine and abroad. Since 2018,
they have been signatories to the Code of Ethics of the Poynter Institute's International Network of
Factcheckers, and since 2020 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
         Word and Deed: This resource checks the election promises of politicians [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
         Stopfake.org: Their primary activity is aimed at combating anti-Ukrainian propaganda, statements,
and facts aimed at discrediting Ukraine [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
         Gvara Media: An independent online publication focused on social change [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        For example of what fake news can do is to look at a recent event. Realistic images of an explosion
near the Pentagon created by artificial intelligence that went viral on Twitter caused the stock market
to briefly plummet. Tweets with images purportedly depicting an explosion near the Pentagon building
in Arlington, Virginia, were posted by many verified Twitter accounts, including a russian state media
outlet with millions of followers and a verified account claiming to be the Bloomberg news agency.
Although at first glance the viral photos appeared to be genuine, they are full of hints that they were
created with the help of artificial intelligence, which proves that it is all a hoax [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        A similar event happened last year. On Nov. 10, a verified Twitter account posed as Eli Lilly wrote
in a viral tweet, "We are excited to announce insulin is free now." By the next day, the drugmakers’
shares plunged by about $22 billion [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods and Techniques of Fact-Checking</title>
      <p>Fact-checking, as a process of verifying the truthfulness of information, includes a number of
methods and techniques that help to carry out this work effectively and accurately.</p>
      <p>These methods include:
 Checking the reliability of sources: A fact-checker determines the reliability, reputation and
competence of the source of information, as well as the ability to confirm or refute the information
presented.</p>
      <p> Comparison of data: Fact-checkers analyze information from different sources and compare it
with known facts or data to detect discrepancies or inaccuracies.</p>
      <p> Checking the context: Fact-checkers analyze the circumstances in which the information was
provided and check for distortions or deletions that could change its meaning.</p>
      <p> Involving experts: Fact-checkers can call on experts in the relevant field to verify specific facts
or data, using their knowledge and experience.</p>
      <p> Objectivity of experts: Experts involved in fact-checking should be objective and free from
conflicts of interest. Their experience and reputation, as well as their qualifications, should be
thoroughly checked.</p>
      <p> Use of verified data: It is important to use reliable sources of information, such as official
statistics and scientific studies, to compare and corroborate information.</p>
      <p> Technologies for fact-checking: Modern technologies, such as artificial intelligence, machine
learning and data analytics, can automate the fact-checking process, increasing its speed and accuracy.
They help to identify contradictory statements and manipulation of information.</p>
      <p>
        To check the authenticity of content on the Internet special software, applications, and browser
plugins (Trooclick, Genius, factcheck.org, and others) are used [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Such services are able to analyze facts
and identify not only fake news but also manipulative content, identify inaccuracies or biases in the
presentation of the material. They scan selected text and check it using official statistics, media reports,
their own database, etc. Some of the services are able to analyze headlines, text complexity, popularity,
and the structure of the site's code.
      </p>
      <p>Additionally, there are also services that can aid verification process of multimedia content such as
photos and videos.</p>
      <p>
        Here are some of the services for photos verification:
 Google.Photo allows to upload a photo and find similar or identical images. Results can be
sorted by size, upload date, and image format [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
         Google.Lens works in a similar way to Google.Photo, but allows to search for individual parts
of an image, such as buildings in a photo [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
         TinEye - this service allows to find different versions of images, including modified versions,
such as cropped, retouched or with added elements. Results can be filtered by relevance, size or
differences from the original [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
         EXIF Data/Jeffrey's Image Metadata Viewer - these tools allow to get technical information
about a photo, such as the date it was taken, the location, and the device it was taken with. But, this data
may not be available for some photos that have been uploaded and re-uploaded multiple times [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
         FotoForensics - this service analyses images using the ELA (error level analysis) method. It
compares the image with its compressed version. It helps to identify which part of the image has been
altered, whether it is darker or brighter compared to a similar area that has remained unchanged [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>These tools help fact-checkers detect photographic manipulation and verify the authenticity of
images.</p>
      <p>Fact-checking technologies are powerful tools, but they cannot replace the role of a fact-checker.
The manual element remains an integral part of the process, as context, cultural differences and
subjectivity of information need to be taken into consideration.</p>
      <p>The efficient usage of technology, combined with expert analysis, can lead to more objective and
reliable results in fact-checking.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Problem Statement</title>
      <p>Fact-checking is a complex and intensive process that requires a significant amount of time and
effort to verify the accuracy of facts, news, and information disseminated in the media and online
sources. Given the vast volume of information available, manually verifying all data is highly
resourceintensive in terms of time and resources.</p>
      <p>That's why the automation of fact-checking processes is becoming increasingly important. The usage
of automated algorithms and software solutions allows for the efficient analysis of large amounts of
data, rapid verification of facts, and the identification of potential attempts at disinformation and
information manipulation. This, in turn, enhances the speed and accuracy of the fact-checking process,
enabling a quicker response to the spread of false information and preventing its adverse consequences.</p>
      <p>The primary objective of our work is to develop an information system based on software tool to
support the work of fact-checkers, with the goal of providing a convenient and effective means of
detecting fake news.</p>
      <p>Functionality Requirements:
 Identification of fake news: The tool should be capable of analyzing news content to determine
its credibility. This may involve analyzing the text, source, author, and other relevant attributes of the
news.</p>
      <p> Search for similar news: The tool should have the capability to search for similar news stories
within a database or Internet. This feature will assist fact-checkers in identifying instances of news
republishing or the spread of fake news.</p>
      <p> Parsing comments: The tool should be able to collect comments on news articles from various
sources, social media, and other platforms. This will enable fact-checkers to analyze public reactions
and opinions regarding a specific news item.</p>
      <p> Analysis of publication dates and authorship: The tool should be able to analyze the publication
date and author information. This functionality will help fact-checkers assess the relevance of the news
and evaluate the credibility of the author.</p>
      <p>Interface Requirements:
 User-Friendly Interface: The tool should feature an intuitive and user-friendly interface that
enables fact-checkers to use it effortlessly and access the necessary information.</p>
      <p> Visualization of Results: The tool should be capable of visually presenting the results of news
analysis. This visual representation will allow fact-checkers to quickly assess the credibility and
dissemination of fake news.</p>
      <p>Performance Requirements:
 Processing Speed: The tool should have the capability to process a substantial volume of news
rapidly, enabling fact-checkers to promptly identify and respond to fake news.</p>
      <p> Efficient Resource Utilization: The tool should optimize the utilization of resources, including
computing power and memory, to ensure the efficiency of the fact-checkers work.</p>
      <p>The tool should include features such as searching for similar news within the database, searching
for news on Telegram channels, parsing comments, identifying fake news, analyzing publication dates
and authorship. Additionally, the tool should offer a user-friendly interface, the ability to visualize
results, and efficient performance.</p>
      <p>So, we have discussed the concept of fact-checking and its significance in the modern information
environment. We have also described the primary types of fact-checking. Following an analysis of the
methods and techniques employed by fact-checkers, we have identified the key steps in the
factchecking process, including fact verification, source and citation checks, context analysis, and the
utilization of specialized tools and data.</p>
      <p>Comprehending these methods and techniques is essential for the effective execution of
factchecking tasks. The study has pinpointed the principal challenges associated with fact-checking, such
as politicization and the impact of social media.</p>
      <p>The development of such a tool will help to increase the efficiency and accuracy of information
verification, as well as help to provide reliable and accurate information to a wide audience by
automating the fact-checking process.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Information system architecture</title>
      <p>The information system architecture includes several modules. It is shown in Figure 1. It consists of
six modules: visualization module, database-search module, telegram-search module, network-search
module, analysis and verification module and data-save module. System has its own database also.</p>
      <p>Let's consider each module separately. The visualization module is responsible for launching the
system and presenting the graphical interface to an user. Upon launching the application, the module
initializes graphical components such as windows, icons, images, and control buttons for user
interaction with the tool. The module handles the display of windows in response to user actions. It also
establishes the necessary connections and displays the system's results. After data entry, the module
passes this information to the subsequent processing stages.</p>
      <p>The database search module searches and compares the entered data with the existing database of
fake and true news. It utilizes algorithms and methods to identify matches and determine the credibility
of the news. If the news is found in the database, the component returns the corresponding verification
result. However, frequently incoming news may not be unique, as different information channels, in
various formats, cover certain events even though they pertain to the same event. Consequently, the
same facts can be processed multiple times. Machine-based comparison of the received message with
ones stored in the database enables the automatic processing of similar requests. The task of comparing
and determining text similarity is relatively straightforward. Currently, there are many tools available
for implementing this.</p>
      <p>
        One of the most popular methods of converting text into a numerical vector is to use vector
embedding. A vector embedding is a numerical representation of a word or phrase that preserves the
semantic relationships between words [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. For example, words with similar meanings will have similar
vector representations.
      </p>
      <p>
        One of the simplest approaches is the bag-of-words approach. In this approach, the text is considered
as a set of individual words, and a vector is created that shows which words are present in the text and
how often they occur. Each word is considered a separate feature, and its number in the text is the value
of this feature [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>Another approach is TF-IDF (Term Frequency-Inverse Document Frequency), which is a statistical
method for assessing the importance of a word in a document by taking into account its frequency in
the document and weighing the frequency of the word in other documents. The TF-IDF is calculated
for each word in a corpus of documents and gives a numerical value that reflects its relevance to each
individual document.</p>
      <p>
        The TF-IDF is often used to represent documents in a collection as numerical vectors those reflect
the importance of each word in a set of words (the number of words in the set determines the dimension
of the vector) in each document. Such model is called a vector model and makes it possible to compare
texts by comparing their respective vectors using given metric (Euclidean distance, cosine measure,
Manhattan distance, Chebyshev distance, etc.), i.e. to perform a cluster analysis [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        Next is the Word2Vec model, which learns to predict words based on their context. It creates vector
representations of words in which similar words have similar vectors [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>
        Another one is the GloVe model (Global Vectors for Word Representation), which combines
statistics of word occurrences in texts with matrix decomposition [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>
        The FastText model uses the Word2Vec approach, but treats words as a set of character substrings.
It allows to obtain vector representations even for words that were not included in the training corpus
[
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. There is also the Universal Sentence Encoder Multilingual model from Google. It is a powerful
tool for converting text sentences into vector representations (embeddings). The main goal of the model
is to preserve the semantic relations between sentences in the vector space [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
      </p>
      <p>For our work, we have chosen the Universal Sentence Encoder Multilingual model, which is based
on transformers. It provides a convenient and powerful way to convert text into numerical vectors. Its
multilingual support allows to work with text in different languages, which is important for analyzing
news and content from different sources. In addition, the model has good performance and high
accuracy, which makes it an ideal choice for our fact-checking and news content analysis tasks.</p>
      <p>The primary steps for implementing machine fuzzy text matching are:
 Cleaning of stop words: prepositions, conjunctions, pronouns.
 Cleaning of invalid data: links, etc.</p>
      <p> Conversion to a numeric vector format: at this stage, a trained neural network model processes
the text and transforms it into a numeric vector array, for example:
““Окупанти застосували хімічну зброю на
[[ 0.00342214 -0.00027666 -0.02173961 ... -0.01806361 0.02199496
[-0.06651231 -0.08472251 0.0423876 ... -0.02535929 -0.01699802
[-0.04076068 -0.093428 -0.02202856 ... -0.0024216 -0.00762561
...
[ 0.04959109 -0.0801333 -0.06416482 ... -0.02055199 0.0410036 -0.04278645]
[-0.03331999 -0.09735778 0.00543897 ... 0.03228471 0.01209306 -0.04241692]
[-0.0637934 -0.08842427 -0.02345617 ... 0.04337208 -0.03944801 0.04867269]]
Our array has 512 elements. Each element is represented by a floating point number.</p>
      <p>Our program utilizes a database that is sourced from a dump of the "ПЕРЕВІРКА" Telegram bot.
This database serves as the foundation for data retrieval and operations within our application.</p>
      <p>The resulting vector arrays can be easily compared using the cosine similarity measure and the
percentage of similarity between two texts can be determined.</p>
      <p>The database structure contains the News table with the following fields:
 image: A field containing a link to the image associated with the news item. This allows you to
store one or more images related to a particular news item.</p>
      <p> video: A field containing a link to a video associated with the news item. It allows you to store
one or more videos related to a particular news item.</p>
      <p> fakeStatus: A Boolean field that indicates whether the news is fake. A value of "true" indicates
a true news item, and a value of "false" indicates a fake news item.</p>
      <p> createdAt: A field containing the date and time the news was created. This allows you to track
when a particular news item was added to the database.</p>
      <p> text: A text field that contains the text of the news item. It stores the main content of the news
item, which will be used for further analysis and verification.</p>
      <p> id: A unique identifier for the news item. This field allows you to uniquely identify and access
a specific news item in the database.</p>
      <p>This database structure allows to store information about news, including images, videos, fake news
status, date of creation and the news text itself. A unique identifier helps to unify and uniquely identify
each news item in the system.</p>
      <p>The Telegram search module allows to automatically find news in the channels and groups of this
platform that have similar content to the entered news text. This module provides an opportunity to
effectively analyze and monitor news content, receive up-to-date information from a large number of
sources and identify news that have a similar context or topic. Usage of the Telegram search module
allows to quickly and conveniently review news from different sources in one place, as well as analyze
their similarities and identify related topics or events. This will help users to track up-to-date
information, conduct research, and draw informed conclusions based on a wide range of sources.</p>
      <p>The Internet search module allows users to search for news that was not found in the database. It
queries a variety of information sources, including news websites, social media and other sources, to
find related information about a news item. The module displays a list of URLs of web pages where the
news is mentioned and the date of its publication. It also provides special buttons next to each result,
which are needed for further analysis and verification of the news.</p>
      <p>The analysis and verification module uses special algorithms and methods to assess the reliability of
the news. It performs text processing, detection of suspicious patterns, context analysis and other
techniques to determine whether a news item is fake or not. The verification result is presented in an
appropriate way, allowing the user to understand how reliable the news is.</p>
      <p>The general algorithm for checking news for authenticity can be summarized as follows:
Step 1. Input data: Receiving the news for analysis and verification.</p>
      <p>Step 2. Pre-processing:
 Parsing data from the web page
 Identification of the author.
 Parsing the publication date of the news.</p>
      <p> If supported by the website, parsing of comments.</p>
      <p>Step 3. Context analysis:
 Analyzing the context of the news, including keywords, topics and other parameters.
 Identify suspicious patterns or incorrect information.
 Check whether links to other sources are present. Each news item should be linked to a
source.</p>
      <p>Step 4. Language and style analysis:
 Determining the language of the news text and checking compliance with language rules.
 Take into account the style and phrases used that may indicate fake news.</p>
      <p>Step 5. Checking the source:
 Verifying the news from other trusted sources.
 Checking the credibility and reliability of the news source.</p>
      <p> Checking whether the source is satirical and the news itself is a joke.</p>
      <p>Step 6. Using of external sources:
 Querying external sources or databases for more information about the news.</p>
      <p> Compare the results and confirm or deny the facts of the news.</p>
      <p>Step 7. Yielding a result:
 Based on the collected information and analysis, determining whether the news is fake.
 Displaying the result, which indicates whether the news is authentic or fake.</p>
      <p>Step 8. Output data: Receiving the result of the fake news checking.</p>
      <p>This algorithm describes the main steps and procedures performed in the News Analysis and
Verification module. The implementation of specific algorithms and methods may vary depending on
the needs. In our software, the algorithm of the analysis and verification module is based on rules and
weights. First of all, we parse the elements of the web page where the news is placed, such as the title,
author, date of publication, and author. Next, we use rules to check the authenticity of the news. For
example, we check whether the source of the news is unreliable: if true, it is set to 1, otherwise to 0. We
also check whether the news text is typed: if there are typos, it is set to 1, otherwise to 0. Additionally,
we check for manipulative words in the news text: if they are found, it is set to 1, otherwise to 0. Also,
if there is no author of the news, it is set to 1, otherwise to 0. We use weights for each rule. In our case,
each rule has a weight of 0.25. Finally, we calculate the newsworthiness score by calculating the
weighted sum of the rule values.</p>
      <p>The value is calculated using the formula:
fake_news_score = (source weight · source rule + text weight · text rule + manipulation weight·
· manipulation rule + author weight · author rule).</p>
      <p>The data storage module is responsible for saving news and verification results in the database. It
provides efficient storage and management of the collected data, ensuring its security and availability.
This component may use a database or other storage mechanisms to ensure efficient access to data and
to ensure data integrity and confidentiality. Each of these modules interacts with the others, transmitting
and receiving the necessary data for the system to operate efficiently. This interaction ensures the
integration of functionality and improves the system's ability to verify the accuracy of news. We have
utilized the capabilities of the Python programming language and the JetBrains PyCharm development
environment to construct our system. While the program's user interface was created using the
customtkinter library, we have also leveraged a host of other powerful tools to extend its functionality.</p>
      <p>For the transformation of embeddings, we have turned to the renowned TensorFlow library, a
stalwart in the field of machine learning and deep neural networks. Its extensive capabilities are
instrumental in converting textual data into numerical vectors, thereby enabling effective comparison
and similarity measurement. Additionally, for calculating similarity metrics, we have incorporated the
sklearn library. This dynamic library equips our system with the ability to assess the similarity between
text vectors, employing various metrics like cosine similarity, Euclidean distance, and more.</p>
      <p>In the realm of communication with Telegram, we have utilized Telethon, a Python library that
facilitates interaction with the Telegram API. This integration enables our system to access and search
Telegram channels, expanding the horizons of information retrieval. When parsing news from Internet
content, our system use the Google API Client library. This tool enhances the capacity of our system to
conduct web searches and retrieve relevant data. These integrated tools and libraries fortify the core of
our system, allowing it to function as a comprehensive fact-checking and news analysis platform.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Experiments</title>
      <p>To launch the program, we need to download the software and run it. The user opens the program
window on the "Check by database" tab. It is shown in Figure 2.</p>
      <p>On the side of the program is a window with controls:
 "Database check" screen;
 "Telegram check" screen;
 "Checking the network";
 element for changing the application theme;
 element for changing the interface scale.</p>
      <p>Let’s test the application's intaraction with databases. We already have a ready-made database of
fake news. Let's create an array of NumPy vectors for it. To do this, one is to press the "Select File"
button and then "Convert". After that, the application asks where to save the file. Once the file is
successfully saved, one can select it using the Select File button to continue working. Select database
and an array of vectors. Enter the news to be checked in the text box. Enter the number of most similar
results those user wants to display. Press the "Search" button. The result is shown in Figure 3.</p>
      <p>Let’s test the developed application with Telegram. We have a ready-made database of links to
Telegram channels. Since the algorithm works using the Telegram API, to get started, one needs to get
the API ID and API HASH values and enter them into the special fields of application. This procedure
needs to be done only once, and these values will be saved in the future.</p>
      <p>After entering the API ID and API HASH values, one needs to enter the news we need, select file
with the channel database, and enter the date for which application should search for the news. The
result is shown in Figure 4. After pressing the "News" button, user will see all similar posts from all
channels. As a result, program outputs the link to the channel, news text, link to the post and date of
publication. The result is shown in Figure 5. Let's test the parsing of comments. Since we have searched
for news by channel, and the results contain links to posts, let's take one of them.</p>
      <p>To start parsing, one needs to paste the link into a special field, and if necessary, select black or
white list filtering, and click the "Comments" button. The result of parsing can be seen in Figure 6. Let's
try to perform the same action, but with the whitelist filter enabled. The result is shown in Figure 7.</p>
      <p>Let's test the search for news on the web. To do this, one is to enter the news in the search field and
press the "Search" button. The search results will be displayed in the text box. The result is shown in
Figure 8. As it can be seen from fig. 8 the results include the headline of the news and the link to it.
Next, we can test the news verification on the web page. Since the search returned certain results, we
can take the link from there. Let’s paste the selected link into a special field and press the "Check. The
verification process yield the result shown in Figure 9. As we can see, the Fake News Score is 0.0,
which means the news is most likely true. Let's try checking out https://panorama.pub, a site that
publishes only joke and fake news The result is shown in Figure 10.</p>
      <p>As we can see, Fake News Score: 1. This means that the news is fake.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion and future works</title>
      <p>In this article, we have presented a comprehensive system for fact-checking and analyzing news
content, which utilizes the power of machine learning, natural language processing, and data
integration. Our system is designed to streamline the fact-checking process and enhance its efficiency,
providing users with a valuable tool for verifying the credibility of news items. We have integrated
various modules, each serving a distinct purpose, to create a unified platform for news analysis and
verification. The effectiveness of our system is based on its ability to automate and expedite tasks that
typically demand hours of manual fact-checking. A human fact-checker may spend considerable time
searching both social media and Internet publications, trusted information sources, analyzing context,
and scrutinizing language and style for signs of misinformation or manipulation. With our system, these
tasks are performed in seconds, offering a rapid response and flexibility in customization. Users can
customize the system to their specific needs and preferences. One of the key strengths of our system is
its integration with machine learning techniques, including state-of-the-art NLP transformers. We are
currently exploring the development of a module for automatic text classification using these advanced
technologies. This addition would further enhance our system's ability to discern the authenticity of
news by leveraging the latest advancements in machine learning.
8. References</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Gorokhovsky O. M.</surname>
          </string-name>
          <article-title>Factchek as a trend of investigations: opportunities and prospects: a practical guide</article-title>
          .
          <source>Dnipro: LIRA</source>
          ,
          <year>2017</year>
          . 133 p.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>[2] Challenges of the digital space: social networks, verification and fact-checking</article-title>
          . URL: https://www.aup.com.
          <article-title>ua/vikliki-cifrovogo-prostoru-socialn.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Zhadko</surname>
            <given-names>V. O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kharytonenko</surname>
            <given-names>O. I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poltavets</surname>
            <given-names>Y. S.</given-names>
          </string-name>
          <article-title>Hybrid war and journalism. Problems of information security</article-title>
          . Kyiv: Drahomanov National Pedagogical University,
          <year>2018</year>
          . 356 p.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Krainikova</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Prokopenko</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Waves of disinformation in the hybrid RussianUkrainian war</article-title>
          .
          <source>Current Issues of Mass Communication</source>
          ,
          <volume>33</volume>
          ,
          <fpage>12</fpage>
          -
          <lpage>25</lpage>
          . https://doi.org/10.17721/CIMC.
          <year>2023</year>
          .
          <volume>33</volume>
          .
          <fpage>12</fpage>
          -
          <lpage>25</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <article-title>[5] VoxCheck</article-title>
          . URL : https://voxcheck.voxukraine.org.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>[6] Word and deed</article-title>
          . URL : https://www.slovoidilo.ua.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Stopfake</surname>
          </string-name>
          .org. URL : https://www.stopfake.org/ru/glavnaya-2.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>[8] Pentagon explosion hoax goes viral</article-title>
          . URL: https://www.bleepingcomputer.com/news/ security/pentagon-explosion
          <article-title>-hoax-goes-viral-after-verified-twitter-accounts-push.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <article-title>[9] Eli Lilly lost billions after fake Twitter account promotes free insulin</article-title>
          . URL: https://www.beckershospitalreview.com/pharmacy/eli-lilly
          <article-title>-lost-billions-after-fake-twitteraccount-promotes-free-insulin</article-title>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <article-title>Fake control: how to recognize fakes and not get caught in a lie. website</article-title>
          . URL: https://vsim.ua/ITnovyny/feyk-kontrol
          <article-title>-yak-rozpiznati-feyki-i-ne-potrapiti-na-gachok-brehni-novi-11170324</article-title>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Google</given-names>
            <surname>Photo</surname>
          </string-name>
          . URL: https://images.google.com.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Google</surname>
          </string-name>
          .Lens. URL: https://lens.google/intl/uk.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <article-title>TinEye</article-title>
          . URL: https://tineye.com.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <article-title>Exif Data Viever</article-title>
          . URL: https://exifdata.com.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Fotoforensics</surname>
          </string-name>
          . URL: https://fotoforensics.com.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Reisinger</surname>
            , Joseph; Mooney,
            <given-names>Raymond J</given-names>
          </string-name>
          . (
          <year>2010</year>
          ).
          <article-title>Multi-Prototype Vector-Space Models of Word Meaning. Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics</article-title>
          . Los Angeles, California: Association for Computational Linguistics. pp.
          <fpage>109</fpage>
          -
          <lpage>117</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>E.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rustam</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Ashraf</surname>
          </string-name>
          , P. B. Washington,
          <string-name>
            <given-names>M.</given-names>
            <surname>Narra</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Shafique</surname>
          </string-name>
          ,
          <article-title>"Inquest of Current Situation in Afghanistan Under Taliban Rule Using Sentiment Analysis and Volume Analysis,"</article-title>
          <source>in IEEE Access</source>
          , vol.
          <volume>10</volume>
          , pp.
          <fpage>10333</fpage>
          -
          <lpage>10348</lpage>
          ,
          <year>2022</year>
          , doi: 10.1109/ACCESS.
          <year>2022</year>
          .
          <volume>3144659</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Che-Wen</surname>
            <given-names>Chen</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shih-Pang</surname>
            <given-names>Tseng</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ta-Wen</surname>
            <given-names>Kuan</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jhing-Fa</surname>
            <given-names>Wang</given-names>
          </string-name>
          “
          <article-title>Outpatient Text Classification Using Attention-Based Bidirectional LSTM for Robot-Assisted Servicing in Hospital</article-title>
          ” vol.
          <volume>19</volume>
          ,
          <year>2020</year>
          . URL: https://www.researchgate.net/publication/339343963_Outpatient_
          <article-title>Text_Classification_Using_Attention-Based_Bidirectional_LSTM_for_Robot-Assisted_Servicing_in_Hospital.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Yugesh</surname>
            <given-names>Verma</given-names>
          </string-name>
          , “
          <article-title>Word2Vec vs GloVe - A Comparative Guide to Word Embedding Techniques” URL: https://analyticsindiamag.com/word2vec-vs-glove-a-comparative-guide-to-wordembedding-techniques.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <article-title>Multilingual Universal Sentence Encoder for Semantic Retrieval</article-title>
          . URL: https://ai.googleblog.com/
          <year>2019</year>
          /07/multilingual-universal
          <article-title>-sentence-encoder</article-title>
          .html.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>