<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Information System for the Fact-checker Support</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Yurii</forename><surname>Baryshev</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Vinnytsia National Technical University</orgName>
								<address>
									<addrLine>Khmelnytske shoes str., 95</addrLine>
									<postCode>21021</postCode>
									<settlement>Vinnytsia</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Leonid</forename><surname>Kupershtein</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Vinnytsia National Technical University</orgName>
								<address>
									<addrLine>Khmelnytske shoes str., 95</addrLine>
									<postCode>21021</postCode>
									<settlement>Vinnytsia</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vladyslav</forename><surname>Maidanovych</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Vinnytsia National Technical University</orgName>
								<address>
									<addrLine>Khmelnytske shoes str., 95</addrLine>
									<postCode>21021</postCode>
									<settlement>Vinnytsia</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Olesia</forename><surname>Voitovych</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Vinnytsia National Technical University</orgName>
								<address>
									<addrLine>Khmelnytske shoes str., 95</addrLine>
									<postCode>21021</postCode>
									<settlement>Vinnytsia</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Serhii</forename><surname>Prokopenko</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Semen Kuznets Kharkiv National University of Economics</orgName>
								<address>
									<addrLine>Nauky Ave., 9-A</addrLine>
									<postCode>61166</postCode>
									<settlement>Kharkiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Information System for the Fact-checker Support</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F7DE3E8F8A5D616CC4BE500129A926FD</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T20:01+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Fact-checking</term>
					<term>automation</term>
					<term>fake news</term>
					<term>information attack</term>
					<term>cybersecurity</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The research results of fact-checking automation are presented. The main types of fact-checking were described and the methods and techniques used by fact-checkers were analyzed. Particular attention was paid to the development of the system architecture, which consists of various modules, including a visualization module, database search, internet search, analysis and verification of news on websites and search in telegram channels. Algorithms were developed to work with these modules, allowing for efficient fact-checking. The software architecture was designed and experimental application was developed.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In a world where inaccurate news can spread quickly and have harmful impact, the process of factchecking is becoming vital and increasingly topical. It is a key tool in the fight against so-called fake news and disinformation. Fact-checking is not only a way to identify false statements but also a means of restoring the trust to the media. However, the process of fact-checking can be labor-intensive and time-consuming, especially when dealing with large amounts of information. Fact-checkers often have to spend hours, and sometimes days, searching for evidence to confirm or refute certain facts. In addition, the verification process itself can be challenging, as it requires not only extensive knowledge but also the ability to think critically and analyze information. In this context, the development of an automated fact-checking tool can be not only a significant contribution to the field but also a real asset for journalists, fact-checkers, and anyone working in the informational and media sectors. It is believed that these kind of tools can greatly simplify the fact-checking process, reduce the time required for it, and increase the efficiency and accuracy of the process.</p><p>The goal of our work is to increase the efficiency of the fact-checking process by automating the search, collection and analysis of information using modern information intelligent technologies.</p><p>In the modern world, inundated with vast amounts of information, it is imperative that we have confidence in the accuracy of the information we consume. In this context, this work is highly relevant, as it aims to develop effective tools to assist in verifying the authenticity of information. Future research in this field could greatly enhance capacity to identify and counteract fake news and disinformation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Analysis of Fact-Checking</head><p>Fact-checking is a modern tool for journalistic investigations with a unique concept and technical features aimed at exposing populism, manipulations, and false information in the rhetoric of speakers at various levels and in media publications <ref type="bibr" target="#b0">[1]</ref>. This procedure is crucial in the media industry, where Information Technology and Implementation (IT&amp;I-2023), <ref type="bibr">November 20-21, 2023</ref>, Kyiv, Ukraine EMAIL: kupershtein.lm@gmail.com (L.Kupershtein); vladmaidanovich1@gmail.com (V.Maidanovych); yuriy.baryshev@vntu.edu.ua (Y.Baryshev); voytovych.op@gmail.com (O.Voitovych); prokopenko.serhii@gmail.com (S. Prokopenko). ORCID: 0000-0003-4712-3916 (L.Kupershtein); 0009-0008-1983-479X (V.Maidanovych); 0000-0001-8964-7000 (O.Voitovych); 0000-0003-0718-4917 (S. Prokopenko); 0000-0001-8324-8869 (Y. <ref type="bibr">Baryshev)</ref> journalists, editors, and other professionals are required to effectively verify information before disclosing it. Fact-checking encompasses the verification of aspects such as factual data, quotes, sources of information, and other content.</p><p>It is important to note that fact-checking is not only about identifying false information, but also about the ability to recognize and identify subjectivity and hidden aspects, as well as to assess the quality of information in the context. Professional journalism is distinguished by the usage of reliable data and facts verified by all possible means <ref type="bibr" target="#b1">[2]</ref>.</p><p>This process also plays a significant role in cybersecurity. Disinformation and information manipulation are increasingly prevalent in the modern world, posing threats in various contexts, including wars and conflicts. Military conflicts involve not only physical combat but also the usage of information warfare, such as disinformation and manipulation, to achieve objectives and influence public opinion <ref type="bibr" target="#b2">[3]</ref>. Fact-checking, as an investigative trend, swiftly entered Ukrainian media landscape. This was primarily in response to the wave of anti-Ukrainian propaganda that rapidly flooded information channels with fake and blatantly false reports <ref type="bibr" target="#b0">[1]</ref>. Ukraine serves as an example of information warfare as an integral component of a military conflict. Such disinformation campaigns and information manipulation aim to sow discord, foster distrust, and create social tensions. Their consequences can be severe, impacting political stability, public safety, and the economic situation in the country. Furthermore, these information attacks can have an international impact, altering perceptions and support from allied countries and influencing political relations. To ensure cybersecurity and defend against such information threats, it is crucial to actively implement fact-checking methods and counter the disinformation. The development and utilization of tools for detecting fake news, promoting critical thinking and information literacy, and facilitating access to reliable information can mitigate the impact of disinformation campaigns and enhance cybersecurity in the context of military conflicts and other challenges in the modern world.</p><p>For instance, in the initial days of the full-scale invasion of Ukraine by the russian federation, the Gvara Media team launched the fact-checking telegram-bot "PEREVIRKA". This bot assists in distinguishing fake news from real news and recognizing manipulation and propaganda. Upon receiving news submissions, the information is forwarded to project volunteers who manually verify its truthfulness and accuracy. In just the first month of the invasion, they processed approximately 30000 requests <ref type="bibr" target="#b3">[4]</ref>.</p><p>In Ukraine, the concept of fact-checking is currently represented by several resources:  VoxUkraine: This is a fact-checking project of the independent analytical platform Vox Ukraine. The team exposes lies, manipulations, and russian propaganda, both in Ukraine and abroad. Since 2018, they have been signatories to the Code of Ethics of the Poynter Institute's International Network of Factcheckers, and since 2020 <ref type="bibr" target="#b4">[5]</ref>.</p><p> Word and Deed: This resource checks the election promises of politicians <ref type="bibr" target="#b5">[6]</ref>.  Stopfake.org: Their primary activity is aimed at combating anti-Ukrainian propaganda, statements, and facts aimed at discrediting Ukraine <ref type="bibr" target="#b6">[7]</ref>.</p><p> Gvara Media: An independent online publication focused on social change <ref type="bibr" target="#b3">[4]</ref>.</p><p>For example of what fake news can do is to look at a recent event. Realistic images of an explosion near the Pentagon created by artificial intelligence that went viral on Twitter caused the stock market to briefly plummet. Tweets with images purportedly depicting an explosion near the Pentagon building in Arlington, Virginia, were posted by many verified Twitter accounts, including a russian state media outlet with millions of followers and a verified account claiming to be the Bloomberg news agency. Although at first glance the viral photos appeared to be genuine, they are full of hints that they were created with the help of artificial intelligence, which proves that it is all a hoax <ref type="bibr" target="#b7">[8]</ref>.</p><p>A similar event happened last year. On Nov. 10, a verified Twitter account posed as Eli Lilly wrote in a viral tweet, "We are excited to announce insulin is free now." By the next day, the drugmakers' shares plunged by about $22 billion <ref type="bibr" target="#b8">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methods and Techniques of Fact-Checking</head><p>Fact-checking, as a process of verifying the truthfulness of information, includes a number of methods and techniques that help to carry out this work effectively and accurately.</p><p>These methods include:</p><p> Checking the reliability of sources: A fact-checker determines the reliability, reputation and competence of the source of information, as well as the ability to confirm or refute the information presented.</p><p> Comparison of data: Fact-checkers analyze information from different sources and compare it with known facts or data to detect discrepancies or inaccuracies.</p><p> Checking the context: Fact-checkers analyze the circumstances in which the information was provided and check for distortions or deletions that could change its meaning.</p><p> Involving experts: Fact-checkers can call on experts in the relevant field to verify specific facts or data, using their knowledge and experience.</p><p> Objectivity of experts: Experts involved in fact-checking should be objective and free from conflicts of interest. Their experience and reputation, as well as their qualifications, should be thoroughly checked.</p><p> Use of verified data: It is important to use reliable sources of information, such as official statistics and scientific studies, to compare and corroborate information.</p><p> Technologies for fact-checking: Modern technologies, such as artificial intelligence, machine learning and data analytics, can automate the fact-checking process, increasing its speed and accuracy. They help to identify contradictory statements and manipulation of information.</p><p>To check the authenticity of content on the Internet special software, applications, and browser plugins (Trooclick, Genius, factcheck.org, and others) are used <ref type="bibr" target="#b9">[10]</ref>. Such services are able to analyze facts and identify not only fake news but also manipulative content, identify inaccuracies or biases in the presentation of the material. They scan selected text and check it using official statistics, media reports, their own database, etc. Some of the services are able to analyze headlines, text complexity, popularity, and the structure of the site's code.</p><p>Additionally, there are also services that can aid verification process of multimedia content such as photos and videos.</p><p>Here are some of the services for photos verification:  Google.Photo allows to upload a photo and find similar or identical images. Results can be sorted by size, upload date, and image format <ref type="bibr">[11]</ref>.  Google.Lens works in a similar way to Google.Photo, but allows to search for individual parts of an image, such as buildings in a photo <ref type="bibr" target="#b10">[12]</ref>.</p><p> TinEye -this service allows to find different versions of images, including modified versions, such as cropped, retouched or with added elements. Results can be filtered by relevance, size or differences from the original <ref type="bibr">[13]</ref>.</p><p> EXIF Data/Jeffrey's Image Metadata Viewer -these tools allow to get technical information about a photo, such as the date it was taken, the location, and the device it was taken with. But, this data may not be available for some photos that have been uploaded and re-uploaded multiple times <ref type="bibr" target="#b11">[14]</ref>.</p><p> FotoForensics -this service analyses images using the ELA (error level analysis) method. It compares the image with its compressed version. It helps to identify which part of the image has been altered, whether it is darker or brighter compared to a similar area that has remained unchanged <ref type="bibr" target="#b12">[15]</ref>.</p><p>These tools help fact-checkers detect photographic manipulation and verify the authenticity of images.</p><p>Fact-checking technologies are powerful tools, but they cannot replace the role of a fact-checker. The manual element remains an integral part of the process, as context, cultural differences and subjectivity of information need to be taken into consideration.</p><p>The efficient usage of technology, combined with expert analysis, can lead to more objective and reliable results in fact-checking.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Problem Statement</head><p>Fact-checking is a complex and intensive process that requires a significant amount of time and effort to verify the accuracy of facts, news, and information disseminated in the media and online sources. Given the vast volume of information available, manually verifying all data is highly resourceintensive in terms of time and resources.</p><p>That's why the automation of fact-checking processes is becoming increasingly important. The usage of automated algorithms and software solutions allows for the efficient analysis of large amounts of data, rapid verification of facts, and the identification of potential attempts at disinformation and information manipulation. This, in turn, enhances the speed and accuracy of the fact-checking process, enabling a quicker response to the spread of false information and preventing its adverse consequences.</p><p>The primary objective of our work is to develop an information system based on software tool to support the work of fact-checkers, with the goal of providing a convenient and effective means of detecting fake news.</p><p>Functionality Requirements:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Identification of fake news: The tool should be capable of analyzing news content to determine its credibility. This may involve analyzing the text, source, author, and other relevant attributes of the news.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Search for similar news: The tool should have the capability to search for similar news stories within a database or Internet. This feature will assist fact-checkers in identifying instances of news republishing or the spread of fake news.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Parsing comments: The tool should be able to collect comments on news articles from various sources, social media, and other platforms. This will enable fact-checkers to analyze public reactions and opinions regarding a specific news item.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Analysis of publication dates and authorship: The tool should be able to analyze the publication date and author information. This functionality will help fact-checkers assess the relevance of the news and evaluate the credibility of the author.</p><p>Interface Requirements:</p><p> User-Friendly Interface: The tool should feature an intuitive and user-friendly interface that enables fact-checkers to use it effortlessly and access the necessary information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Visualization of Results: The tool should be capable of visually presenting the results of news analysis. This visual representation will allow fact-checkers to quickly assess the credibility and dissemination of fake news.</p><p>Performance Requirements:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Processing Speed: The tool should have the capability to process a substantial volume of news rapidly, enabling fact-checkers to promptly identify and respond to fake news.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head></head><p>Efficient Resource Utilization: The tool should optimize the utilization of resources, including computing power and memory, to ensure the efficiency of the fact-checkers work.</p><p>The tool should include features such as searching for similar news within the database, searching for news on Telegram channels, parsing comments, identifying fake news, analyzing publication dates and authorship. Additionally, the tool should offer a user-friendly interface, the ability to visualize results, and efficient performance.</p><p>So, we have discussed the concept of fact-checking and its significance in the modern information environment. We have also described the primary types of fact-checking. Following an analysis of the methods and techniques employed by fact-checkers, we have identified the key steps in the factchecking process, including fact verification, source and citation checks, context analysis, and the utilization of specialized tools and data.</p><p>Comprehending these methods and techniques is essential for the effective execution of factchecking tasks. The study has pinpointed the principal challenges associated with fact-checking, such as politicization and the impact of social media.</p><p>The development of such a tool will help to increase the efficiency and accuracy of information verification, as well as help to provide reliable and accurate information to a wide audience by automating the fact-checking process. Let's consider each module separately. The visualization module is responsible for launching the system and presenting the graphical interface to an user. Upon launching the application, the module initializes graphical components such as windows, icons, images, and control buttons for user interaction with the tool. The module handles the display of windows in response to user actions. It also establishes the necessary connections and displays the system's results. After data entry, the module passes this information to the subsequent processing stages.</p><p>The database search module searches and compares the entered data with the existing database of fake and true news. It utilizes algorithms and methods to identify matches and determine the credibility of the news. If the news is found in the database, the component returns the corresponding verification result. However, frequently incoming news may not be unique, as different information channels, in various formats, cover certain events even though they pertain to the same event. Consequently, the same facts can be processed multiple times. Machine-based comparison of the received message with ones stored in the database enables the automatic processing of similar requests. The task of comparing and determining text similarity is relatively straightforward. Currently, there are many tools available for implementing this.</p><p>One of the most popular methods of converting text into a numerical vector is to use vector embedding. A vector embedding is a numerical representation of a word or phrase that preserves the semantic relationships between words <ref type="bibr" target="#b13">[16]</ref>. For example, words with similar meanings will have similar vector representations.</p><p>One of the simplest approaches is the bag-of-words approach. In this approach, the text is considered as a set of individual words, and a vector is created that shows which words are present in the text and how often they occur. Each word is considered a separate feature, and its number in the text is the value of this feature <ref type="bibr" target="#b14">[17]</ref>.</p><p>Another approach is TF-IDF (Term Frequency-Inverse Document Frequency), which is a statistical method for assessing the importance of a word in a document by taking into account its frequency in the document and weighing the frequency of the word in other documents. The TF-IDF is calculated for each word in a corpus of documents and gives a numerical value that reflects its relevance to each individual document.</p><p>The TF-IDF is often used to represent documents in a collection as numerical vectors those reflect the importance of each word in a set of words (the number of words in the set determines the dimension of the vector) in each document. Such model is called a vector model and makes it possible to compare texts by comparing their respective vectors using given metric (Euclidean distance, cosine measure, Manhattan distance, Chebyshev distance, etc.), i.e. to perform a cluster analysis <ref type="bibr" target="#b15">[18]</ref>.</p><p>Next is the Word2Vec model, which learns to predict words based on their context. It creates vector representations of words in which similar words have similar vectors <ref type="bibr" target="#b16">[19]</ref>.</p><p>Another one is the GloVe model (Global Vectors for Word Representation), which combines statistics of word occurrences in texts with matrix decomposition <ref type="bibr" target="#b16">[19]</ref>.</p><p>The FastText model uses the Word2Vec approach, but treats words as a set of character substrings. It allows to obtain vector representations even for words that were not included in the training corpus <ref type="bibr" target="#b16">[19]</ref>. There is also the Universal Sentence Encoder Multilingual model from Google. It is a powerful tool for converting text sentences into vector representations (embeddings). The main goal of the model is to preserve the semantic relations between sentences in the vector space <ref type="bibr" target="#b17">[20]</ref>.</p><p>For our work, we have chosen the Universal Sentence Encoder Multilingual model, which is based on transformers. It provides a convenient and powerful way to convert text into numerical vectors. Its multilingual support allows to work with text in different languages, which is important for analyzing news and content from different sources. In addition, the model has good performance and high accuracy, which makes it an ideal choice for our fact-checking and news content analysis tasks.</p><p>The primary steps for implementing machine fuzzy text matching are: Our array has 512 elements. Each element is represented by a floating point number. Our program utilizes a database that is sourced from a dump of the "ПЕРЕВІРКА" Telegram bot. This database serves as the foundation for data retrieval and operations within our application.</p><formula xml:id="formula_0"></formula><p>The resulting vector arrays can be easily compared using the cosine similarity measure and the percentage of similarity between two texts can be determined.</p><p>The database structure contains the News table with the following fields:  image: A field containing a link to the image associated with the news item. This allows you to store one or more images related to a particular news item.</p><p> video: A field containing a link to a video associated with the news item. It allows you to store one or more videos related to a particular news item.</p><p> fakeStatus: A Boolean field that indicates whether the news is fake. A value of "true" indicates a true news item, and a value of "false" indicates a fake news item.</p><p> createdAt: A field containing the date and time the news was created. This allows you to track when a particular news item was added to the database.  text: A text field that contains the text of the news item. It stores the main content of the news item, which will be used for further analysis and verification.</p><p> id: A unique identifier for the news item. This field allows you to uniquely identify and access a specific news item in the database.</p><p>This database structure allows to store information about news, including images, videos, fake news status, date of creation and the news text itself. A unique identifier helps to unify and uniquely identify each news item in the system.</p><p>The Telegram search module allows to automatically find news in the channels and groups of this platform that have similar content to the entered news text. This module provides an opportunity to effectively analyze and monitor news content, receive up-to-date information from a large number of sources and identify news that have a similar context or topic. Usage of the Telegram search module allows to quickly and conveniently review news from different sources in one place, as well as analyze their similarities and identify related topics or events. This will help users to track up-to-date information, conduct research, and draw informed conclusions based on a wide range of sources.</p><p>The Internet search module allows users to search for news that was not found in the database. It queries a variety of information sources, including news websites, social media and other sources, to find related information about a news item. The module displays a list of URLs of web pages where the news is mentioned and the date of its publication. It also provides special buttons next to each result, which are needed for further analysis and verification of the news.</p><p>The analysis and verification module uses special algorithms and methods to assess the reliability of the news. It performs text processing, detection of suspicious patterns, context analysis and other techniques to determine whether a news item is fake or not. The verification result is presented in an appropriate way, allowing the user to understand how reliable the news is.</p><p>The general algorithm for checking news for authenticity can be summarized as follows:</p><p>Step 1. Input data: Receiving the news for analysis and verification.</p><p>Step  Determining the language of the news text and checking compliance with language rules.  Take into account the style and phrases used that may indicate fake news. Step 5. Checking the source:</p><p> Verifying the news from other trusted sources.  Checking the credibility and reliability of the news source.  Checking whether the source is satirical and the news itself is a joke. Step 6. Using of external sources:</p><p> Querying external sources or databases for more information about the news.  Compare the results and confirm or deny the facts of the news. Step 7. Yielding a result:</p><p> Based on the collected information and analysis, determining whether the news is fake.  Displaying the result, which indicates whether the news is authentic or fake. Step 8. Output data: Receiving the result of the fake news checking. This algorithm describes the main steps and procedures performed in the News Analysis and Verification module. The implementation of specific algorithms and methods may vary depending on the needs. In our software, the algorithm of the analysis and verification module is based on rules and weights. First of all, we parse the elements of the web page where the news is placed, such as the title, author, date of publication, and author. Next, we use rules to check the authenticity of the news. For example, we check whether the source of the news is unreliable: if true, it is set to 1, otherwise to 0. We also check whether the news text is typed: if there are typos, it is set to 1, otherwise to 0. Additionally, we check for manipulative words in the news text: if they are found, it is set to 1, otherwise to 0. Also, if there is no author of the news, it is set to 1, otherwise to 0. We use weights for each rule. In our case, each rule has a weight of 0.25. Finally, we calculate the newsworthiness score by calculating the weighted sum of the rule values.</p><p>The value is calculated using the formula: fake_news_score = (source weight • source rule + text weight • text rule + manipulation weight• • manipulation rule + author weight • author rule). The data storage module is responsible for saving news and verification results in the database. It provides efficient storage and management of the collected data, ensuring its security and availability. This component may use a database or other storage mechanisms to ensure efficient access to data and to ensure data integrity and confidentiality. Each of these modules interacts with the others, transmitting and receiving the necessary data for the system to operate efficiently. This interaction ensures the integration of functionality and improves the system's ability to verify the accuracy of news. We have utilized the capabilities of the Python programming language and the JetBrains PyCharm development environment to construct our system. While the program's user interface was created using the customtkinter library, we have also leveraged a host of other powerful tools to extend its functionality.</p><p>For the transformation of embeddings, we have turned to the renowned TensorFlow library, a stalwart in the field of machine learning and deep neural networks. Its extensive capabilities are instrumental in converting textual data into numerical vectors, thereby enabling effective comparison and similarity measurement. Additionally, for calculating similarity metrics, we have incorporated the sklearn library. This dynamic library equips our system with the ability to assess the similarity between text vectors, employing various metrics like cosine similarity, Euclidean distance, and more.</p><p>In the realm of communication with Telegram, we have utilized Telethon, a Python library that facilitates interaction with the Telegram API. This integration enables our system to access and search Telegram channels, expanding the horizons of information retrieval. When parsing news from Internet content, our system use the Google API Client library. This tool enhances the capacity of our system to conduct web searches and retrieve relevant data. These integrated tools and libraries fortify the core of our system, allowing it to function as a comprehensive fact-checking and news analysis platform.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Experiments</head><p>To launch the program, we need to download the software and run it. The user opens the program window on the "Check by database" tab. It is shown in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 2: Database check window</head><p>On the side of the program is a window with controls:  "Database check" screen;  "Telegram check" screen;  "Checking the network";  element for changing the application theme;  element for changing the interface scale. Let's test the application's intaraction with databases. We already have a ready-made database of fake news. Let's create an array of NumPy vectors for it. To do this, one is to press the "Select File" button and then "Convert". After that, the application asks where to save the file. Once the file is successfully saved, one can select it using the Select File button to continue working. Select database and an array of vectors. Enter the news to be checked in the text box. Enter the number of most similar results those user wants to display. Press the "Search" button. The result is shown in Figure <ref type="figure" target="#fig_2">3</ref>.</p><p>Let's test the developed application with Telegram. We have a ready-made database of links to Telegram channels. Since the algorithm works using the Telegram API, to get started, one needs to get the API ID and API HASH values and enter them into the special fields of application. This procedure needs to be done only once, and these values will be saved in the future.</p><p>After entering the API ID and API HASH values, one needs to enter the news we need, select file with the channel database, and enter the date for which application should search for the news. The result is shown in Figure <ref type="figure" target="#fig_3">4</ref>. After pressing the "News" button, user will see all similar posts from all channels. As a result, program outputs the link to the channel, news text, link to the post and date of publication. The result is shown in Figure <ref type="figure" target="#fig_4">5</ref>. Let's test the parsing of comments. Since we have searched for news by channel, and the results contain links to posts, let's take one of them.</p><p>To start parsing, one needs to paste the link into a special field, and if necessary, select black or white list filtering, and click the "Comments" button. The result of parsing can be seen in Figure <ref type="figure" target="#fig_5">6</ref>. Let's try to perform the same action, but with the whitelist filter enabled. The result is shown in Figure <ref type="figure" target="#fig_6">7</ref>.      Let's test the search for news on the web. To do this, one is to enter the news in the search field and press the "Search" button. The search results will be displayed in the text box. The result is shown in Figure <ref type="figure" target="#fig_7">8</ref>. As it can be seen from fig. <ref type="figure" target="#fig_7">8</ref> the results include the headline of the news and the link to it. Next, we can test the news verification on the web page. Since the search returned certain results, we can take the link from there. Let's paste the selected link into a special field and press the "Check. The verification process yield the result shown in Figure <ref type="figure" target="#fig_8">9</ref>. As we can see, the Fake News Score is 0.0, which means the news is most likely true. Let's try checking out https://panorama.pub, a site that publishes only joke and fake news The result is shown in Figure <ref type="figure" target="#fig_9">10</ref>.</p><p>As we can see, Fake News Score: 1. This means that the news is fake.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion and future works</head><p>In this article, we have presented a comprehensive system for fact-checking and analyzing news content, which utilizes the power of machine learning, natural language processing, and data integration. Our system is designed to streamline the fact-checking process and enhance its efficiency, providing users with a valuable tool for verifying the credibility of news items. We have integrated various modules, each serving a distinct purpose, to create a unified platform for news analysis and verification. The effectiveness of our system is based on its ability to automate and expedite tasks that typically demand hours of manual fact-checking. A human fact-checker may spend considerable time searching both social media and Internet publications, trusted information sources, analyzing context, and scrutinizing language and style for signs of misinformation or manipulation. With our system, these tasks are performed in seconds, offering a rapid response and flexibility in customization. Users can customize the system to their specific needs and preferences. One of the key strengths of our system is its integration with machine learning techniques, including state-of-the-art NLP transformers. We are currently exploring the development of a module for automatic text classification using these advanced technologies. This addition would further enhance our system's ability to discern the authenticity of news by leveraging the latest advancements in machine learning.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Information system architecture</figDesc><graphic coords="5,121.40,72.00,351.75,192.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>2 .</head><label>2</label><figDesc>Pre-processing:  Parsing data from the web page  Identification of the author.  Parsing the publication date of the news.  If supported by the website, parsing of comments. Step 3. Context analysis:  Analyzing the context of the news, including keywords, topics and other parameters.  Identify suspicious patterns or incorrect information.  Check whether links to other sources are present. Each news item should be linked to a source. Step 4. Language and style analysis:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Database search result</figDesc><graphic coords="9,143.50,325.43,322.50,205.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Search interface for Telegram channels</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Search result in Telegram channels</figDesc><graphic coords="9,135.70,556.36,305.00,190.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: The result of parsing comments without a filter</figDesc><graphic coords="10,133.00,248.93,343.00,172.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Parsing result with whitelist filter</figDesc><graphic coords="10,180.25,447.21,248.53,246.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: News search result in the network</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: The result of checking the news on the web page</figDesc><graphic coords="11,148.40,141.24,298.00,269.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: The result of checking a joke site</figDesc><graphic coords="11,164.50,435.87,280.15,206.60" type="bitmap" /></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Information system architecture</head><p>The information system architecture includes several modules. It is shown in Figure <ref type="figure">1</ref>. It consists of six modules: visualization module, database-search module, telegram-search module, network-search module, analysis and verification module and data-save module. System has its own database also.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Factchek as a trend of investigations: opportunities and prospects: a practical guide</title>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">M</forename><surname>Gorokhovsky</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>LIRA</publisher>
			<biblScope unit="page">133</biblScope>
			<pubPlace>Dnipro</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<ptr target="https://www.aup.com.ua/vikliki-cifrovogo-prostoru-socialn" />
		<title level="m">Challenges of the digital space: social networks, verification and fact-checking</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">O</forename><surname>Zhadko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><forename type="middle">I</forename><surname>Kharytonenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Poltavets</surname></persName>
		</author>
		<title level="m">Problems of information security</title>
				<meeting><address><addrLine>Kyiv</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page">356</biblScope>
		</imprint>
		<respStmt>
			<orgName>Drahomanov National Pedagogical University</orgName>
		</respStmt>
	</monogr>
	<note>Hybrid war and journalism</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Waves of disinformation in the hybrid Russian-Ukrainian war</title>
		<author>
			<persName><forename type="first">T</forename><surname>Krainikova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Prokopenko</surname></persName>
		</author>
		<idno type="DOI">10.17721/CIMC.2023.33.12-25</idno>
		<ptr target="https://doi.org/10.17721/CIMC.2023.33.12-25" />
	</analytic>
	<monogr>
		<title level="j">Current Issues of Mass Communication</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="12" to="25" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title/>
		<author>
			<persName><surname>Voxcheck</surname></persName>
		</author>
		<ptr target="https://voxcheck.voxukraine.org" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<ptr target="https://www.slovoidilo.ua" />
		<title level="m">Word and</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title/>
		<author>
			<persName><surname>Stopfake</surname></persName>
		</author>
		<ptr target="https://www.stopfake.org/ru/glavnaya-2" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<ptr target="https://www.bleepingcomputer.com/news/security/pentagon-explosion-hoax-goes-viral-after-verified-twitter-accounts-push" />
		<title level="m">Pentagon explosion hoax goes viral</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<ptr target="https://www.beckershospitalreview.com/pharmacy/eli-lilly-lost-billions-after-fake-twitter-account-promotes-free-insulin.html" />
		<title level="m">Eli Lilly lost billions after fake Twitter account promotes free insulin</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<ptr target="https://vsim.ua/IT-novyny/feyk-kontrol-yak-rozpiznati-feyki-i-ne-potrapiti-na-gachok-brehni-novi-11170324.html" />
		<title level="m">Fake control: how to recognize fakes and not get caught in a lie</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title/>
		<author>
			<persName><surname>Google</surname></persName>
		</author>
		<author>
			<persName><surname>Lens</surname></persName>
		</author>
		<ptr target="https://lens.google/intl/uk" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<ptr target="https://exifdata.com" />
		<title level="m">Exif Data Viever</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<ptr target="https://fotoforensics.com" />
		<title level="m">Fotoforensics</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Multi-Prototype Vector-Space Models of Word Meaning</title>
		<author>
			<persName><forename type="first">Joseph</forename><forename type="middle">;</forename><surname>Reisinger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Raymond</forename><forename type="middle">J</forename><surname>Mooney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics</title>
		<title level="s">Association for Computational Linguistics</title>
		<meeting><address><addrLine>Los Angeles, California</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="109" to="117" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Inquest of Current Situation in Afghanistan Under Taliban Rule Using Sentiment Analysis and Volume Analysis</title>
		<author>
			<persName><forename type="first">E</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rustam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ashraf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">B</forename><surname>Washington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Narra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Shafique</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2022.3144659</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="10333" to="10348" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Outpatient Text Classification Using Attention-Based Bidirectional LSTM for Robot</title>
		<author>
			<persName><forename type="first">Che-Wen</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shih-Pang</forename><surname>Tseng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ta-Wen</forename><surname>Kuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jhing-Fa</forename><surname>Wang</surname></persName>
		</author>
		<ptr target="https://www.researchgate.net/publication/339343963_Outpatient_Text_Classifi-cation_Using_Attention-Based_Bidirectional_LSTM_for_Robot-Assisted_Servicing_in_Ho-spital" />
	</analytic>
	<monogr>
		<title level="m">-Assisted Servicing in Hospital</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">19</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Word2Vec vs GloVe -A Comparative Guide to Word Embedding Techniques</title>
		<author>
			<persName><forename type="first">Yugesh</forename><surname>Verma</surname></persName>
		</author>
		<ptr target="https://analyticsindiamag.com/word2vec-vs-glove-a-comparative-guide-to-wordembed-ding-techniques" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Multilingual Universal Sentence Encoder for Semantic Retrieval</title>
		<ptr target="https://ai.googleblog.com/2019/07/multilingual-universal-sentence-encoder.html" />
		<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
