<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">An Intelligent Information System for Generating a Scientist&apos;s Scientometrics Using Content Analysis Methods *</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Mykola</forename><surname>Dyvak</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">West Ukrainian National University</orgName>
								<address>
									<addrLine>11 Lvivska Street</addrLine>
									<postCode>46001</postCode>
									<settlement>Ternopil</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Andriy</forename><surname>Yushko</surname></persName>
							<email>a.yushko@wunu.edu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">West Ukrainian National University</orgName>
								<address>
									<addrLine>11 Lvivska Street</addrLine>
									<postCode>46001</postCode>
									<settlement>Ternopil</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Andriy</forename><surname>Melnyk</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">West Ukrainian National University</orgName>
								<address>
									<addrLine>11 Lvivska Street</addrLine>
									<postCode>46001</postCode>
									<settlement>Ternopil</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tiande</forename><surname>Pan</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">West Ukrainian National University</orgName>
								<address>
									<addrLine>11 Lvivska Street</addrLine>
									<postCode>46001</postCode>
									<settlement>Ternopil</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<orgName type="department">International Scientific and Practical Conference Applied Information Systems and Technologies Digital</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">An Intelligent Information System for Generating a Scientist&apos;s Scientometrics Using Content Analysis Methods *</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">71C9360F30BB8350EE61D1307A853D47</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T20:11+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>intelligent information system</term>
					<term>scientometrics</term>
					<term>researcher</term>
					<term>content analysis methods</term>
					<term>large language model</term>
					<term>irrelevant publications. *</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The paper proposes methods and software tools for developing a scientometric profile of a researcher using content analysis techniques. A scientometric profile is a system of indicators that assesses a researcher's scientific productivity and influence. The growing volume of scientific information in various databases, such as Scopus and Web of Science, has made it challenging to manually track and analyze individual publishing activities. For scientific and higher education institutions, monitoring both the quantity and quality of publications is crucial. Additionally, understanding researchers' main areas of interest helps support their professional development and foster interdisciplinary collaboration. Existing tools for monitoring scientific metrics typically offer limited functionality, lack the ability to process large volumes of data efficiently, and struggle to filter irrelevant information automatically. This paper presents an approach to building a researcher's scientometric profile using content analysis, supported by large language models, specifically Ollama. A mathematical model was developed to filter out irrelevant publications based on the researcher's scientometric profile. The system for collecting and analyzing scientometric indicators was implemented, and experimental studies were conducted using profiles of researchers from West Ukrainian National University.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the modern scientific environment, the problem of managing the publishing activity of professors and teaching staff is gaining more and more importance. The increase in the volume of scientific information placed in various scientometric databases, such as Scopus, Web of Science, and others, greatly complicates the manual process of tracking and analyzing the publication activity of each individual researcher. For scientific institutions and institutions of higher education, it is important not only to monitor the number and quality of publications, but also to understand the main scientific interests of employees, which allows to support their professional activities and develop interdisciplinary cooperation <ref type="bibr" target="#b0">[1]</ref><ref type="bibr" target="#b1">[2]</ref><ref type="bibr" target="#b2">[3]</ref><ref type="bibr" target="#b3">[4]</ref>.</p><p>Existing tools for monitoring scientific publications, as a rule, have limited functionality, are not able to efficiently process large volumes of information and automatically filter irrelevant publications. They often require significant human resources to analyze and select relevant data, which is not always possible for large scientific teams <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b4">5]</ref>.</p><p>The purpose of this article is to present the developed software system that automates the collection, analysis and processing of the publishing activity of the teaching staff. The system not only allows you to collect information from scientific databases about publications, projects, grants and participation in scientific events, but also forms a profile of a scientist, determining his scientific interests. Using this profile, the system is able to filter irrelevant publications, automatically assessing their relevance to the scientist's interests. This decision contributes to increasing the efficiency of scientific activity, allowing to focus attention on really important and relevant scientific achievements.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Algorithms and approaches for selecting keywords and determining the researcher's scientific interests</head><p>Modern research actively uses algorithms for automatic analysis of text data to select keywords that reflect the main scientific interests of the researcher. The development of such approaches is aimed at simplifying the process of collecting, analyzing and systematizing scientific materials, which allows not only to identify the main areas of work, but also to identify interdisciplinary connections.</p><p>The main methods used to analyze texts for the purpose of extracting keywords can be divided into several categories:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Statistical methods</head><p>One of the basic approaches is to calculate the frequency of use of terms in texts. The TF-IDF metric (Term Frequency-Inverse Document Frequency) is the most popular among statistical methods and allows taking into account both the frequency of a term in a document and its significance in the context of the entire corpus of texts <ref type="bibr" target="#b5">[6]</ref>. This increases the accuracy of extracting significant terms, as frequent but insignificant words are given less weight. Figure <ref type="figure" target="#fig_0">1</ref> shows an example of the implementation of the TF-IDF metric in the Python programming language using the scikit-learn library. As a result of executing the code, we will get a table with the top 10 keywords and their TF-IDF values.</p><p>The TF-IDF value of each keyword reflects its weight in the context of the article's annotationthe higher the TF-IDF value, the more important the term is for this text</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Rule-based methods</head><p>Rule-based approaches, such as Named Entity Recognition (NER), allow the extraction of certain categories of words, such as names of organizations, names of people, geographic locations, and other important entities <ref type="bibr" target="#b6">[7]</ref>. In the Python programming language, you can use the transformers library from Hugging Face, which allows you to load a pre-trained model for recognizing named entities (Fig. <ref type="figure" target="#fig_1">2</ref>). As you can see from the code above, we use the pipeline method with a pre-trained dbmdz/bertlarge-cased-finetuned-conll03-english model that is specially tuned for Named Entity Recognition (NER). The aggregation_strategy="simple" parameter allows you to aggregate the results for greater convenience.</p><p>The next step is to run NER on the text. This pulls up a list of colds with the specified types (eg organizations, technology names, scientific concepts).</p><p>After that, keyword filtering is performed by selecting entities that may be relevant. For example ORG (organizations) and MISC (various terms such as technology or scientific concepts).</p><p>After passing all the stages, we get a list of keywords selected from the text, in our case it is: Google Cloud Reduce and MapReduce.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Natural language processing (NLP) models</head><p>Thanks to the development of natural language processing methods and the emergence of deep models such as BERT, GPT and others, it became possible to significantly improve the accuracy of text analysis <ref type="bibr" target="#b7">[8]</ref>. These models take into account the context of words, which allows not only to highlight keywords, but also to understand their relationship and semantic meaning.</p><p>Figure <ref type="figure" target="#fig_2">3</ref> shows a code fragment for selecting article categories by their annotations using the ready-made facebook/bart-large-mnli model from the Transformers library.</p><p>In Figure <ref type="figure" target="#fig_3">4</ref>, the output shows which categories most closely match each text, as well as the confidence level of the model corresponding to each category.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Text vectorization</head><p>To vectorize the text and create its numerical representation, you can use the Word2Vec or Doc2Vec methods from the gensim library in Python. Word2Vec creates vectors for individual words, while Doc2Vec allows you to get a vector representation for an entire document <ref type="bibr" target="#b8">[9]</ref> (Fig. <ref type="figure" target="#fig_4">5</ref>). As a result of executing this code, we will get a vector representation for three annotation texts, which is shown in Figure <ref type="figure">6</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 6: Vector representation of annotations</head><p>In the future, the obtained vectors can be used to compare the similarities between documents or to cluster documents based on topics. For example, we can calculate cosine similarity between vectors to find out how similar documents are to each other (Fig. <ref type="figure" target="#fig_5">7</ref>). The cosine similarity value of two documents of 0.98 indicates that these documents have a very high level of similarity in terms of their vector representations. Cosine similarity measures the angle between the vectors of two texts: a value close to 1 means that the vectors are nearly parallel, indicating a high degree of similarity between the texts.</p><p>These methods can be used both separately and in combination to achieve more accurate results in determining the researcher's key scientific interests. The use of these approaches allows automating the processes of analysis of scientific activity, which, in turn, contributes to the formation of a comprehensive profile of a researcher capable of reflecting the dynamics of his scientific work and interdisciplinary connections.</p><p>Each of the described methods has its own unique application and can complement other methods in complex tasks of text analysis. In the next section, we will look at how you can use the Ollama model with its powerful language models to identify keywords in text. This approach will make it possible to apply the latest capabilities of deep learning to improve the accuracy of extracting relevant terms and analyzing complex textual data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Methodology for creating a scientometric portrait of scientist using large Ollama language models</head><p>A researcher's profile is a comprehensive description of the researcher's professional activities, scientific achievements, and interests. It includes such key elements as name, surname, position, academic title, scientific interests, number of published works, participation in scientific grants and projects. The formation of a scientist's profile is an important task, since it can be used to solve such tasks as automated filtering of publications that match the researcher's scientific interests, and optimized selection of a scientific supervisor for young scientists or graduate students whose scientific activity coincides with the topic.</p><p>To form a profile of a scientist, first of all, it is necessary to collect basic metadata, which will become the foundation for further processing. The web scraping method can help us in this, which will allow us to collect basic information from the official website of the organization where the scientist works. This method provides automated extraction of such data as name, surname, position, academic title, circle of scientific interests, links to scientometric profiles of the author (Scopus, Web of Science, ORCID, Google Scholar, DSpace).</p><p>The use of web scraping at the initial stage provides automatic filling of the profile with publicly available information, which significantly reduces the time spent on manual data collection and creates an accurate starting point for further analysis.</p><p>To implement the web scraping process, you can use specialized libraries that allow you to automatically read and extract information from web pages <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>. For example, using the Cheerio library in JavaScript, it is possible to retrieve and process the HTML content of a page, extracting the required metadata such as name, title, academic interests, etc. The following example demonstrates the basic code for obtaining information about a scientist from the official website of the Western Ukrainian National University, focusing on the necessary profile elements.</p><p>Figure <ref type="figure" target="#fig_6">8</ref> shows a fragment of the code for parsing the metadata of scientists from the official website of the organization. Now that we have a basic set of web-scraping metadata, we can move on to the next stepfleshing out a scientist's profile using Ollama's large language models. Large Language Models (LLM) are a powerful tool for analyzing and processing textual data. Thanks to the ability to understand the context and extract meaningful units.</p><p>The main advantage of Ollama is the ability to run and manage large language models (LLM) locally on a computer, without the need for cloud services. This ensures increased confidentiality of data, reduces costs and allows users to fully control information processing processes <ref type="bibr" target="#b9">[10]</ref>.</p><p>The models presented in the Ollama platform are specialized in the processing of scientific texts and have a wide range of applications, such as automatic text classification, extraction of keywords and phrases, identification of scientific interests and creation of a generalized profile.</p><p>Table <ref type="table" target="#tab_0">1</ref> provides a comparative analysis of the major language models supported by the Ollama platform <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>.</p><p>The above table shows the main features of the models, their advantages and disadvantages and allows you to choose the appropriate model for solving this or that problem. Effective in code generation and analysis of large volumes of data.</p><p>Narrow specialization; may be less effective in other tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Code Llama 1,5B ~3</head><p>Optimized for tasks related to programming; support for various programming languages.</p><p>Narrow specialization; not suitable for general NLP tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Gemma 2 1,1B ~2</head><p>Multimodal model; processes text and images simultaneously.</p><p>Requires more resources to process multimodal data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Phi-3 1,4B ~2,8</head><p>Specialized in scientific and research tasks; high accuracy.</p><p>May be less effective in general tasks; needs setting.</p><p>Phi-3 1,4B ~2,8 Specialized in scientific and research tasks; high accuracy.</p><p>May be less effective in general tasks; needs setting.</p><p>Figure <ref type="figure" target="#fig_7">9</ref> demonstrates the process of forming a profile of a scientist, which includes the main stages: data collection, analysis of text documents using Ollama models, parsing of scientific interests, classification of information and final creation of a profile. As can be seen from the diagram above, the profile formation process can be divided into six main stages:</p><p>1. Request -the process begins with formulating a request to create a profile of a scientist; 2. Metadata web scraping -the first stage of information collection. Web scraping is used to obtain basic metadata, such as name, title, research interests, etc., from organizations' websites or other available sources; 3. Data pre-processing -collected data undergo a process of cleaning and structuring to prepare them for further analysis; 4. Prompt -a generated request or command for data analysis using a language model; 5. Analysis of scientific publications -the language model analyzes scientific publications, projects and dissertations to highlight key topics, scientific interests and specializations; 6. Update of profile data -based on the analysis of publications and additional data, the profile of a scientist is formed or supplemented.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">A mathematical model of filtering irrelevant publications based on the profile of a scientist</head><p>In scientometric databases, a problem often arises when, due to the coincidence of the author's last name, first name, and patronymic, publications that do not belong to the scientist are added to the scientist's profile. This distorts the indicators of scientific activity and complicates the objective assessment of the researcher's contribution. The development of a mathematical model for filtering irrelevant publications based on the profile of a scientist allows to effectively solve this problem.</p><p>To build mathematical models in conditions of limited data sampling, it is advisable to use methods based on interval analysis <ref type="bibr" target="#b13">[14]</ref><ref type="bibr" target="#b14">[15]</ref><ref type="bibr" target="#b15">[16]</ref><ref type="bibr" target="#b16">[17]</ref><ref type="bibr" target="#b17">[18]</ref>. Using detailed scholarly profile data, such as author research interests, affiliations, and other unique characteristics, it is possible to accurately identify publications that actually belong to a particular scholar. This ensures an increase in the accuracy of scientometric indicators and will contribute to a more objective analysis of scientific activity. The process of building a mathematical model for filtering irrelevant publications based on the profile of a scientist can be divided into several steps:</p><p>Step 1. Formulation of the author's scientific interests. The author's scientific interests can be represented as a vector of keywords that provides an opportunity to describe the main areas of research. Let    12 , , , n I k k k , where i k is a keyword or phrase describing the author's interests.</p><p>Step 2. Vector representation of the publication. Each post can also be represented as a vector of keywords. Let Step 3. Calculating relevance using cosine similarity. To measure the similarity between the scientific interests of the author I and the publication vector </p><formula xml:id="formula_0">, (<label>1</label></formula><formula xml:id="formula_1">)</formula><p>where  is the scalar product operation. The value of   , j relevance I P ranges from 0 to 1, where a value close to 1 means high relevance.</p><p>Step 4. Filtering of irrelevant publications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>If the value</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head> </head><p>, j relevance I P is less than some threshold T , then the publication is considered irrelevant and is filtered out.</p><formula xml:id="formula_2">j P is relevant if    , j relevance I P T</formula><p>So, as we can see, the model we received allows us to automatically filter out irrelevant publications based on the scientific profile of the author.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9.">Software implementation of the system for collecting and analyzing scientific and scientific-pedagogical activities of the academic team</head><p>In the modern conditions of the information society, it is important to have effective tools for collecting and analyzing scientific activity <ref type="bibr" target="#b18">[19]</ref><ref type="bibr" target="#b19">[20]</ref><ref type="bibr" target="#b20">[21]</ref>. The developed system is aimed at automating the processes of data collection about the scientific and scientific and pedagogical achievements of the academic staff, filtering this data based on relevance to their interests, and forming reports by university, faculty, department, which allows to improve the quality of management and planning of scientific work. Conventionally, our system can be divided into several interacting modules, namely:</p><p>1. The authorization and authentication module, which ensures secure user access to the system, access management and protection of personal data; 2. Data collection module: responsible for obtaining information from scientometric databases (for example, Scopus, Crossref, NRAT) and the profile of a scientist; 3. Data processing and analysis module: cleans, normalizes and pre-processes collected data for preparation for further analysis; 4. Filtering module: implements a mathematical filtering model using machine learning algorithms and criteria defined on the basis of the scientist's profile; 5. Reporting module: provides an opportunity to generate a general report on the scientific activity of the university, faculty or department; 6. User interface: provides user interaction with the system, providing the ability to perform CRUD operations with the main entities (for example, publications, dissertations, grants, projects, scientific activities). The system architecture was implemented using advanced technologies that ensure reliability, performance and flexibility. The core technology stack is based on JavaScript as both client-side and server-side programming languages, which helps ensure codebase consistency and eases application development. The server part was developed using Node.js, which allows you to create high-performance and scalable server applications with high request processing capabilities even in real time.</p><p>To optimize the interaction between the client and the server, GraphQL is used, which gives the client the opportunity to get only the data that is needed, which reduces the load on the network and server resources. Which, in turn, will improve system performance from the point of view of building complex queries.</p><p>The MongoDB database acts as a storage, which provides speed and flexibility when working with large volumes of unstructured data. It also provides efficient work with various data types used in describing the structure for data from various scientometric information systems, and provides easy scalability of the database in accordance with the load and needs.</p><p>In addition, the Ollama platform is integrated into the system, which provides a mechanism for working with various models of machine learning and artificial intelligence. Thanks to such capabilities, the system can more accurately determine the relevance of publications based on the profile of a scientist, calculating complex relationships between data.</p><p>The system interface is developed based on the principles of building intuitiveness and ease of use, which provides convenient user access to the main functionality without the need for additional training of personnel.</p><p>Figure <ref type="figure" target="#fig_10">10</ref> shows the initial screen of the page with the authorization and authentication forms. As you can see from the screenshot above, the authorization form is quite simple, as it requires the user to enter only the e-mail address and the password that was created during registration in the system (Fig. <ref type="figure" target="#fig_11">11</ref>). The registration form requires the user to fill out a database of information about himself, such as surname, first name, patronymic, position, faculty, department, etc. Also, when registering, the employee must specify his identifiers in other scientometric databases to ensure the process of automated information collection. Another of the main fields of this form is the scientist's last name and first name in Latin, as these data will be needed to search for publications in the Crossref database.</p><p>After successful authorization in the system, the user gets to the "Overview" page (Fig. <ref type="figure" target="#fig_12">12</ref>) where he can see quantitative indicators of publication activity. As can be seen from the screenshot above, the user has the opportunity to filter all indicators by faculty, department and publication period.</p><p>There is also an opportunity to create a world based on a specific division by clicking the "Download report" button. This opportunity is available only to employees with appropriate access rights (for example, the head of the department, the dean of the faculty, the vice-rector for scientific work).</p><p>If the user has entered the system and there is no added data yet, he will see a welcome window and a button that will allow synchronization of all publication activity from other scientometric databases (Fig. <ref type="figure" target="#fig_13">13</ref>). After pressing the "Synchronization" button, a window will open with a description of the databases in which information will be searched (Fig. <ref type="figure" target="#fig_14">14</ref>). Also, as can be seen from the image above, the user is provided with brief information about the period for which his data will be collected (by default 5 years) and how exactly this process will take place.</p><p>After successful synchronization, the user of the system receives a message in his e-mail box, which informs that his publishing activity has been successfully collected.</p><p>By going to the "Publications" section, the user will be able to view all the publications that the system managed to find (Fig. <ref type="figure" target="#fig_15">15</ref>). If the system could not find any publication of the author, he can add it manually by pressing the corresponding button. After that, the user will open a form where he will need to fill in all the necessary fields (Fig. <ref type="figure" target="#fig_16">16</ref>). The page for viewing dissertations protected by the user, where there is an addition form, has a similar appearance (Fig. <ref type="figure" target="#fig_18">17</ref>  Also, our system provides a section where you can view the list of granted grants (Fig. <ref type="figure" target="#fig_19">18</ref>).   In the planned events section, users can plan new events and view existing ones (Fig. <ref type="figure" target="#fig_22">20</ref>). Figure <ref type="figure" target="#fig_23">21</ref> shows the Add Activities form, which includes a variety of required fields marked with an asterisk to collect information. It includes fields such as "Name", "Type", "Faculty", "Department", "Date", "Place", "Number of participants" and "Co-organizers". There are also sections for the contact details of the organizing committee, where you need to enter a postal address, phone number and email address. The R&amp;D funding page displays a list of research and development (R&amp;D) funded projects, including the name, manager, terms, amount of funding, type of funding, and category of each project (Fig. <ref type="figure" target="#fig_24">22</ref>).</p><p>It is also possible to quickly search by faculty, department and deadline. As already mentioned earlier, the system provides for the possibility of automatic creation of a scientist's profile, which can be used in the future for the tasks of filtering publications.   As can be seen from the figure above, the user has the opportunity not only to view his profile, but also to edit the necessary information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="10.">Conclusion</head><p>This work emphasizes the need for automation of collection, processing and analysis of publication activity in the modern scientific environment. The increase in the volume of scientific information complicates the manual control and analysis of data, especially in large academic groups. The developed system described in this paper not only provides automated collection of information from scientometric databases such as Scopus and Web of Science, but also forms a profile of a scientist, which includes information about his scientific interests, publications, grants and participation in scientific events. This allows you to optimize the processes of managing scientific activities, making them more efficient and objective.</p><p>An important part of the work is the use of modern algorithms for automatic text analysis, such as TF-IDF, Named Entity Recognition (NER) and text vectorization, which contribute to the selection of keywords and the identification of scientific interests of researchers. The application of deep language models, such as BERT, GPT, as well as the capabilities of the Ollama platform for localized processing of big data, allows you to achieve high accuracy in text analysis, taking into account the semantic context and the relationship between terms.</p><p>In addition, a mathematical model of filtering irrelevant publications is built in the work, which is based on the profile of a scientist, which solves the problem of filtering the author's original works, thereby significantly increasing the accuracy of scientometric indicators.</p><p>Also, the use of vector representation of scientific interests and publications with the calculation of cosine similarity is proposed for the first time. This approach contributes to the objective assessment of scientific contributions, reducing the risk of inaccuracies due to random coincidence of surnames or errors in databases.</p><p>Another important component of this work is the integration of the Ollama platform into its own system, which allows the use of language models for accurate identification of scientific interests, as well as for automatic categorization and clustering of scientific materials. This greatly facilitates the formation of reports for scientific institutions, which allows you to quickly obtain generalized data on the activities of the university, faculties and departments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Declaration on Generative AI</head><p>During the preparation of this work, the authors used ChatGPT and Grammarly to check grammar and spelling, paraphrase, and reword the text. These tools help identify and correct grammatical errors, typos, and other writing mistakes, improving the clarity and professionalism of the text. After using these tools, the authors reviewed and edited the content as needed and take full responsibility for the publication's content.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Implementation of the statistical method of selecting keywords from the test using the TF-IDF metric.</figDesc><graphic coords="2,78.95,451.49,442.72,263.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Implementation of the NER method for searching keywords.</figDesc><graphic coords="3,77.58,235.85,445.32,213.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Implementation of the BART method.</figDesc><graphic coords="4,79.48,62.35,441.61,163.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The result of text categorization using the BART method.</figDesc><graphic coords="4,193.73,254.99,213.20,200.73" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Implementation of the Doc2Vec vectorization method.</figDesc><graphic coords="4,79.08,553.98,442.50,194.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Calculation of cosine similarity of two documents.</figDesc><graphic coords="5,152.55,350.96,309.04,76.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Code listing for parsing metadata.</figDesc><graphic coords="6,76.83,338.23,447.00,369.69" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: Scheme of formation of a scientist's profile.</figDesc><graphic coords="7,136.07,580.21,328.20,176.98" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>jP</head><label></label><figDesc>, you can use the cosine similarity:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: Authorization form.</figDesc><graphic coords="10,137.95,62.35,324.74,170.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head>Figure 11 :</head><label>11</label><figDesc>Figure 11: Registration form in the system.</figDesc><graphic coords="10,114.32,318.83,371.84,302.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Figure 12 :</head><label>12</label><figDesc>Figure 12: Viewing the number of publications in each of the author's categories.</figDesc><graphic coords="11,78.33,62.35,443.90,113.99" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_13"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The user's welcome page in the system.</figDesc><graphic coords="11,76.65,344.34,447.33,116.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_14"><head>Figure 14 :</head><label>14</label><figDesc>Figure 14: Synchronization confirmation window.Also, as can be seen from the image above, the user is provided with brief information about the period for which his data will be collected (by default 5 years) and how exactly this process will take place.After successful synchronization, the user of the system receives a message in his e-mail box, which informs that his publishing activity has been successfully collected.</figDesc><graphic coords="11,99.13,536.97,402.35,140.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_15"><head>Figure 15 :</head><label>15</label><figDesc>Figure 15: Publications display page.</figDesc><graphic coords="12,76.55,104.94,447.55,220.25" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_16"><head>Figure 16 :</head><label>16</label><figDesc>Figure 16: Form for adding a new publication.</figDesc><graphic coords="12,79.08,414.87,442.36,221.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_17"><head></head><label></label><figDesc>).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_18"><head>Figure 17 :</head><label>17</label><figDesc>Figure 17: Dissertation viewing page.</figDesc><graphic coords="13,127.82,62.35,344.95,167.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_19"><head>Figure 18 :</head><label>18</label><figDesc>Figure 18: Grants View Page.</figDesc><graphic coords="13,125.57,293.99,349.03,166.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_20"><head>Figure 19</head><label>19</label><figDesc>Figure19shows the form for adding a new grant.</figDesc><graphic coords="13,106.48,522.33,387.70,194.35" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_21"><head>Figure 19 :</head><label>19</label><figDesc>Figure 19: Grant addition form.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_22"><head>Figure 20 :</head><label>20</label><figDesc>Figure 20: View and add events page.</figDesc><graphic coords="14,88.08,62.35,424.38,210.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_23"><head>Figure 21 :</head><label>21</label><figDesc>Figure 21: Event planning form.</figDesc><graphic coords="14,78.33,390.02,443.36,222.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_24"><head>Figure 22 :</head><label>22</label><figDesc>Figure 22: R&amp;D funding review page.</figDesc><graphic coords="15,76.55,62.35,447.65,220.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_25"><head>Figure 23</head><label>23</label><figDesc>Figure23shows the scientist's profile view page.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_26"><head>Figure 23 :</head><label>23</label><figDesc>Figure 23: Scientist profile page.</figDesc><graphic coords="15,76.55,344.84,447.55,189.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Comparison table of major language models in Ollama</figDesc><table><row><cell>Model</cell><cell>Number of parameters</cell><cell>Size (GB)</cell><cell>Advantages</cell><cell>Disadvantages</cell></row><row><cell></cell><cell></cell><cell></cell><cell>High productivity in general</cell><cell>Requires significant computing</cell></row><row><cell>Llama 3.2</cell><cell>1,2B</cell><cell>~2</cell><cell>NLP tasks; support for many</cell><cell>resources for efficient</cell></row><row><cell></cell><cell></cell><cell></cell><cell>languages.</cell><cell>operation.</cell></row><row><cell>Mistral</cell><cell>1,3B</cell><cell>~2,5</cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Tracing the influence of large language models across the most impactful scientific works</title>
		<author>
			<persName><forename type="first">D.-M</forename><surname>Petroșanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pîrjan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tăbușcă</surname></persName>
		</author>
		<idno type="DOI">10.3390/electronics12244957</idno>
		<ptr target="https://doi.org/10.3390/electronics12244957" />
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">24</biblScope>
			<biblScope unit="page">4957</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Deep semisupervised learning-based network anomaly detection in heterogeneous information systems</title>
		<author>
			<persName><forename type="first">N</forename><surname>Lutsiv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Maksymyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beshley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Lavriv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Andrushchak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vokorokos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gazda</surname></persName>
		</author>
		<idno type="DOI">10.32604/cmc.2022.018773</idno>
		<ptr target="https://doi.org/10.32604/cmc.2022.018773" />
	</analytic>
	<monogr>
		<title level="j">Comput., Mater. &amp; Contin</title>
		<imprint>
			<biblScope unit="volume">70</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="413" to="431" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Intelligent nodes for distributed sensor network</title>
		<author>
			<persName><forename type="first">A</forename><surname>Sachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kochan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Turchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tymchyshyn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Vasylkiv</surname></persName>
		</author>
		<idno type="DOI">10.1109/IMTC.1999.776072</idno>
		<ptr target="https://doi.org/10.1109/IMTC.1999.776072" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th IEEE Instrumentation and Measurement Technology Conference (Cat. No.99CH36309)</title>
				<meeting>the 16th IEEE Instrumentation and Measurement Technology Conference (Cat. No.99CH36309)<address><addrLine>Venice, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1479" to="1484" />
		</imprint>
	</monogr>
	<note>IMTC/99</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Analysis of the developed quantitative method for automatic attribution of scientific and technical text content written in Ukrainian</title>
		<author>
			<persName><forename type="first">V</forename><surname>Lytvyn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vysotska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pukach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Nytrebych</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Demkiv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Senyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Malanchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sachenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kovalchuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Huzyk</surname></persName>
		</author>
		<idno type="DOI">10.15587/1729-4061</idno>
		<idno>.2018.149596</idno>
		<ptr target="https://doi.org/10.15587/1729-4061" />
	</analytic>
	<monogr>
		<title level="j">Eastern-European J. Enterp. Technol</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="19" to="31" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A methodology for machine-learning content analysis to define the key labels in the titles of online customer reviews with the rating evaluation</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Zaki</forename><surname>Ahmed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Rodríguez</forename><surname>Díaz</surname></persName>
		</author>
		<idno type="DOI">10.3390/su14159183</idno>
		<ptr target="https://doi.org/10.3390/su14159183" />
	</analytic>
	<monogr>
		<title level="j">Sustainability</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page">9183</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Research on the TF-IDF algorithm combined with semantics for automatic extraction of keywords from network news texts</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1515/jisys-2023-0300</idno>
		<ptr target="https://doi.org/10.1515/jisys-2023-0300" />
	</analytic>
	<monogr>
		<title level="j">J. Intell. Syst</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page">1</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Named entity recognition (NER) and relation extraction in scientific publications</title>
		<author>
			<persName><forename type="first">A</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garg</surname></persName>
		</author>
		<idno type="DOI">10.35940/ijrte.b7846.0712223</idno>
		<ptr target="https://doi.org/10.35940/ijrte.b7846.0712223" />
	</analytic>
	<monogr>
		<title level="j">Int. J. Recent Technol. Eng. (IJRTE)</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="110" to="113" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">LNLF-BERT: transformer for long document classification with multiple attention levels</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">M</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">C</forename><surname>The</surname></persName>
		</author>
		<idno type="DOI">10.1109/access.2024.3492102</idno>
		<ptr target="https://doi.org/10.1109/access.2024.3492102" />
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Sentiment classification: review of text vectorization methods: bag of words, tf-idf, word2vec and doc2vec</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">D</forename><surname>Abubakar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Umar</surname></persName>
		</author>
		<idno type="DOI">10.56471/slujst.v4i.266</idno>
		<ptr target="https://doi.org/10.56471/slujst.v4i.266" />
	</analytic>
	<monogr>
		<title level="j">SLU J. Sci. Technol</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="27" to="33" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Implementation of generative AI using metaverse-based LLM</title>
		<author>
			<persName><forename type="first">H.-S</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-S</forename><surname>Shim</surname></persName>
		</author>
		<idno type="DOI">10.29279/jitr.2024.29.2.123</idno>
		<ptr target="https://doi.org/10.29279/jitr.2024.29.2.123" />
	</analytic>
	<monogr>
		<title level="j">Korea Ind. Technol. Converg. Soc</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="123" to="132" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gruen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Maldoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Messing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sanderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zimmer</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2410.23432</idno>
		<ptr target="https://doi.org/10.48550/arXiv.2410.23432" />
		<title level="m">Web scraping for research: legal, ethical, institutional, and scientific considerations</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Forward learning of large language models by consumer devices</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Pau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Aymone</surname></persName>
		</author>
		<idno type="DOI">10.3390/electronics13020402</idno>
		<ptr target="https://doi.org/10.3390/electronics13020402" />
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page">402</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Large language models meet next-generation networking technologies: A review</title>
		<author>
			<persName><forename type="first">C.-N</forename><surname>Hang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-D</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morabito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-W</forename><surname>Tan</surname></persName>
		</author>
		<idno type="DOI">10.3390/fi16100365</idno>
		<ptr target="https://doi.org/10.3390/fi16100365" />
	</analytic>
	<monogr>
		<title level="j">Future Internet</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="issue">10</biblScope>
			<biblScope unit="page">365</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Algorithms of parallel calculations in task of tolerance ellipsoidal estimation of interval model parameters</title>
		<author>
			<persName><forename type="first">M</forename><surname>Dyvak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stakhiv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pukas</surname></persName>
		</author>
		<idno type="DOI">10.2478/v10175-012-0022-9</idno>
		<ptr target="https://doi.org/10.2478/v10175-012-0022-9" />
	</analytic>
	<monogr>
		<title level="j">Bull. Pol. Acad. Sci</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="issue">1</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Dyvak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Voytyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Porplytsya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pukas</surname></persName>
		</author>
		<idno type="DOI">10.1109/TCSET.2018.8336426</idno>
		<ptr target="https://doi.org/10.1109/TCSET.2018.8336426" />
		<title level="m">2018 14th international conference on advanced trends in radioelecrtronics, telecommunications and computer engineering</title>
				<meeting><address><addrLine>TCSET)</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1272" to="1276" />
		</imprint>
	</monogr>
	<note>Modeling the process of air pollution by harmful emissions from vehicles</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Features of structure identification the macromodels for nonstationary fields of air pollutions from vehicles</title>
		<author>
			<persName><forename type="first">N</forename><surname>Ocheretnyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Voytyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dyvak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Martsenyuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of international conference on modern problem of radio engineering, telecommunications and computer science</title>
				<meeting>international conference on modern problem of radio engineering, telecommunications and computer science</meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="444" to="444" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Parameters identification method of interval discrete dynamic models of air pollution based on artificial bee colony algorithm</title>
		<author>
			<persName><forename type="first">M</forename><surname>Dyvak</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACIT49673.2020.9208972</idno>
		<ptr target="https://doi.org/10.1109/ACIT49673.2020.9208972" />
	</analytic>
	<monogr>
		<title level="m">2020 10th international conference on advanced computer information technologies (ACIT)</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="130" to="135" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Selection the &quot;saturated&quot; block from interval system of linear algebraic equations for recurrent laryngeal nerve identification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Dyvak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pukas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Oliynyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Melnyk</surname></persName>
		</author>
		<idno type="DOI">10.1109/DSMP.2018.8478528</idno>
		<ptr target="https://doi.org/10.1109/DSMP.2018.8478528" />
	</analytic>
	<monogr>
		<title level="m">IEEE second international conference on data stream mining &amp; processing (DSMP)</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="444" to="448" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Content analysis using specific natural language processing methods for big data</title>
		<author>
			<persName><forename type="first">M</forename><surname>Pirnau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Botezatu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Priescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hosszu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tabusca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Coculescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Oncioiu</surname></persName>
		</author>
		<idno type="DOI">10.3390/electronics13030584</idno>
		<ptr target="https://doi.org/10.3390/electronics13030584" />
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page">584</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Illustration of a software-aided content analysis methodology applied to educational research</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gkevrou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Stamovlasis</surname></persName>
		</author>
		<idno type="DOI">10.3390/educsci12050328</idno>
		<ptr target="https://doi.org/10.3390/educsci12050328" />
	</analytic>
	<monogr>
		<title level="j">Educ. Sci</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page">328</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Content analysis of three-dimensional model technologies and applications for construction: current trends and future directions</title>
		<author>
			<persName><forename type="first">N</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sturgill</surname></persName>
		</author>
		<idno type="DOI">10.3390/s24123838</idno>
		<ptr target="https://doi.org/10.3390/s24123838" />
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page">3838</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
