<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Analysis of Recommendation System Methods for Accuracy of Predicted Estimates</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nataliya Boyko</string-name>
          <email>nataliya.i.boyko@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Petro Telishevskyi</string-name>
          <email>petro.telishevskyi.kn.2017@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Beata Kushka</string-name>
          <email>beatakushka@yahoo.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Data Mining, Recommended System</institution>
          ,
          <addr-line>Algorithms, SVD, method, Item-Based Method, NMF</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Hot Deck method, the Barlet method, the Resampling algorithms</institution>
          ,
          <addr-line>Zet, Zetbraid, EM estimation</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>Profesorska Street 1, Lviv, 79013</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The Internet plays a major role in people's lives today. In the network, people communicate with friends, meet new people, develop themselves, buy goods and services, and spend time on entertainment, namely, watching movies, listening to music and so on. Because there is enough content on the Internet, it means that it is difficult for people to physically choose what they want at the moment. Therefore, web services use different levels of referral systems. Because, recommendation systems help us choose from a thousand not a lot of content that interests us, but what is not interesting to reject. Recommendation systems have been implemented for a long time, but have only recently been developed and applied. Namely, with the active development of the Internet. The most successful recommendation systems are systems based on collaborative filtering. Investigate the collaborative filtering, namely, User-based, Item-based and SVD. Conducting experimental studies on the data using methods for recommendation. A comparative characterization between the two methods after the experiments.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>RMSE</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>One of the manifestations of information uncertainty is uncertainty caused by data gaps. The
objective characteristics of certain processes can be changed or even distorted due to the loss of some
data during their receipt, transmission or storage [3,8]. There is a need to recover such missed data
and, importantly, to select the algorithms by which they will be recovered, because incorrect or
insufficiently reliable recovery can cause more damage than the data gaps themselves [1, 5].</p>
      <p>There are algorithms that allow you to process gaps with the necessary information, such as the</p>
      <p>2021 Copyright for this paper by its authors.
 Amazon sells 35% of all products with recommendations.</p>
      <p>The most popular are systems that work on the model of explicit feedback (explicitfeedback), their
main essence is that users leave reviews about a product or service and based on these reviews,
recommendations are built in this work, we will explore a different approach, namely
implicitfeedback. This model allows based on user behavior to predict the user's attitude to a
particular product or service. As a result, we can automatically generate a content rating and use it for
other users.</p>
      <p>A recommended system is a system that recommends content to us among the vast amount of
information from our needs. Users who use the resource, where there are recommendations, determine
their interests using estimates [6, 10].</p>
      <p>To create resources for watching videos, to store preferences, recommendation systems use user
profiles, because these profiles store an estimate of the content of the resource. When a user evaluates,
new recommendations for the user are calculated and compared with his profile [3, 12].</p>
      <p>Ratings can be explicit or implicit. The evaluation that the user makes on the content that interests
him is called explicit. And when the benefits are made on purchases, on the pages to which users
went, it is an implicit rating.</p>
      <p>Most recommendation systems use one of three approaches:
 Collaborative filtering;
 Content-based filtering;
 Hybrid filtration.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Methods</title>
      <p>To date, there are many technologies and tools for analysis, classification. These include machine
learning, BigData, Natural Language Processing (NLP), and referral systems. In many cases, such
systems are created to predict and analyze user behavior.</p>
      <p>Recommendation systems are designed to simplify and improve, speed up the user's search for the
necessary content. They are very important, because they allow the user to interact only with the
content that may interest him, and therefore increase the efficiency of the system.</p>
      <p>Wang P. said that: "Recommendation systems are systems that aim to select specific objects that
meet user requirements, where each of these objects is stored in computer memory and is
characterized by a set of attributes."</p>
      <p>Using specially collected information, recommendation systems can predict whether the user is
interested in this content, rank. Or choose a specific set of N items that may be of interest to the user.</p>
      <p>From these statements there are two problems which solve systems of recommendations. The first
problem is the problem of predicting when it is necessary to predict whether this content will be
interesting. As well as the problem (top-N) of choosing a set of data that might interest the user. This
helps companies increase profits by reducing purchases of goods that will not be sold any time soon.</p>
      <p>Modern recommendation systems provide high forecasting accuracy, but only if there is sufficient
information about the user and his preferences. When a user does not provide enough information
about himself, the system is unable to make the correct prediction</p>
      <p>Lack of sufficient information can lead to the following problems.</p>
      <p>The problem of similarity arises when it is difficult to determine the similarity between users
because the number of features about the user is less than necessary for a quality recommendation.
Researcher Yu also made an interesting conclusion about this problem, that in fact the content suffers
from this problem no less than users. Because they typically use fewer features which is often not
enough for popular algorithms such as feature-based, content-based algorithms.</p>
      <p>A Cold Start Problem occurs when a newly registered user has not yet rated any content and the
system has no information about it.</p>
      <p>The problem of changing the taste. For example, today the buyer is looking for a product for
himself and tomorrow he is looking for a gift for his mother, so the user may have incorrect ideas
about his tastes.</p>
      <p>The problem of new things happens when the system does not know about specific content
because no one has evaluated it yet.</p>
      <p>Unpredictable things depend on the user's preferences, especially in music, it is difficult for the
system to evaluate such things, because everyone can have their own reaction to this content.</p>
      <p>Basic algorithms in recommendation systems:
 Content based</p>
      <p>This approach recommends that users use similar things to those they have liked in the past.
Keyword search is often used to find similar things. In this case, the company Pandora music online
service in the framework of MUSIC GENOME PROJECT to create for each of its songs a vector
consisting of approximately 450 features. This allowed the use of more standard machine learning
approaches that gave the output the likelihood that the user would like the music.</p>
      <p> Demographic based</p>
      <p>Another approach is when there is a lot of information about users such as Facebook, Linkedin.
Therefore, you can use this information to set the direction of recommendations regardless of
previous user behavior.</p>
      <p> Coloborative filter</p>
      <p>This method is based on user-content interaction. It only analyzes the rating and ignores
information about the content or user. The key idea is that such users like similar things and if users
have watched the same movies, there is a high probability that in the future they will also get the same
recommendation. All you need is some kind of assessment that arises from the interaction of user and
content.</p>
      <p>There are two types of such data:
 Explicit - this is when the available rating matches the rating or preferences. Popular services
offer users to evaluate things that interest them in different ways. For example Netflix, Youtube
use the like / dislike system. At the same time, Amazon and Aliexpress trading platforms use a
system of stars, the maximum number of which is 10.
 Implicit is the process by which a service collects information about a user's interaction with
content with information about clicks, views, or purchases. It also takes into account the time: how
much the user spent on the page and many other factors.</p>
      <p>A striking example is online cinemas, where each film has a certain rating. It allows you to see in
numerical terms the user's preferences for the film. The only problem is that some users do not leave
any feedback, so the recommendation system may be inaccurate for this and implicit evaluation is
used because it allows you to increase the amount of information about the user</p>
      <p>Currently, many commercial pages on the Internet have their own recommendation pages. Ahead
of others are sites such as Netflix, Amazon, Google, Linkedin. Many researchers also presented their
versions of recommendation systems.</p>
      <p>Today, there is a lot of research on referral systems, but they all face the problem of insufficient
data. And implicit evaluation allows you to improve these results, because it does not require any
additional action from the user.</p>
      <p>
        There are several major problems when working with referral systems. Problems (
        <xref ref-type="bibr" rid="ref1 ref2">1-2</xref>
        ) were
encountered at the beginning of the development of recommendation systems. Now they are
successfully solved within the framework of modern algorithms.
      </p>
      <p>1. The problem of similarity. A hybrid algorithm is used to solve this problem. Due to which we
do not always need information about the user to give an accurate and adequate recommendation.
2. The problem of lack of data. It is solved by using implicit estimation, and also colloborative
algorithm. This is usually sufficient for recommendations with high accuracy.</p>
      <p>The main problem is the problem of cold start.
3. Cold start problem. Consider a problem with new content that has not yet received enough
ratings to be recommended. This problem can be solved thanks to the content-based algorithm,
hybrid algorithm. Due to certain categories that will be when mastering the content it is possible to
show it along with similar content. However, not every content is labeled with enough classes or
certain features to be used in recommendations. In this case, text mining uses several basic
algorithms.</p>
      <p>Here is an example based on the recommendations of websites.
1. Preprocessing stage
In this case, highlight the keywords, highlight the theme of the page, discard the stop words.
2. TF-IDF (TF — term frequency, IDF — inverse document frequency)</p>
      <p>At this stage, the calculation is performed for each word w in each document d. tf (w, d ) 
where nw d - the total number of word entries in the document, nd - the total number of words in the
document. This method is used to discard redundant information and not store it in memory. Because
in the relevance feedback algorithm, only 200 words with maximum weight will suffice.
3. Relewant feedback</p>
      <p>This algorithm brings us to the problem. It is based on page ratings that the user likes but without
overall page ratings. But it is usually used for new content, because it loses in efficiency to more
traditional algorithms.</p>
      <p>The first step is to find a TF-IDF for sites that the user likes (if the user has just started using the
service, it is good practice to let him mark several sites (movies, books,…) that he visited and that he
liked).</p>
      <p>The second step is to find the similarity of the page to the user's preferences, which is calculated as
a scalar product of the vectors of user weights and sites k (u, d )  
wWU</p>
      <p>Vww * tfdf (w, d ) , where
WU - words from the user profile. As a result, this algorithm allows you to explore the similarity of
any page that has text to any user who has certain preferences.</p>
      <p>The problem with the cold start for the user is that the system meets the user for the first time and
there is no information about him in his memory. This problem can occur constantly. The first
example is when a user has not registered and information about him is stored in cookies, but users
always have the opportunity to delete their information. The system will then consider them as new
users. The second problem arises when a user searches for goods for someone. For example, the user
went to the site in search of a particular product and the relevant recommendations he needs only until
the time of purchase. And then he will need another product. As a result, the user remains the same,
but his tastes can be radically different.</p>
      <p>The most common solution to this problem is to use geolocation. When a user has just registered
or is using the service for the first time, we will show him information that is popular in his area of
residence at a certain point in time. After a few likes or interactions with the content, the system will
be able to show more accurate suggestions.
,
2.1.</p>
    </sec>
    <sec id="sec-4">
      <title>User-based method</title>
      <p>A method that is based on the user liking products that have been selected by users similar to him
(Formula 1) [11, 13].</p>
      <p>
        1
rˆui  ri  iR(u) sim(i, i ) iR(u) sim(i, i)(ru,i  ri ). (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
      </p>
      <p>The degree of similarity sim(i,i) is calculated from the matrix of estimates R. The most
commonly used similarity metric is the Pearson correlation and the cosine distance of the rows
(columns) of the matrix (Formula 2-3).</p>
      <p>sim(u, u) </p>
      <p>iR(u)R(u) (ru,i  r u )(ru,i  r u )
iR(u)R(u) (ru,i  r u )2 iR(u)R(u) (ru,i  r u )2
.
sim(u, u) </p>
      <p>iR(u)R(u) (ru,i ru,i )
iR(u)R(u) ru,i 2 iR(u)R(u) ru
.</p>
      <p>
        (
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
      </p>
    </sec>
    <sec id="sec-5">
      <title>Item-based method</title>
      <p>rˆui  r i </p>
      <p>iR(u) sim(i, i)</p>
      <p>A method that is based on the fact that the user will like products similar to those he has already
chosen.</p>
      <p>User A is characterized by objects that he has viewed or rated. For each of the selected / rated
objects, m neighboring objects are defined, ie there are m most similar objects in terms of user views /
ratings. When building a PC for movies, m takes values from 10 to 30. All neighboring objects are
combined into a collection, which excludes objects viewed or evaluated by user A. And from the
remains of the collection is built top n recommendations. Thus, in the item-based approach, all users
who liked this or that object from the collection take part in creating
recommendations (Formula 4) [14, 15].</p>
      <p>1</p>
      <p>iR(u) sim(i, i)(ru,i  r i ).
2.3.</p>
    </sec>
    <sec id="sec-6">
      <title>SVD algorithm</title>
      <p>Almost all collaborative filtering algorithms have such shortcomings as cold start, triviality of
recommendation results, and so on. One of the fairly new algorithms that reduces the impact of
typical collaborative filtering problems was the SVD algorithm, which was created to improve the
results of conventional algorithms [16, 17].</p>
      <p>SVD is a method of factorization of matrices, which is usually used to reduce the number of data
set functions by reducing the size of space from n to k, where k &lt;n. However, for the purposes of
recommendation systems, we are only interested in matrix factorization, where parts retain the same
dimension. Matrix factorization is performed on the matrix of user position ratings (Formula 5).
Matrix factorization can be considered as a search for 2 matrices, the product of which is the original
matrix (Formula 6) [12, 15].</p>
      <p>expected rating  rˆui  qiT pu ,
when qi and pi can be found so that the difference of the square errors between their product points
and the known rating in the custom element matrix is minimal.</p>
      <p>minimum ( p, q)(u,i)K (rui  qiT pu )2 .
2.4.</p>
    </sec>
    <sec id="sec-7">
      <title>NMF algorithm</title>
      <p>NMF algorithm is a representation of the matrix V in the form of the product of the matrices W and
H, in which all the elements of the three matrices are non-negative [9]. Let table V have size m x n.
Denote by r the rank of the matrices W and H, as a rule r &lt;&lt; min (n, m). In contrast to the exact
representation of the matrix in SVD and NMF, we obtain only an approximate equality (Formula 7)
[3, 12].</p>
      <p>
        V  WH. (
        <xref ref-type="bibr" rid="ref7">7</xref>
        )
      </p>
      <p>The matrices W and H are chosen so as to minimize the loss function: D(V ,WH )  min . In this
case, D is given on the basis of Kulbak-Leibler divergence (Formula 8) [13, 17].</p>
      <p>
        a
D( A, B)  i, j aij log( ij )  aij bij . (
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
bij
      </p>
    </sec>
    <sec id="sec-8">
      <title>3. Experiments</title>
      <p>
        Experiments use the MovieLense100k open access data set. This data set includes detailed
information about users, movies and information about users and their ratings on movies.
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
In this paper, we will use certain fields from this data set:
 User ID;
 Film ID;
 User rating for this movie.
      </p>
      <p>This data set contains information about 1000 users, 1700 movies and 100000 user ratings for
these movies. Table 1 will show the data structure.
where N – total number of assessments;
r
i – projected assessment;
ˆ
r
i – the assessment that was made.</p>
      <p>At the end of the experiments, the results of the image on the graph, where the x-axis will describe
the test number, and the y-axis - the RMSE value for a method.</p>
      <p>For experiments, use the methods mentioned in the previous section:
 User-based Collaborating filter(CF), Item-based CF;
 SVD, NMF.</p>
      <p>The results are shown in Table 2 for these methods that worked with the ‘MovieLense 100k’ data
set. The table records the values of five times the cross-calibration, and then shows the average RMSE
value for each algorithm.
For a comparative result, we show the results of all methods in Figure 1.</p>
      <p>
        (
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
      </p>
      <sec id="sec-8-1">
        <title>Average</title>
        <p>RMSE
1.010936
1.041222
0.936784
0.964507</p>
        <p>As can be seen from the graph, in all tests the accuracy relative to RMSE is the best in the SVD
algorithm. During all tests for this algorithm, the difference in accuracy is not large compared to the
rest of the algorithms.</p>
        <p>The average results show the NMF method, although the results of this algorithm are similar to the
SVD algorithm.</p>
        <p>The average results show the NMF method, although the results of this algorithm are similar to the
SVD algorithm. This is determined by the fact that the presented algorithms use the decomposition of
the matrix.</p>
        <p>The worst results are User-base and Item-based, due to the fact that we use Pearson's correlation to
find similar data. Sometimes the data does not correlate with the required data, which gives worse
accuracy.</p>
        <p>We will also conduct experiments for User-based and Item-based using different types of
correlation. Namely, we use Pearson correlation, Cosine, Mean Squared Difference (MSD), Pearson
baseline. All checks will be performed on our data from MovieLense 100K. Mean Abslolute Error
was used to show which of the algorithms for the User-based and Item-based methods would be the
best.</p>
        <p>Table 3 and Figure 2 will show the results of experiments for User-based. The table records the
values of five times the cross-calibration, and then shows the average value of the absolute error for
each measure of similarity in the User-based method.</p>
      </sec>
      <sec id="sec-8-2">
        <title>Fold 1</title>
        <p>0.8096
0.7800
0.8051</p>
        <p>Figure 2 shows that after using MSD, we get much better results. Therefore, this measure of
similarity should be used, not correlation or cosine, because in such cases for the recommendation
system using the User-based method with MSD the lowest values will be obtained, and they
negatively affect the correctness of the recommendations. Also, the execution time in MSD is the
lowest and this is shown in Table 4.</p>
        <p>As with User-based in Item-based, a measure of MSD similarity should be used. MAE scores are
the lowest. They have a positive effect on the system. Runtime is fastest in MSD and slowest in
Pearson baseline. The final results are shown in Table 6.</p>
        <p>To understand how the SVD and NMF algorithm will behave on the proposed dataset
MovieLense 100K, need training data and tests to be divided into 5 parts. Then it is necessary to
determine with what accuracy for each user the Top 10 recommendations are defined. They are
determined by the Formula 10.</p>
        <p>Recommende d items that are relevant 
.</p>
        <p>
          (
          <xref ref-type="bibr" rid="ref10">10</xref>
          )
| {Recommend ed items} |
        </p>
        <p>The data shown in Table 7, where the numerical indicator is the average accuracy of the sum of the
values of each user for the total number. Accuracy will be displayed as a percentage.</p>
        <p>Thus, from Table 7 results in all parameters equal to approximately 61%. Therefore, the top 10
recommendations will be true for users if you use SVD or NMF algorithms on a data set with
MovieLense 100K.</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>4. Discussion of the results of experiments</title>
      <p>Experiments show that SVD calculations will be better than User-based and Item-based. Because
User-based and Item-based use correlation to find similar data, it doesn't always give us what we
need.</p>
      <p>The disadvantage of User-based and Item-based CFs do not fit into real scenarios and do not solve
the known problem of cold start, ie when a new user or a new element enters the system.</p>
      <p>It is said that cold start affects User and Item -based CF, ie increases RMSE for these methods.</p>
      <p>And we can also assume that the algorithms described in this paper can give a better estimate if
you increase the amount of data.</p>
      <p>Let's list the disadvantages of correlation methods:
 Cold start problem;
 Poor forecasts for new / atypical users / facilities;
 Triviality of recommendations;
 Resource-intensive calculations. In order to make assumptions, we need to keep in mind all
the estimates of all users.</p>
      <p>And the second limitation may be the data set we used. Because the data provided by MovieLense
was used for this study. And in order to see how our tonicity changes, it is desirable to use different
data sets.</p>
      <p>In order to improve the accuracy of the assessment should combine methods described methods of
filtering content. This means the formation of a hybrid approach. Hybridity will consist of a
combination of two different approaches. This will avoid the disadvantages of collaborative filtering
and content filtering. There are several approaches:
 Inclusion of element characteristics in collaborative filtering;
 Construction of a single model-based collaborative filtering and content filtering.</p>
    </sec>
    <sec id="sec-10">
      <title>5. Conclusions</title>
      <p>In an information-saturated world, referral systems play an important role in helping users interact
only with information that is potentially of interest to them. Recommendation systems are a leading
indicator of predicting user behavior in systems with a large amount of information. The systems
provide the applications in which they are used with valuable customer offerings, increasing profits
from business start-ups. Such a system helps the customer to provide the relevant goods, while
generating revenue from the company. The system must also provide the newly added users with
enough offers to meet the demand of the users. One way to increase forecasting accuracy is to
encourage the user to evaluate the quality of the movies, thereby increasing the number of offers.</p>
      <p>This study showed that the SVD algorithm provides better estimates than the described methods,
which are based on correlation. It can also be said that it is advisable to use the User-based and
Itembased CF methods independently, because the results of the other two algorithms are better. The
described methods should be used in hybrid approaches. Further studies should take into account
other parameters that affect the accuracy results.</p>
    </sec>
    <sec id="sec-11">
      <title>6. References</title>
      <p>
        [12] N. Boyko, B. Mandych, Technologies of Object Recognition in Space for Visually Impaired
People, in: Proceedings of The 3rd International Conference on Informatics &amp; Data-Driven
Medicine (IDDM 2020), Växjö, Sweden, November 19-21, 2020, pp. 338-347.
[13] R. Agrawal, J. Gehrke, D. Gunopulos, P. Raghavan, Automatic sub-space clustering of high
dimensional data, Data mining knowledge discovery, vol. 11(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), 2005, pp. 5–33.
[14] V. Estivill-Castro, I. Lee, Amoeba: Hierarchical clustering based on spatial proximity using
Delaunay diagram, in: Proceedings of the 9th Intern. Symp. on spatial data handling, Beijing,
China, 2000, pp. 26–41.
[15] I. Turton, S. Openshaw, C. Brunsdon, Testing space time and more complex hyperspace
geographical analysis tools, Innovations in GIS 7, Taylor &amp; Francis, London, 2000, pp. 87–100.
[16] N. Boyko, N. Tkachuk, Processing of Medical Different Types of Data Using Hadoop and Java
MapReduce, in: Proceedings of The 3rd International Conference on Informatics &amp; Data-Driven
Medicine (IDDM 2020), Växjö, Sweden, November 19-21, 2020, pp. 405-414.
[17] C. Aggarwal, P. Yu, Finding generalized projected clusters in high dimensional spaces, in:
Proceedings of the Intern. conf. on management of data ACM SIGMOD, 2000, pp. 70–81.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G.</given-names>
            <surname>Tak</surname>
          </string-name>
          <article-title>´acs, I. Pil´aszy, B</article-title>
          . N´emeth, D. Tikk,
          <article-title>Matrix factorization and neighbor based algorithms for the netflix prize problem</article-title>
          ,
          <source>in: Proceedings of the 2008 ACM conference on Recommender systems, ACM</source>
          ,
          <year>2008</year>
          , pp.
          <fpage>267</fpage>
          -
          <lpage>274</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B.</given-names>
            <surname>Sarwar</surname>
          </string-name>
          , G. Karypis,
          <string-name>
            <given-names>J.</given-names>
            <surname>Konstan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Riedl</surname>
          </string-name>
          ,
          <article-title>Sarwar Item-Based Collaborative Filtering Recommendation Algorithms</article-title>
          ,
          <source>in: Proceedings 10th Int'l WWW Conf</source>
          ,
          <year>2001</year>
          , pp.
          <fpage>285</fpage>
          -
          <lpage>295</lpage>
          . https://dl.acm.org/doi/10.1145/1454008.1454049.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I.</given-names>
            <surname>Soboroff</surname>
          </string-name>
          , C. Nicholas,
          <article-title>Combining Content and Collaborationin Text Filtering</article-title>
          ,
          <source>in: Proceedings of the Int'l Joint Conf. Artificial Intelligence Workshop: Machine Learning for Information Filtering</source>
          ,
          <year>1999</year>
          , pp.
          <fpage>86</fpage>
          -
          <lpage>91</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <article-title>Recommender systems</article-title>
          .
          <article-title>Introduction to approaches and algorithms</article-title>
          .,
          <year>2017</year>
          . URL: https://www.ibm.com/developerworks/library/os-recommender1/
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Boehm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kailing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kroeger</surname>
          </string-name>
          ,
          <article-title>Density connected clustering with local subspace preferences</article-title>
          ,
          <source>in: Proceedings of the 4th IEEE Intern. conf. on data mining, IEEE Computer Society</source>
          , Los Alamitos,
          <year>2004</year>
          , pp.
          <fpage>27</fpage>
          -
          <lpage>34</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Boyko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Boksho</surname>
          </string-name>
          ,
          <article-title>Application of the Naive Bayesian Classifier in Work on Sentimental Analysis of Medical Data</article-title>
          ,
          <source>in: Proceedings of The 3rd International Conference on Informatics &amp; Data-Driven Medicine (IDDM</source>
          <year>2020</year>
          ), Växjö, Sweden,
          <source>November 19-21</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>230</fpage>
          -
          <lpage>239</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Harel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Koren</surname>
          </string-name>
          ,
          <article-title>Clustering spatial data using random walks</article-title>
          ,
          <source>in: Proceedings of the 7th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</source>
          , San Francisco, California,
          <volume>200</volume>
          , pp.
          <fpage>281</fpage>
          -
          <lpage>286</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.J.</given-names>
            <surname>Peuquet</surname>
          </string-name>
          ,
          <article-title>Representations of space and time</article-title>
          , Guilford Press, New York, NY,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.-Y.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.-J.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-J. Li</surname>
          </string-name>
          ,
          <article-title>P2P Spatial query processing by Delaunay triangulation</article-title>
          ,
          <source>Lecture notes in computer science</source>
          , vol.
          <volume>3428</volume>
          , Springer/Heidelberg,
          <year>2005</year>
          , pp.
          <fpage>136</fpage>
          -
          <lpage>150</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ankerst</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <article-title>Towards an effective cooperation of the user and the computer for classification</article-title>
          ,
          <source>in: Proceedings of the 6th ACM SIGKDD Intern. conf. on knowledge discovery and data mining</source>
          , Boston, Massachusetts, USA,
          <year>2000</year>
          , pp.
          <fpage>179</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>С</surname>
          </string-name>
          . Zhang, Y. Murayama,
          <article-title>Testing local spatial autocorrelation using</article-title>
          , vol.
          <volume>14</volume>
          ,
          <string-name>
            <surname>Intern</surname>
          </string-name>
          .
          <source>J. of Geogr. Inform. Science</source>
          ,
          <year>2000</year>
          , pp.
          <fpage>681</fpage>
          -
          <lpage>692</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>