=Paper= {{Paper |id=Vol-2161/paper36 |storemode=property |title=Ontology-based Linked Data Summarization in Semantics-aware Recommender Systems |pdfUrl=https://ceur-ws.org/Vol-2161/paper36.pdf |volume=Vol-2161 |authors=Vito Walter Anelli,Tommaso Di Noia,Andrea Maurino,Matteo Palmonari,Anisa Rula |dblpUrl=https://dblp.org/rec/conf/sebd/AnelliNMPR18 }} ==Ontology-based Linked Data Summarization in Semantics-aware Recommender Systems== https://ceur-ws.org/Vol-2161/paper36.pdf
    Using Ontology-based Data Summarization to Develop
          Semantics-aware Recommender Systems?

      Vito Walter Anelli1 , Tommaso Di Noia1 , Andrea Maurino2 , Matteo Palmonari2 ,
                                      Anisa Rula2
                1
                 Polytechnic University of Bari, Via Orabona, 4, 70125 Bari, Italy
                            firstname.lastname@poliba.it
        2
          University of Milano-Bicocca, Piazza dell’Ateneo Nuovo, 1, 20126 Milano, Italy
                            firstname.lastname@unimib.it



        Abstract. In the current information-centric era, recommender systems are gaining
        momentum as tools able to assist users in daily decision-making tasks. Within the
        recommendation process, Linked Data have been already proposed as a valuable
        source of information to enhance the predictive power of recommender systems but
        an open issues is still related to feature selection of the most relevant subset of data
        in the whole semantic web. In this paper, we show how ontology-based (linked)
        data summarization can drive the selection of properties/features useful to a rec-
        ommender system. In particular, we compare a fully automated feature selection
        method based on ontology-based data summaries with more classical ones, and we
        evaluate the performance of these methods in terms of accuracy and aggregate di-
        versity of a recommender system exploiting the top-k selected features.



1     Introduction
Semantics-aware Recommender Systems (RSs) exploiting information held in knowledge
graphs, as the ones available as Linked Data (LD), represent one of the most interesting
and challenging application scenarios for LD [6]. A high number of solutions and tools
have been proposed in the last years showing the effectiveness of adopting LD as knowl-
edge sources to feed a recommendation engine (see [5] and references therein for an
overview). Nevertheless, how to automatically select the “best” subset of a LD dataset to
feed a LD-based RS without affecting the performance of the recommendation algorithm
is still an open issue. Notice that the selection of the top-k features to use in a RSs means
to discover which properties in a LD-dataset (e.g., DBpedia) encode the knowledge useful
in the recommendation task and which ones are just noise [20]. In most of the approaches
proposed so far, usually, the FS process is performed by human experts that manually
choose properties resulting more “suitable” for a given scenario. Over the years, many
algorithms and techniques for feature selection , e.g., Information Gain, Information Gain
Ratio, Chi Squared Test and Principal Component Analysis, have been proposed with ref-
erence to machine learning tasks but they do not consider a characteristic which makes
unique LD: they come with semantics attached.
     The main objective of this paper is to investigate how ontology-based data summa-
rization [28] can be used as a new and semantic-oriented feature selection technique for
?
    Discussion Paper. An extended version of this work has been presented at [7].
    SEBD 2018, June 24-27, 2018, Castellaneta Marina, Italy. Copyright held by the author(s).
LD-based RSs. We define a feature selection method that automatically extracts the top-k
properties that are deemed to be more important to evaluate similarity between instances
of a given class on top of data summaries built with the help of an ontology. We per-
form an experimental evaluation on three well-known datasets in the RS domain (Movie-
lens, LastFM, LibraryThing) in order to analyze how the choice of a particular FS tech-
nique may influence the performance of recommendation algorithms in terms of typical
accuracy and diversity metrics. Experimental results show that information provided in
ontology-based data summaries selects features that achieve comparable, or, in most of
the cases, better performance than state-of-the-art, semantic-agnostic analytical methods
such as Information Gain [19]. We believe that these results are interesting also because
of practical reasons. LD summaries are published on-line and summary-based FS can be
performed even without acquiring the entire dataset and efficiently (on top of summary
information). The use of frequency associated with schema patterns in a FS approach was
initially tested in a previous work[24].
     The paper is organized as follows: in Section 2, we introduce the ontology-based data
summarization approach used in this work, while in Section 3, we describe the feature
selection and recommendation methods. Section 4 is devoted to the explanation and dis-
cussion of the experimental results. Section 5 briefly reviews related literature for schema
and data summarization as well as on recommender systems while Section 6 discuss con-
clusions and future work.
2     Ontology-driven Linked Data Summarization
While relevance-oriented data summarization approaches are aimed at finding subsets of a
dataset or an ontology that are estimated to be more relevant for the users [30], vocabulary-
oriented approaches are aimed at profiling a dataset, by describing the usage of vocabular-
ies/ontologies used in the dataset. The summaries returned by these approaches are com-
plete, i.e., they provide statistics about every element of the vocabulary/ontology used in
the dataset [28]. Statistics captured by these summaries that can be useful for the feature
selection process are the ones concerning the usage of properties for a certain class of
items to recommend.
    Patterns and frequency. In our approach we use pattern-based summaries ex-
tracted using the ABSTAT framework. Pattern-based summaries describe the content
of a dataset using schema patterns having the form hC, P, Di, where C and D, are
types (either classes or datatypes) and P is a property. For example, the pattern
hdbo:Film, dbo:starring, dbo:Actori tells that films exist in the dataset, in
which star some actors. Differently from similar pattern-based summaries [17], AB-
STAT uses the subclass relations in the data ontology, represented in a Type Graph, to
extract only minimal type patterns from relational assertions, i.e, the patterns that are
more type-wise specific according to the ontology. A pattern hC, P, Di is a minimal type
pattern for a relational assertion ha, P, bi according to a type graph G iff C and D are
the types of a and b respectively, which are minimal in G. In a pattern hC, P, Di, C
and D are referred to as source and target types respectively. A minimal type pattern
hdbo:Film, dbo:starring, dbo:Actori (simply referred to as pattern in the fol-
lowing) tells that there exist entities that have dbo:Film and dbo:Actor as minimal
    ABSTAT summaries for several datasets can be explored at http://abstat.disco.
    unimib.it
    If no ontology is specified, all types are minimal and patterns are extracted like in frameworks
    that do not adopt minimalization
types which are connected through the property P . Non minimal patterns can be inferred
from minimal patterns and the type graph. Therefore, they can be excluded as redundant
without information loss, making summaries more compact [28]. Each pattern hC, P, Di
is associated with a frequency, which reports the number of relational assertions ha, P, bi
from which the pattern has been extracted.
    Local cardinality descriptors. For this work, we have extended ABSTAT to extract
local cardinality descriptors, i.e., cardinality descriptors of RDF properties, which are
specific to the patterns in which the properties occur. To define these descriptors, we first
introduce the concept of restricted property extensions. The extension of a property P
restricted to a pattern hC, P, Di is the set of pairs hx, yi such that the relational assertion
hx, P, yi is part of the dataset and hC, P, Di is a minimal-type pattern for hx, P, yi. Given
a pattern π with a property P , we can define the functions (that return the closest integer):
    minS(π), maxS(π), avgS(π): denoting respectively the minimum, maximum and
average number of distinct subjects associated to unique objects in the extension of P
restricted to π;
    minO(π), maxO(π), avgO(π): denoting respectively the minimum, maximum and
average number of distinct objects associated to unique subjects in the extension of P
restricted to π.
    ABSTAT can also compute global cardinality descriptors by adjusting the above men-
tioned definition so as to consider unrestricted property extensions. Local cardinality de-
scriptors carry information about the semantics of properties as used with specific types
of resources (in specific patterns) and can be helpful for selecting features used to com-
pute the similarity between resources. For example, to compute similarity for movies, one
would like to discard properties that occur in patterns π with dbo:Film as source type
and avgS(π) = 1. We remark that the values of local cardinality descriptors for patterns
with a property P may differ from values of global cardinality descriptors for P . Some
examples of local cardinality descriptors can be found in the faceted-search interface (AB-
STATBrowse). In conclusion, ABSTAT takes a linked dataset and - if specified - one or
more ontologies as input, and returns a summary that consists of: a type graph, a set of
patterns, their frequency, local an global cardinality descriptors.

3    Semantics-aware Feature Selection
Feature selection is the process of selecting a subset of relevant attributes in a dataset.
Thanks to the feature selection process it is possible to improve the prediction perfor-
mance, and to give a better understanding of the process that generates the data [12].
There are three typical measures of feature selection (i) ”filters”, statistical measures to
assign a score to each feature (here the feature selection process is a preprocessing step
and can be independent from learning[11]); (ii) ”wrapper” where the learning system is
used as a black box to score subsets of features [14]; (iii) embedded methods that per-
form the selection within the process of training [12]. In the following, we discuss two
approaches used for the feature selection task: the first operates on the summarization of
the datasets and the second operates on the instances of the datasets.
    Feature Selection with Ontology-based Summaries. As described in Section 2, the
ABSTAT framework provides two useful statistics: the pattern frequency and the cardinal-
ity descriptors that are used in the feature selection process as described in Figure 1. The

    http://abstat.disco.unimib.it/browse
process starts by considering all patterns Π = {π1 , π2 , . . . , πn } of a given class C occur-
ring as a source type. The example in Figure 1 shows a subset of Π with dbo:Film as
source type. The first step of our approach (FILTERBY) filters out properties based on the
local cardinality descriptors. In particular, it filters only properties for which the average
number of distinct subjects associated with unique objects is more than one (avgS > 1).
The second step of the process (SELECTDISTINCTP) selects all properties of the pat-
terns in Π by applying the maximum of the pattern frequency (# in the Figure). Then,
the properties are ranked (ORDERBY) in a descending order on pattern frequency and
then k properties (TOPK) are selected (k=2).
    In some datasets, such as DBpedia, properties may use redundant information
by using same properties with different namespaces, e.g., dbo:starring and
dbp:starring. For this reason, in such case, a pre-processing step for removing repli-
cated properties to avoid redundant ones is requested (see Section 4).
𝝅            P                  D         #      avgS
1 dbo:director             foaf:Person     93k      4
2 dbo:director             dbo:Person      39k      5
3 dbo:director             dbo:Producer    16k      7
4 dbo:wikiPageExternalLink owl:Thing      110k      1                                                                    P
5 dbo:starring             dbo:Person      49k      4                                                             dcterms:subject
6 dbo:starring             dbo:Actor      218k      7                                                             dbo:starring
7 dbo:starring             foaf:Person    306k      4
8 owl:sameAs               owl:Thing      757k      1
9 dcterms:subject          skos:Concept   934k     18
                                                                                                           𝑻𝑶𝑷𝑲(𝑲 = 𝟐)

                         𝝅          P               D   #
                          1 dbo:director    foaf:Person   93k
            𝑭𝑰𝑳𝑻𝑬𝑹𝑩𝒀      2 dbo:director    dbo:Person    39k 𝑺𝑬𝑳𝑬𝑪𝑻𝑫𝑰𝑺𝑻𝑰𝑵𝑪𝑻𝑷        P         #                         P
          . 𝒂𝒗𝒈𝑺6𝟏        3 dbo:director    dbo:Producer 16k      (𝑴𝑨𝑿 # )    dbo:director     93k   𝑶𝑹𝑫𝑬𝑹𝑩𝒀(#) dcterms:subject
                          5 dbo:starring    dbo:Person    49k                 dbo:starring    306k                dbo:starring
                          6 dbo:starring    dbo:Actor    218k                 dcterms:subject 934k                dbo:director
                          7 dbo:starring    foaf:Person  306k
                          9 dcterms:subject skos:Concept 934k



                 Fig. 1: Feature selection with ABSTAT with source type dbo:Film
     Feature Selection with State-of-the-art Techniques. In this work we consider RDF
properties as features, so among the different feature selection techniques available in the
literature, we initially selected Information Gain, Information Gain Ratio, Chi-squared
test and Principal Component Analysis as their computation can be adapted to categorical
features as LD and we then evaluated their effect over the recommendation results. The
features selected from each technique have been used as an input of the recommendation
algorithm that uses the Jaccard index as similarity measure. In order to identify the best
technique among the one we selected, they have been evaluated by using Information
Gain (IG), Gain Ration and Chi Squared Test. At the end, IG[19] resulted as the best
performing one. Then features are ranked according to their IG value and the top-k ones
are returned.
Feature pre-processing. LD datasets usually have a quite large feature set that can
be very sparse depending on the knowledge domain. For instance, taking into ac-
count the movies available in Movielens, properties as dbp:artDirection or
dbp:precededBy are very specific and have a lot of missing values. On the other hand,
properties as dbo:wikiPageExternalLink or owl:sameAs always have different
and unique values, so they are not informative for a recommendation task.
    For the sake of conciseness we do not report all the results here. Results obtained with other FS
    techniques can be found at http://ow.ly/zAA530d0wu0
    For this reason, before starting the feature selection process with IG, we reduced re-
dundant or irrelevant features. The pre-processing step has been done following [23]: we
fixed a threshold tm = td = 97% both for missing values and for distinct values and,
then, we discarded features for which we had more than tm of missing values and more
than td of distinct values.
    Recommendation Method. We implemented a content-based recommender system
using an item-based nearest neighbors algorithm as in [20], where the similarity is com-
puted by means of Jaccard’s index (widely adopted for categorical features). In this work,
the neighborhood of a resource includes all the nodes in the graph reachable starting from
the resource i (respectively j) following the properties selected by the feature selection
phase. The neighbors are thus one-hop features. The similarity values are then used to
recommend to each user the items which result most similar to the ones she has liked
in the past. Ratings are predicted as a normalized sum of neighbors ratings, weighted by
their similarity values [21].

4    Experimental Evaluation
Datasets. The evaluation has been carried out on the three well-known datasets belong-
ing to different domains, i.e. movies(Movielens 1M), books (LibraryThing), and music
(Last.fm). The datasets contains, respectively, 1,000,209, 626,000 and 92,834 ratings.
Movielens 1M and LibraryThing provide explicit ratings over 1-5 and 1-10 scales whereas
Last.fm [3] provides users listening counts.
Measures. For evaluating the quality of our recommendation algorithm we are interested
in measuring its performances in terms of accuracy of the predicted results and diversity.
To evaluate recommendation accuracy, we used Precision (Precision@N) and Mean Re-
ciprocal Rank (MRR). Precision@N is a metric denoting the fraction of relevant items in
the Top-N recommendations. MRR computes the average reciprocal rank of the first rele-
vant recommended item [26]. A good recommender system should provide recommenda-
tions equally distributed among the items, otherwise, even if accurate, they indicate a low
degree of personalization [1]. To evaluate aggregate diversity, we considered catalog
coverage (the percentage of recommended items in the catalog) and aggregate entropy
[1]. Please note that here we are considering a global diversity rather than a personalized
one [8].
    Implementation. We tried different ranking and filtering functions to study their ef-
fect on feature selection. For lack of space we include the best combination from those
proposed in section 3, AbsOccAvgS, that considers as input of FILTERBY the avgS and
SELECTDISTINCTP the maximum of the pattern frequency. Both for ABSTAT and IG
we considered both the Onlydbp configuration in which, if among the first N features se-
lected there are both dbo: and dbp: feature we consider only the dbp: one. Conversely
in Onlydbo we take into account only the dbo: one.
    ABSTAT Baseline. As a baseline for ABSTAT-based feature selection we use TF-IDF
as is a well-known measure to identify most relevant terms (properties in this case) for
a document (a class in this case). We adopt TF-IDF in our context where by document
we refer to a set of patterns having the same subject-type and by term we refer to a
property. TF-IDF is based on the number of properties occurring in a document (TF) and
the logarithm of the ratio between the total number of documents and the ones containing
the property (IDF).
    http://ir.ii.uam.es/hetrec2011/datasets.html
                            Precision@10 MRR@10 catalogCoverage@10 aggrEntropy@10
             Top K features   5      20   5     20    5     20       5      20
             dbo.IG         .0841 .1076 .2961 .3390 .3372 .5226    7.94    8.44
             dbo.AbsOccAvgS .1066 .1067 .3388 .3402 .5344 .5208    8.68    8.51
             dbo.TfIdf      .0823 .0856 .2994 .3123 .3520 .3908    7.83    7.99
             dbp.IG         .0688 .1046 .2134 .3336 .2799 .5065    6.54    8.31
             dbp.AbsOccAvgS .1065 .1059 .3408 .3360 .5426 .5105    8.64    8.31
             dbp.TfIdf      .0549 .0745 .1924 .2687 .2530 .3575    6.33    7.41
                 Table 1: Experimental results on the Movielens dataset.
                            Precision@10 MRR@10 catalogCoverage@10 aggrEntropy@10
             Top K features   5      20   5     20    5     20       5       20
             dbo.IG         .0571 .0579 .2346 .2274 .3988 .4037    10.47   10.44
             dbo.AbsOccAvgS .0561 .0593 .2328 .2329 .3982 .4030    10.54   10.48
             dbo.TfIdf      .0579 .0605 .2374 .2477 .4086 .3991    10.55   10.20
             dbp.IG         .0586 .0586 .2350 .2299 .4027 .4043    10.49   10.40
             dbp.AbsOccAvgS .0623 .0612 .2467 .2342 .3943 .4043    10.42   10.45
             dbp.TfIdf      .0215 .0132 .1608 .1218 .1314 .2696     8.81    9.96
                  Table 2: Experimental results on the LastFM dataset.
                            Precision@10 MRR@10 catalogCoverage@10 aggrEntropy@10
             Top K features   5      20   5     20    5     20       5      20
             dbo.IG         .0411 .1319 .1989 .4083 .4425 .5053    11.06   11.20
             dbo.AbsOccAvgS .1283 .1292 .3986 .4063 .4915 .4949    11.14   11.14
             dbo.TfIdf      .1024 .1132 .3064 .3554 .4026 .4508    10.76   10.96
             dbp.IG         .0678 .1319 .2553 .4083 .4364 .5053    10.83   11.20
             dbp.AbsOccAvgS .1319 .1316 .4026 .4113 .4926 .5055    11.14   11.20
             dbp.TfIdf      .0790 .1170 .2371 .3572 .3894 .4698    10.69   11.04
               Table 3: Experimental results on the LibraryThing dataset.
    Results Tables 1, 2, 3 show the experimental results obtained on, respectively, Movie-
lens, Last.FM and LibraryThing datasets in terms of Precision, MRR, catalogCover-
age, and aggrEntropy. Results are computed over lists of top-10 items recommended
by the RS. We conducted experiments using top-k selected features for different k, i.e.,
k = 5, 10, 15, 20, and all configurations but we report only results for k = 5, 20 and
best configurations . We highlight in bold only the values for which there is a statistical
significant difference. For Lastfm dataset the differences are not statistical significant so
the two methods are equivalent in selecting features.
    Discussion. As an overall result, ABSTAT-based FS leads to the best results in terms
of accuracy and diversity for both the movie and books domains while IG leads to better
results (although not statistically significant) for music.
    Specifically, considering the results on Movielens (Table 1), ABSTAT produces bet-
ter accuracy with respect to IG in all the configurations both with 5 and 20 features.
In terms of aggregate diversity, i.e. itemCoverage and aggrEntropy, ABSTAT is still the
best choice, overcoming IG in almost all the situations. On Lastfm (Table 2) there are
no particular differences, and hence the choice of the method seems irrelevant: both
summarization-based and statistical methods are comparable. Eventually, on Library-
Thing (Table 3), ABSTAT strongly beats IG in almost all the configurations. In particular,
it gets more than twice of the precision and MRR respect to IG in top-5 features sce-
nario. Summing up, ABSTAT beats IG in almost all the configurations on the two datasets
Movielens and LibraryThing, while they act in the same way on the Lastfm dataset.
    In order to investigate the reasons behind the different behaviors depending on the
selected knowledge domain, we measured: (i) the number of minimal patterns and (ii) the
average number of triples per resource and the corresponding variance. Regarding the for-
mer we may say that a higher number of minimal patterns means a richer and more diverse

  The interested reader can find results for all values of k and configurations on GitHub: http:
  //ow.ly/zAA530d0wu0
ontological representation of the knowledge domain. As for the latter, a high variance in
the number of triples associated to resources is a clue of an unbalanced representation of
the items to recommend. Hence, items with a higher number of triples associated result
“more popular” in the knowledge graph compared to those with only a few. This may
reflect in the rising of a stronger content popularity bias while computing the recommen-
dation results. If we look at the values represented in Table 4 we may assert that a higher
                  Domain Number of Minimal Patterns Average Number of Triples Variance
                  Movies           57757                     74,015           549,313
                  Books            41684                     44,966           169,478
                  Music            40481                     80,502           981,509
             Table 4: Ontological and data dimensions of the three datasets
sparsity in the knowledge graph data may give chance to statistical methods to beat onto-
logical ones. In other words, it seems that the higher the sparsity of the knowledge graph
at the data level, the lower the influence of the ontological schema in the selection of the
most informative features to build a pure content-based recommendation engine.
5    Related Work
Summarization. Different approaches have been proposed for schema and data summa-
rization[28]. Several data profiling approaches are aimed to describe linked data by re-
porting statistics about the usage of the vocabularies, types and properties. SchemeEx ex-
tracts interesting measures , by considering the co-occurrence of types and properties [15].
Linked Open Vocabularies, RDFStats [16] and LODStats [2] provide such statistics. In
contrast, ABSTAT represents connections between types using schema patterns, for which
it also provides cardinality descriptors. does not include cardinality descriptors for prop-
erties or patterns. TermPicker extracts [25] patterns consisting in triples hS, E, Oi, where
S and O are sets of types and E is a set of predicates. Instead, ABSTAT and Loupe ex-
tract patterns each consisting in a triple hC, P, Di where C and D are types and P a
property. TermPicker summaries do not describe cardinality and are extracted from RDF
data without considering relationships between types. According to [18], which proposes
a method to define and discover classes of cardinality constraints with some preliminary
results, current approaches focus only on mining keys or pseudo-keys (e.g., [27]). We dis-
cover richer statistics about property cardinality like the above mentioned work, and in
addition we compute cardinality descriptors for properties occurring in specific schema
patterns.
     Recommender Systems. One of the first approaches for using LD in a recommender
system was proposed by Heitmann and Hayes[13]. A system for recommending artists
and music using DBpedia was presented in [22]. The task of cross-domain recommenda-
tion leveraging DBpedia as a knowledge-based framework was addressed in [10], while in
[29] the authors present a semantics-aware approach to deal with cold-start situations. In
[9] the authors use a a hybrid graph-based algorithm built upon DBpedia and collaborative
information. To the best of our knowledge, the only approaches proposing an automatic
selection of LD features are [19, 24]. Finally, we observe that even approaches that do not
perform automatic FS like [19, 4, 9] used (hand-crafted) FS to improve their performance.
6    Conclusions
In this work we investigated the role of ontology-based data summarization for feature
selection in recommendation tasks. Here we compare results coming from ABSTAT, a
    http://lov.okfn.org/
schema summarization tool, with classical methods for feature selection and we show that
the former are allowed to compute better predictions not just in terms of precision of the
recommended items but also considering other dimensions such as diversity. Experiments
have been carried out in three different knowledge domains thus showing the effectiveness
of a feature selection based on schema summarization over classical techniques such as
Information Gain.

References
 1. G. Adomavicius and Y. Kwon. Improving aggregate recommendation diversity using ranking-based techniques. IEEE
    Trans. on Knowl. and Data Eng., 24(5), 2012.
 2. S. Auer, J. Demter, M. Martin, and J. Lehmann. LODStats - An Extensible Framework for High-Performance Dataset
    Analytics. In EKAW (2), 2012.
 3. I. Cantador, P. Brusilovsky, and T. Kuflik. 2nd workshop on information heterogeneity and fusion in recommender systems
    (hetrec 2011). In RecSys. ACM, 2011.
 4. M. de Gemmis, P. Lops, C. Musto, F. Narducci, and G. Semeraro. Semantics-aware content-based recommender systems.
    In Recommender Systems Handbook. 2015.
 5. T. Di Noia. Knowledge-enabled recommender systems: Models, challenges, solutions. In Proceedings of the 3rd Interna-
    tional Workshop on Knowledge Discovery on the WEB, Cagliari, Italy, September 11-12, 2017., 2017.
 6. T. Di Noia, I. Cantador, and V. C. Ostuni. Linked open data-enabled recommender systems: ESWC 2014 challenge on
    book recommendation. In Semantic Web Evaluation Challenge - SemWebEval 2014 at ESWC 2014, Anissaras, Crete,
    Greece, May 25-29, 2014, Revised Selected Papers, pages 129–143, 2014.
 7. T. Di Noia, C. Magarelli, A. Maurino, M. Palmonari, and A. Rula. Using ontology-based data summarization to develop
    semantics-aware recommender systems. In The Semantic Web - 15th International Conference, ESWC 2018, Heraklion,
    Crete, Greece, June 3-7, 2018, Proceedings, pages 128–144, 2018.
 8. T. Di Noia, V. Ostuni, J. Rosati, P. Tomeo, and E. Di Sciascio. An analysis of users’ propensity toward diversity in
    recommendations. In RecSys 2014 - Proceedings of the 8th ACM Conference on Recommender Systems, pages 285–288.
    Association for Computing Machinery, Inc, 2014.
 9. T. Di Noia, V. C. Ostuni, P. Tomeo, and E. Di Sciascio. Sprank: Semantic path-based ranking for top-N recommendations
    using linked open data. ACM TIST, 8(1):9:1–9:34, 2016.
10. I. Fernández-Tobı́as, I. Cantador, M. Kaminskas, and F. Ricci. A generic semantic-based framework for cross-domain
    recommendation. In 2nd Workshop on Information Heterogeneity and Fusion in Recommender Systems, 2011.
11. X. Geng, T.-Y. Liu, T. Qin, and H. Li. Feature selection for ranking. In SIGIR. ACM, 2007.
12. I. Guyon and A. Elisseeff. An introduction to variable and feature selection. Journal of Machine Learning Research, 3,
    2003.
13. B. Heitmann and C. Hayes. Using linked data to build open, collaborative recommender systems. In AAAI Spring
    Symposium: Linked Data Meets Artificial Intelligence, 2010.
14. R. Kohavi and G. H. John. Wrappers for feature subset selection. Artificial Intelligence, 97(1-2), 1997.
15. M. Konrath, T. Gottron, S. Staab, and A. Scherp. Schemex - efficient construction of a data catalogue by stream-based
    indexing of linked data. Web Semant., 16, 2012.
16. A. Langegger and W. Wöß. RDFStats - an extensible RDF statistics generator and library. In DEXA Workshops, pages
    79–83. IEEE, 2009.
17. N. Mihindukulasooriya, M. Poveda Villalon, R. Garcia-Castro, and A. Gomez-Perez. Loupe - An Online Tool for Inspect-
    ing Datasets in the Linked Data Cloud. In ISWC Posters & Demonstrations, 2015.
18. E. Muñoz. On learnability of constraints from RDF data. In ESWC. Springer, 2016.
19. C. Musto, P. Lops, P. Basile, M. de Gemmis, and G. Semeraro. Semantics-aware graph-based recommender systems
    exploiting linked open data. In UMAP, 2016.
20. P. Nguyen, P. Tomeo, T. Di Noia, and E. Di Sciascio. An evaluation of SimRank and Personalized PageRank to build a
    recommender system for the web of data. In WWW ’15 Companion, pages 1477–1482. ACM, 2015.
21. X. Ning, C. Desrosiers, and G. Karypis. A comprehensive survey of neighborhood-based recommendation methods. In
    Recommender Systems Handbook, pages 37–76. 2015.
22. A. Passant. Dbrec: Music Recommendations Using DBpedia. In 9th ISWC, pages 209–224. Springer-Verlag, 2010.
23. H. Paulheim and J. Fümkranz. Unsupervised generation of data mining features from linked open data. In WIMS, 2012.
24. A. Ragone, P. Tomeo, C. Magarelli, T. Di Noia, M. Palmonari, A. Maurino, and E. Di Sciascio. Schema-summarization in
    linked-data-based feature selection for recommender systems. In Proceedings of the Symposium on Applied Computing,
    SAC 2017, Marrakech, Morocco, April 3-7, 2017, pages 330–335, 2017.
25. J. Schaible, T. Gottron, and A. Scherp. Termpicker: Enabling the reuse of vocabulary terms by exploiting data from the
    linked open data cloud. In The Semantic Web. Latest Advances and New Domains - 13th International Conference, ESWC
    2016, Heraklion, Crete, Greece, May 29 - June 2, 2016, Proceedings, pages 101–117, 2016.
26. Y. Shi, A. Karatzoglou, L. Baltrunas, M. Larson, N. Oliver, and A. Hanjalic. CLiMF: learning to maximize reciprocal
    rank with collaborative less-is-more filtering. In RecSys. ACM, 2012.
27. T. Soru, E. Marx, and A. N. Ngomo. ROCKER: A refinement operator for key discovery. In WWW. ACM, 2015.
28. B. Spahiu, R. Porrini, M. Palmonari, A. Rula, and A. Maurino. ABSTAT: ontology-driven linked data summaries with
    pattern minimalization. In 2nd Workshop on Summarizing and Presenting Entities and Ontologies, co-located with ESWC.,
    2016.
29. P. Tomeo, I. Fernández-Tobı́as, I. Cantador, and T. Di Noia. Addressing the cold start with positive-only feedback
    through semantic-based recommendations. International Journal of Uncertainty, Fuzziness and Knowledge-Based Sys-
    tems, 25(Supplement-2):57–78, 2017.
30. G. Troullinou, H. Kondylakis, E. Daskalaki, and D. Plexousakis. RDF Digest: Efficient Summarization of RDF/S KBs.
    In ESWC, 2015.