<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Robust Training of Sequential Recommender Systems with Missing Input Data⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Federico Siciliano</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shoval Lagziel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iftah Gamzu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gabriele Tolomei</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Amazon</institution>
          ,
          <country country="IL">Israel</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sapienza University of Rome</institution>
          ,
          <addr-line>Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In the realm of sequential recommender systems, understanding users' preferences based on their past actions is paramount. Yet, the susceptibility of these models to input perturbations has limited their practicality. Addressing this, we present an innovative approach to mitigate the impact of missing input items, a challenge that has been overlooked. Our method involves a novel training process that anticipates data loss and employs an optimization loss to predict multiple future items. Extensive evaluations on diverse datasets and recommender models underscore its efectiveness. Notably, our approach enhances NDCG@10 by up to 18% with one missing item and an impressive 230% with five missing items, underscoring its substantial impact on system resilience and performance. This work sheds light on the intricate dynamics of sequential recommendation and ofers a solution to real-world data limitations.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Recommender Systems</kwd>
        <kwd>Sequential Recommendation</kwd>
        <kwd>Model Stability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Sequential recommendation models have raised interest in
recent years for their promising increasing performance in
various domains such as e-commerce, health and education
[
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. However, machine learning models are sensitive to
input perturbations [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], and particularly, sequential
recommendation models were shown to be vulnerable to even
a single change in the training data [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ]. The
robustness of recommender systems to data perturbations is a
desired property and is essential in various domains.
Suppose a user regularly uses an e-commerce platform to buy
clothes. The platform collects data on the user’s past
purchases and browsing behavior to make personalized
recommendations for future purchases. However, the user decides
to take a break from the platform for a few weeks and shops
for clothes elsewhere. During this break, the e-commerce
platform is unable to collect data on the user’s behavior,
resulting in missing data. When the user returns to the
platform, the recommender system must take into account the
missing data and still provide personalized
recommendations based on the user’s past purchases. Missing data can
even be dangerous in some domains, such as healthcare [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
where patients might have been treated at diferent clinics,
and this might result in incorrect diagnoses or treatments.
Specifically, in sequential recommendation systems, the
recommendation is based on the sequence of user actions, so
the most recent actions might have an even stronger efect
on the generated recommendations. Considering this, we
explore the impact of missing data in the last items of the
sequence and how to mitigate it by training the models
diferently. To the best of our knowledge, this is the first
work verifying that existing sequential recommender
systems sufer from this efect and applying a method to make
sequential recommender models more robust to this type of
data perturbation. We can summarise our contributions as
follows:
• Our investigation shows that several sequential
recommendation models heavily rely on the last items
in the sequence.
• We apply a modified training method to make the
models more robust to such missing data
perturbations.
• Our model outperforms (as measured by Hit Rate
and Normalized Discounted Cumulative Gain)
classical models in cases of missing data while
maintaining or improving performance in the next item
prediction task.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Sequential Recommendation</title>
        <p>
          Sequential recommendation is a subfield of recommendation
systems [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] that focuses on recommending items to users
based on their recent interactions. The goal of sequential
recommendation is to predict the next item a user will likely
interact with, given their previous interactions. One of the
earliest sequential recommendation methods is the Markov
Chain model [
          <xref ref-type="bibr" rid="ref10 ref11 ref9">9, 10, 11</xref>
          ], which models users’ interactions
as a Markov process and uses the transition probabilities
between items to make recommendations. Recently, there
has been a growing interest in using deep learning
techniques for sequential recommendation. These methods
include using deep neural networks, such as Recurrent Neural
Networks (RNN) [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], Long Short-Term Memory (LSTM)
[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], Gated Recurrent Units (GRUs) [
          <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
          ] and attention
mechanisms [
          <xref ref-type="bibr" rid="ref16 ref17 ref18">16, 17, 18</xref>
          ], to model users’ interactions and
make recommendations, allowing the model to focus on
the most relevant parts of the user’s interaction history
when making recommendations. Additionally, there has
been an increased focus on Explainable AI in sequential
recommenders [
          <xref ref-type="bibr" rid="ref19 ref20">19, 20</xref>
          ], some of which are based on
counterfactuals [
          <xref ref-type="bibr" rid="ref21 ref22 ref23 ref24 ref25">21, 22, 23, 24, 25</xref>
          ], which are aimed at making
the recommendations more tailored to the user [
          <xref ref-type="bibr" rid="ref26 ref27 ref28">26, 27, 28</xref>
          ]
and providing more transparency into the decision-making
process of the model. Overall, the field of sequential
recommendation is rapidly evolving, with a wide range of methods
and techniques being proposed and evaluated[
          <xref ref-type="bibr" rid="ref29 ref30">29, 30</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Robustness of recommender systems</title>
        <p>
          One of the main challenges in the field of recommendation
systems is ensuring the robustness of the models to data
perturbations [
          <xref ref-type="bibr" rid="ref31 ref32 ref33 ref34">31, 32, 33, 34</xref>
          ]. Data perturbations refer to
small changes in the input data, such as missing values or
noisy observations, that can significantly impact the model’s
performance. Many common recommendation methods
are sensitive to such perturbations [
          <xref ref-type="bibr" rid="ref35 ref4 ref5">4, 5, 35</xref>
          ] and can lead
to poor performance or even complete failure. Recently,
there has been an increased focus on developing robust
sequential recommenders that can handle data perturbations.
One approach is to use regularization techniques, such as
dropout [
          <xref ref-type="bibr" rid="ref36 ref37">36, 37</xref>
          ], to reduce the impact of noise in the input
data. Another approach is to use ensemble methods, such
as bagging [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ] and boosting, to combine the predictions
of multiple models. Another area of research on
robustness is the use of generative models, such as Variational
Autoencoders (VAEs) [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ] or Generative Adversarial
Networks (GANs) [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ], to learn the underlying distribution of
the data and generate new samples, which can be used to
augment the training data [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ] and improve the robustness
of the models. Additionally, there has been work on
imputation techniques [
          <xref ref-type="bibr" rid="ref42 ref43">42, 43</xref>
          ], to infer the missing data to
improve recommender systems, and on training instability
[
          <xref ref-type="bibr" rid="ref44">44</xref>
          ]. Finally, other works focus on methods for evaluating
the robustness of the model without using ranking
evaluation metrics but rather by assessing the stability of the
generated rankings in the presence of missing items in the
data [
          <xref ref-type="bibr" rid="ref35 ref4 ref5 ref6">4, 5, 6, 35</xref>
          ]. Overall, robustness is a critical issue in
sequential recommendation. There are many ongoing eforts
to develop methods that can handle data perturbations and
improve the performance of the models in practice.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Setting</title>
      <p>The setting in question consists of  users and  items.
Each user  has interacted with at least one item  at a
given time , . The goal of a recommender system is to
predict the compatibility between a given user and the items
with which it has not yet interacted, knowing the items the
user has interacted with. In the Sequential Recommendation
case, the problem takes the form of predicting the next item
in a sequence: given a sequence of  items {1, 2, ..., }
with which a user  has interacted, the goal is to predict the
 + 1-th item (+1). A sequential neural network takes as
input a sequence of at most  elements and performs, for
each time step, the prediction of  values. These represent
the estimated compatibility between user , to which the
sequence of items belong, and all items .</p>
      <sec id="sec-3-1">
        <title>3.1. Classic Training Method</title>
        <p>The goal of the network, at each timestep , is to predict the
next item in the sequence +1. During network training,
if the number of possible items  is too large, it becomes
intractable to calculate predictions for all of them, therefore,
only a chosen few are calculated. In particular, computing
the one corresponding to the next item, called positive item,
and at least one corresponding to an item that is irrelevant,
called negative item, chosen randomly at each epoch. An
attempt is then made to increase the former value at the
expense of the latter. Notably, the negative items are indeed
chosen randomly, but excluding items already in the input
sequence. To achieve this, the loss used for the models
we have considered is the Binary Cross-Entropy; this is
typically used in a classification scenario.</p>
        <p>The definition of loss, positive items, and negative items
are defined leads to a specific ranking that the network
should achieve at time step  of a sequence of  elements.
The ranking can be described to be as follows: first the
positive item +1, followed by, in indiferent order, all the
other items in the sequence  such that  ∈ {1, ..., ,  +
2, ..., ,  + 1}, and finally, again in indiferent order, all
the remaining items  such that  ∈/ {1, 2, ...,  + 1}. This
particular ranking would result in zero loss.</p>
        <p>We can simplify, as shown in Figure 1a, the functioning
of the network by imagining that for the sequence [1], the
model should output item 2, for the sequence [1, 2] it
should output item 3, and so on.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Problem Statement</title>
      <p>A sequential recommender system receives as input a
sequence  = {1, 2, ..., } and tries to predict the next
item in the sequence, item +1. How would the network
behave if the last item  is missing? If the last item is
removed from the sequence [1, 2], we would be left only
with the sequence [1]. Since the network is trained to
predict only item 2, it will have no preference on predicting
3. Furthermore, considering an out-of-sequence division
of the dataset (more info on data division in Section 5.4.3),
removing two items from the sequence results in replicating
a training sequence. Thus, the efect of removal may be
even more detrimental if the model is overfitting on the
training set. Lack of user-item interactions in real-world
scenarios pose a challenge for sequential recommenders.
For example, a streaming service ofering a movie trilogy
may not have data on a user’s interaction with the second
movie if it was watched on a diferent platform. This could
result in the recommender suggesting the second movie
as the top recommendation without considering the third.
This issue also applies to non-sequential items like sequels
to movies or books. We start by demonstrating that existing
sequential recommender models sufer from this efect, and
then we devise a method to make them robust to this type
of data perturbation.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Methodology</title>
      <sec id="sec-5-1">
        <title>5.1. More Positive Items</title>
        <p>We assume that, in a real scenario, an ideal model should
yield, at a given time step , a ranking containing, in order,
all future items in the sequence, and only upon finishing
these, all other (negative) items. Therefore, the solution we
have applied is to choose  positive items, such that
the network learns to simultaneously predict  future
instances. Please note that we are not trying to predict the
whole sequence of future interactions but only the relevance
of the items at time step . In this case, the loss would become
as in 1.
...</p>
        <sec id="sec-5-1-1">
          <title>Model</title>
        </sec>
        <sec id="sec-5-1-2">
          <title>Model</title>
        </sec>
        <sec id="sec-5-1-3">
          <title>Model</title>
          <p>
            ...
Model
Model
(a) Simplified visualization of a Sequential Recommender System
functioning. Each  represents the item at time step . Each
line represents a diferent inference.
(b) Simplified visualization of a Sequential Recommender System
functioning in the case of  = 3 positives items. Each
 represents the item at time step . Each line represents a
diferent inference.
(1)
where ⃗ represents the output of the network, ⃗ =
{1, ...,  } the identifier of the  positive items, and
⃗ = {1, ...,  )} those of the selected 
negative items. The loss takes the same form as that presented
in [
            <xref ref-type="bibr" rid="ref45">45</xref>
            ]. Although the authors mention the loss function, its
potential for improving the model’s robustness in the face
of missing data has not been explored. Our work fills this
gap by showing how this loss function can be efectively
used to increase the robustness of the model and improve
its performance in the presence of missing data. Replicating
the simplified illustration in Figure 1a, we can visualize the
idea of predicting multiple positive items as in Figure 1b.
          </p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Margin Loss</title>
        <p>Considering more positive items poses a clear limitation,
as it becomes more challenging for the next item to rank
high in the network’s ranking. This is because the Binary
Cross-Entropy loss does not distinguish between positive
items; a perfect model would rank all  positive items in
the first  positions, regardless of their order. This might
limit the model performance, as the item may end-up in
the  -th position, thus reducing common metrics that take
into account the order of results, such as NDCG. To solve
this problem, we decide to use the Margin Loss. Given pairs
of inputs 1 and 2, and a preferred ordering of them ,
such that  = 1 if we assume that the first input should be
ranked higher than the second input, vice-versa for  = − 1,
the margin loss ℓ takes values according to ℓMRG(1, 2,  |
) = max(0, (2 − 1) + ). This tells
us that if the network outputs for the two inputs respect
the expected ordering and are at least  apart, the
loss is equal to 0; otherwise, it is proportional to the
distance between them. Input pairs are formed between
all pairs of positive items. The expected order is the
order in which the user interacted with them: an item at
a time step  must come first in ranking than one at a
time step  + . The Margin Loss formula is ℓMRG,pos(⃗ |
⃗, ) = ∑︀=1 ∑︀=+1 max(0,  −  +
), where ⃗ represents the output of the network,
⃗ = {1, ...,  )} the identifier of the  positive
items and  the margin value. The equation holds
only if the order of the identifiers of the positive elements
follows the expected order.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Mixed Loss</title>
        <p>The margin loss applied on the positive items is not enough
to train the neural network as we desire. It is always
necessary to discourage the model from predicting negative
items. We, therefore, decide to use it in conjunction with
the traditional Cross-Entropy loss. This naturally brings up
the need to add some hyperparameters to weigh the
importance of the two losses. We also separate the components of
the Binary Cross-Entropy loss pertaining to positive items
and negative items. This Mixed Loss formula is ℓMIX(⃗ |
⃗, ⃗, ) = BCE,pos +  1BCE,neg +  2ℓMRG,pos(⃗ |
⃗, ).</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Experiments</title>
        <p>
          5.4.1. Datasets
We select three datasets that are widely used in this field [
          <xref ref-type="bibr" rid="ref46 ref47">46,
47</xref>
          ]: MovieLens-1M [
          <xref ref-type="bibr" rid="ref48">48</xref>
          ], MovieLens-100K [
          <xref ref-type="bibr" rid="ref48">48</xref>
          ] and Amazon
Beauty [
          <xref ref-type="bibr" rid="ref49">49</xref>
          ]. The first two are movie ratings taken from
the MovieLens website1 and difer on the period they were
collected and the size of the set. The third dataset2 contains
reviews and metadata from Amazon, spanning May 1996
Oct 2018. The three datasets have 165, 106 and 5 interactions
per user, respectively. The statistics for all the considered
datasets are shown in Table 1.
1https://movielens.org
2https://nijianmo.github.io/amazon/index.html
We select three sequential recommendation models. The
ifrst, GRU4REC [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], is an RNN based on Gated Recurrent
Unit. SASRec [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], on the other hand, is a sequential
selfattention based model that uses an attention mechanism
to make predictions based on a relatively small number of
actions. TiSASRec (Time Interval aware Self-attention based
sequential recommendation) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] is instead a modification
of this that adds to the input the time intervals between
elements in the sequence.
5.4.3. Preprocessing
Consistent with other work, we use implicit ratings, so
we do not consider the score but simply the existence of an
interaction of a given user with a given product. Given a user
, the products he interacted with are ordered in a sequence
 based on the timestamp. An out-of-sequence split (i.e.
the last two items in each sequence are kept aside to be the
target output of validation and test, respectively, while the
rest of the sequence is used for training) is performed to
partition the data into training, validation and test sets, in
line with what has been done by other works in the same
domain.
5.4.4. Evaluation
In line with what has been done in other works involving
Neural Recommenders [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], in order to avoid to avoid heavy
computation, the evaluation is carried out in the following
way: the prediction made by the network is taken for the
positive item (the next item in the sequence) and 100 items
chosen randomly, not in the input sequence. The predictions
(for 100 negative items + the positive item) are then sorted
according to the values obtained; this represents the final
ranking.
        </p>
        <p>
          We want to emphasize that while we use multiple positive
items during training, this is not done during the evaluation
phase. The reason for this decision is that changing the
evaluation method could naturally result in our proposed losses
appearing better, thus rendering the comparison invalid.
By adhering to the traditional evaluation setting, we align
ourselves with the evaluation methods used in other works
in this field. However, we acknowledge that this places
our proposed method at a disadvantage compared to the
baseline method for obvious reasons. We are training the
model to predict multiple items to increase its robustness,
but only one of these items will be used during evaluation.
On the other hand, the baseline model focuses solely on
a single positive item, the same one used for evaluation,
which inherently gives it an advantage. In Section 6, we will
demonstrate how our model still manages to achieve
superior results. In addition to the standard metrics, to evaluate
the sensitivity of the models in cases of input data
perturbations, we utilized the recently introduced metric Rank
List Sensitivity (RLS)[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], enabling to compare rankings
produced with and without perturbations. RLS is defined as
 = 1 ∑︀  (,, ,). where  is the number
of samples,  is a similarity function, , and , are
two rankings produced for sample . In our specific case, 
represents the ranking when sample  is unaltered, while 
represents the case when the input sequence is perturbed,
i.e. items are removed. The similarity () of two rankings
 and  can be calculated using the Jaccard similarity
[
          <xref ref-type="bibr" rid="ref50">50</xref>
          ], but it does not consider order. On the other hand,
Rank-biased Overlap (RBO) is more valuable for a
recommendation system as it considers top-ranked items as more
significant using specific weighting (see Equation 2).
 (,  ) = | ∩  |
| ∪  |
        </p>
        <p>(, ) = (1 − ) ∑︁ − 1 |[1 : ] ∩  [1 : ]|

=1
(2)
5.4.5. Hyperparameter Optimization
The hyper-parameters to be optimised are the number of
positive items to be used, the number of negative items to
be used and the Mixed Loss parameters. The number of
positive items  and negative items  varies in the
set {1, 3, 10}, and the Mixed Loss parameters  1 and  2 in
the set {1, 10− 1, ..., 10− 5}.
5.4.6. Implementation
All code is written in Python 3. In particular, with Pytorch
and Pytorch Lightning.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>In this section, we present experimental results showing the
strong reliance of sequential recommender models on the
last items in the sequence as well as the performance of the
proposed training method to mitigate this efect.</p>
      <sec id="sec-6-1">
        <title>6.1. Last Items Importance</title>
        <p>Figure 2a visualizes the efect of removing an item at
diferent positions in the sequence on the model outputs, using
the SASRec model and the MovieLens-1M dataset, has on
the ranking of the top item. We identify this item with the
term previous top-ranked item: that item which, prior to
the input data perturbation, was at the top of the ranking.
In the case of the base model, removing the last item can
push the previously top-ranked item by over 25 positions on
average. While SASRec is trained using dropout, this does
not seem to be suficient to make it robust to missing data.
In contrast, when the model is trained with more positive
items, the removal of the last item results in a significantly
lower drop in the ranking of the previous top item: 5
positions or less. We also observe that the diference between
the diferent models becomes less pronounced as we move
towards the earlier items in the sequence. These results</p>
        <p>2
Location from the end of the sequence of the removed item
1
1
2</p>
        <p>3
Removed items per sequence
4
5</p>
        <p>2 3
Removed items per sequence
4
1 2 3</p>
        <p>Removed items per sequence
Base model MP model</p>
        <p>MP+ML model
1</p>
        <p>2 3
Removed items per sequence
4
demonstrate that incorporating more positive items in the
training process and using our proposed Mixed Loss can
help to mitigate the impact of missing data and improve the
robustness of Sequential Recommender Systems.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Performance of Diferent Training</title>
      </sec>
      <sec id="sec-6-3">
        <title>Methods in Cases of Missing Last Items</title>
        <p>NDCG@10 score is used to gauge the impact of the modified
training method on the performance of the models. Figure
3 provides a visualization of the results. The results are also
expressed integrally in Tables 2, 3 and 4.
6.2.1. A Clear Advantage in Handling Missing Data
One striking observation is that the models trained with
more positive items and the Mixed Loss consistently
outperform the base model when it comes to dealing with missing
data. Although the base model performs slightly better in
the absence of missing data, the new models are able to
sustain their performance even as the number of missing
items increases. This is especially evident in the case of the
Amazon Beauty dataset, where the new training method
is able to maintain acceptable performance as the missing
data becomes more prominent.</p>
        <p>As mentioned in Section 5.4.4, the slight predominance of
the base model in the absence of missing data is expected
because the evaluation setting naturally favors the base model:
both the loss function and the evaluation technique consider
1
1
2
2
3
3
4
4
5
5
5
25
20</p>
        <p>2
Location from the end of the sequence of the removed item
1
0
1</p>
        <p>2 3
Removed items per sequence
4
5
only a single item. We emphasize the significance that our
model, trained in a manner that deviates slightly from the
traditional evaluation setting, is able to retain minimal
performance loss in the same setting while gaining robustness
to missing data.
6.2.2. Length of sequences
It is worth noting that the average sequence length of the
three datasets is vastly diferent (see Section 5.4.1). The
results in Figure 3 indicate that the impact of missing data is
much less severe for datasets with longer sequences, such as
ML-1M. The model trained with the classic training method
is even able to compensate for this deficiency, particularly
in the case of SASRec. However, as the average sequence
length decreases, such as in the ML-100k dataset, the
robustness to missing data seems to decline rapidly, and the
diference between the models becomes more pronounced
when the number of missing items increases. This trend
is especially evident in the Amazon Beauty dataset, where
the diference between the models is particularly noticeable
when the number of missing items is higher than 2, probably
because the average length of the sequences for this dataset
is 5.</p>
      </sec>
      <sec id="sec-6-4">
        <title>6.3. Rank List Stability</title>
        <p>The robustness of the models in the face of item removal at
the end of the input sequence is illustrated through the Rank
List Stability with Rank-biased Overlap metric in Figure 2b.
It is evident that the new models exhibit higher stability,
with Cross-Entropy with more positive items proving to
be even more robust than the Mixed Loss model. We
observed a similar trend for all datasets and models, so only
one plot is presented; further results can be found in the
additional repository. While the multiple positive model
(MP) provides in most cases higher performance in the Rank
List Stability metrics compared to the model with multiple
positive and the mixed loss (MP+ML), it is worth noting that
MP+ML provides higher performance on the HR@10 and
NDCG@10. This can be explained by the fact that MP is not
optimized using the ranks of the positive items as done by
the mixed loss (MP+ML model). However, for precisely the
same reason, MP benefits from higher stability.</p>
        <p>More specifically, both models are trained to predict, at
time  and for a given input sequence,  positive items,
specifically [, +1, ..., + ]. However, the MP model
is trained with a loss function that does not consider the
order of the positive items: the same sequence in reverse
order would yield the same loss value. As discussed in
Section 5.4.4, in the classic evaluation setting, only  is
used during evaluation. If the loss function treats  equally
important as the other positive items [+1, ..., + ], it
is more likely to be ranked lower, thus reducing metrics
such as NDCG and Recall. On the other hand, the model
using the Mixed Loss, which aims to prioritize the position
of  at the top of the ranking, has an advantage in achieving
higher metrics in this regard.</p>
      </sec>
      <sec id="sec-6-5">
        <title>6.4. Study on the Number of Positives</title>
        <p>To understand the impact of the number of positive items
used for training, experiments were performed using
diferent numbers of positive items for just one model, SasRec,
and one dataset, MovieLens-1M, due to the computational
time required. As seen in Figure 4a, as the number of
positive items increases, the change in ranking for previous
top-ranked items decreases significantly. However, Figure
4b shows that the performance in the absence of missing
data degrades as the number of positive items increases.
This trend begins to change as the number of missing items
increases, and the gap between the new models and the base
model narrows, with the latter’s performance deteriorating
more.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Implications of the Research</title>
    </sec>
    <sec id="sec-8">
      <title>Findings</title>
      <p>The findings of this study hold both theoretical and
practical implications that contribute to the advancement of
sequential recommender systems and their application in
real-world scenarios. By addressing the specific challenges
posed by missing input data, our research ofers a novel
perspective on enhancing the robustness and reliability of
these systems.
7.1. Theoretical Implications
1. Uncovering Last-Item Dependence: Our research
uncovers the strong reliance of sequential
recommender systems on the last items in the input
sequence. This revelation contributes to a deeper
understanding of the dynamics within these systems,
emphasizing the need for strategies that can
mitigate the performance degradation caused by missing
items.
2. New Training Paradigm: The introduction of a
training approach that anticipates data loss and
simulates prediction of multiple future items presents
a paradigm shift in the methodology for handling
missing input data. This approach establishes a
theoretical foundation for designing more resilient
recommender systems.
7.2. Practical Implications
1. Real-World Data Challenges: In real-world
scenarios, complete user action sequences are often not
available due to various constraints. Our research
highlights the practical significance of addressing
this data scarcity and provides a concrete solution
to mitigate the negative efects of missing items,
improving the usability of recommender systems.
2. Enhanced System Resilience: The proposed
training method significantly improves the performance
of sequential recommender systems when faced with
missing items. This directly translates into a more
reliable and user-centric experience, thus
benefiting various domains, such as e-commerce, content
recommendation, and personalized services.
3. Impact on User Satisfaction: The performance
enhancement demonstrated by our approach can
lead to improved user satisfaction by providing more
accurate and relevant recommendations, even when
there are gaps in the available data. This practical
outcome can foster greater user engagement and
loyalty.
4. General Applicability: The efectiveness of our
method across various datasets and recommender
models underscores its general applicability. This
widens its potential adoption and impact, making
it a valuable tool for researchers and practitioners
alike.</p>
    </sec>
    <sec id="sec-9">
      <title>8. Discussion and Conclusions</title>
      <p>
        Our findings show that the last items in a sequence have
a significant impact on the predictions of sequential
recommenders, and their removal results in unstable rankings.
However, by incorporating multiple future items in the
training process, model robustness can be improved. Our results
demonstrate that the proposed training methods improve
rankings stability (RLS metric) and performance (HR and
NDCG) on various popular sequential recommender models
(SasRec[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], TiSasRec[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and GRU4Rec[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]) and datasets.
In contrast, the performance without missing data is not
noticeably afected but even improves for specific
models/datasets. Using more positive items with Cross-Entropy
loss improves robustness of sequential recommenders to
removal of elements at the end of the input sequence.
However, increasing the number of future items excessively can
lead to stability increase at the cost of decreased
performance. Mixed Loss, combining Cross-Entropy with Margin
Loss, can prioritize the next item over other positives. Our
method opens up opportunities for further research in the
ifeld. Future work may focus on the development of a loss
function that balances performance and robustness as the
number of positive items increases, as well as modifying
the method for models that use bi-directional connections
(e.g., [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]). Moreover, out proposal is easily extendable to
other approaches, as it is solely tied to a diferent training
method and not to a specific architecture. To summarize,
our work represents a step forward in improving the
robustness of sequential recommender models. We demonstrate
the strong influence of the last items in a sequence and the
efectiveness of our method in mitigating the impact of
missing data. Overall, we expect that our findings and proposed
methods will be a valuable tool in the field of sequential
recommender systems.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cremonesi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jannach</surname>
          </string-name>
          ,
          <article-title>Sequenceaware recommender systems</article-title>
          ,
          <source>ACM Computing Surveys (CSUR) 51</source>
          (
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Brilhante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Macedo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Nardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Perego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Renso</surname>
          </string-name>
          ,
          <article-title>Tripbuilder: A tool for recommending sightseeing tours</article-title>
          ,
          <source>in: Advances in Information Retrieval: 36th European Conference on IR Research</source>
          , ECIR
          <year>2014</year>
          , Amsterdam, The Netherlands,
          <source>April 13-16</source>
          ,
          <year>2014</year>
          . Proceedings 36, Springer,
          <year>2014</year>
          , pp.
          <fpage>771</fpage>
          -
          <lpage>774</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <article-title>Dealing with noise problem in machine learning data-sets: A systematic review</article-title>
          ,
          <source>Procedia Computer Science</source>
          <volume>161</volume>
          (
          <year>2019</year>
          )
          <fpage>466</fpage>
          -
          <lpage>474</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Oh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ustun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>McAuley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Rank list sensitivity of recommender systems to interaction perturbations</article-title>
          ,
          <source>in: Proceedings of the 31st ACM International Conference on Information &amp; Knowledge Management</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>1584</fpage>
          -
          <lpage>1594</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Oh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Robustness of deep recommendation systems to untargeted interaction perturbations</article-title>
          ,
          <source>arXiv preprint arXiv:2201.12686</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Filippo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Federico</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pushkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fabrizio</surname>
          </string-name>
          ,
          <article-title>Investigating the robustness of sequential recommender systems against training data perturbations</article-title>
          ,
          <source>in: Advances in Information Retrieval: 46th European Conference on Information Retrieval (ECIR</source>
          <year>2024</year>
          ), Springer,
          <year>2024</year>
          , pp.
          <fpage>205</fpage>
          -
          <lpage>220</lpage>
          . URL: https:// doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>14</fpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>14</fpage>
          ,
          <string-name>
            <surname>first</surname>
            <given-names>Online</given-names>
          </string-name>
          <source>: 16 March</source>
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T. N. T.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Felfernig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Trattner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Holzinger</surname>
          </string-name>
          ,
          <article-title>Recommender systems in the healthcare domain: stateof-the-art and research issues</article-title>
          ,
          <source>Journal of Intelligent Information Systems</source>
          <volume>57</volume>
          (
          <year>2021</year>
          )
          <fpage>171</fpage>
          -
          <lpage>201</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ricci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rokach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shapira</surname>
          </string-name>
          ,
          <article-title>Introduction to recommender systems handbook</article-title>
          , in: Recommender systems handbook, Springer,
          <year>2011</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F.</given-names>
            <surname>Fouss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Faulkner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kolp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pirotte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saerens</surname>
          </string-name>
          , et al.,
          <article-title>Web recommendation system based on a markovchainmodel.</article-title>
          ,
          <source>in: ICEIS (4)</source>
          ,
          <year>2005</year>
          , pp.
          <fpage>56</fpage>
          -
          <lpage>63</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Shani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Heckerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. I.</given-names>
            <surname>Brafman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Boutilier</surname>
          </string-name>
          ,
          <article-title>An mdp-based recommender system</article-title>
          .,
          <source>Journal of Machine Learning Research</source>
          <volume>6</volume>
          (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tomko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. D.</given-names>
            <surname>Salim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          ,
          <article-title>A location-query-browse graph for contextual recommendation</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>30</volume>
          (
          <year>2017</year>
          )
          <fpage>204</fpage>
          -
          <lpage>218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <article-title>Recurrent neural networks with top-k gains for session-based recommendations</article-title>
          ,
          <source>in: Proceedings of the 27th ACM international conference on information and knowledge management</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>843</fpage>
          -
          <lpage>852</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <article-title>A lstm based model for personalized context-aware citation recommendation</article-title>
          ,
          <source>IEEE access 6</source>
          (
          <year>2018</year>
          )
          <fpage>59618</fpage>
          -
          <lpage>59627</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gulcehre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>Empirical evaluation of gated recurrent neural networks on sequence modeling</article-title>
          ,
          <source>arXiv preprint arXiv:1412.3555</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          ,
          <article-title>Session-based recommendations with recurrent neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:1511.06939</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>W.-C. Kang</surname>
            ,
            <given-names>J. McAuley</given-names>
          </string-name>
          ,
          <article-title>Self-attentive sequential recommendation, in: 2018 IEEE international conference on data mining (ICDM)</article-title>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>197</fpage>
          -
          <lpage>206</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. McAuley</surname>
          </string-name>
          ,
          <article-title>Time interval aware selfattention for sequential recommendation</article-title>
          ,
          <source>in: Proceedings of the 13th international conference on web search and data mining</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>322</fpage>
          -
          <lpage>330</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sun</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ou</surname>
          </string-name>
          , P. Jiang,
          <article-title>Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer</article-title>
          ,
          <source>in: Proceedings of the 28th ACM international conference on information and knowledge management</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1441</fpage>
          -
          <lpage>1450</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>I.</given-names>
            <surname>Nunes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jannach</surname>
          </string-name>
          ,
          <article-title>A systematic review and taxonomy of explanations in decision support and recommender systems, User Modeling and User-Adapted Interaction 27 (</article-title>
          <year>2017</year>
          )
          <fpage>393</fpage>
          -
          <lpage>444</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          , et al.,
          <article-title>Explainable recommendation: A survey and new perspectives</article-title>
          ,
          <source>Foundations and Trends® in Information Retrieval</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>101</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ghazimatin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Balalau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Saha</given-names>
            <surname>Roy</surname>
          </string-name>
          , G. Weikum, Prince:
          <article-title>Provider-side interpretability with counterfactual explanations in recommender systems</article-title>
          ,
          <source>in: Proceedings of the 13th International Conference on Web Search and Data Mining</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>196</fpage>
          -
          <lpage>204</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>Counterfactual explainable recommendation</article-title>
          ,
          <source>in: Proceedings of the 30th ACM International Conference on Information &amp; Knowledge Management</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1784</fpage>
          -
          <lpage>1793</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>K. H. Tran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ghazimatin</surname>
            ,
            <given-names>R. Saha</given-names>
          </string-name>
          <string-name>
            <surname>Roy</surname>
          </string-name>
          ,
          <article-title>Counterfactual explanations for neural recommenders</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1627</fpage>
          -
          <lpage>1631</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Tolomei,
          <article-title>The dark side of explanations: Poisoning recommender systems with counterfactual examples</article-title>
          ,
          <source>arXiv preprint arXiv:2305.00574</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ahn</surname>
          </string-name>
          , G. Tolomei, Grease:
          <article-title>Generate factual and counterfactual explanations for gnn-based recommendations</article-title>
          ,
          <source>arXiv preprint arXiv:2208.04222</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <article-title>A reinforcement learning framework for explainable recommendation</article-title>
          , in: 2018 IEEE international conference
          <article-title>on data mining (ICDM)</article-title>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>587</fpage>
          -
          <lpage>596</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Counterfactual data-augmented sequential recommendation</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>347</fpage>
          -
          <lpage>356</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.-S.</given-names>
            <surname>Chua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Causerec: Counterfactual user sequence synthesis for sequential recommendation</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>367</fpage>
          -
          <lpage>377</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sbandi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Siciliano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <article-title>Mitigating extreme cold start in graph-based recsys through reranking</article-title>
          ,
          <source>in: Proceedings of the 33rd ACM International Conference on Information and Knowledge Management</source>
          , CIKM '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>4844</fpage>
          -
          <lpage>4851</lpage>
          . URL: https://doi.org/10.1145/3627673.3680069. doi:
          <volume>10</volume>
          . 1145/3627673.3680069.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>F.</given-names>
            <surname>Betello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Purificato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Siciliano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Trappolini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bacciu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tonellotto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <article-title>A reproducible analysis of sequential recommender systems</article-title>
          ,
          <source>IEEE Access 13</source>
          (
          <year>2025</year>
          )
          <fpage>5762</fpage>
          -
          <lpage>5772</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2024</year>
          .
          <volume>3522049</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>R.</given-names>
            <surname>Burke</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. P. O'Mahony</surname>
            ,
            <given-names>N. J.</given-names>
          </string-name>
          <string-name>
            <surname>Hurley</surname>
          </string-name>
          ,
          <article-title>Robust collaborative recommendation</article-title>
          ,
          <source>in: Recommender systems handbook</source>
          , Springer,
          <year>2015</year>
          , pp.
          <fpage>961</fpage>
          -
          <lpage>995</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>M. O'Mahony</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Hurley</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Kushmerick</surname>
          </string-name>
          , G. Silvestre,
          <article-title>Collaborative recommendation: A robustness analysis</article-title>
          ,
          <source>ACM Transactions on Internet Technology (TOIT) 4</source>
          (
          <year>2004</year>
          )
          <fpage>344</fpage>
          -
          <lpage>377</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Deldjoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Di</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. Di</given-names>
            <surname>Sciascio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Merra</surname>
          </string-name>
          ,
          <article-title>How dataset characteristics afect the robustness of collaborative recommendation models</article-title>
          ,
          <source>in: Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>951</fpage>
          -
          <lpage>960</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>V.</given-names>
            <surname>Guarrasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Siciliano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          , Robustrecsys @ recsys2024:
          <article-title>Design, evaluation and deployment of robust recommender systems</article-title>
          ,
          <source>in: Proceedings of the 18th ACM Conference on Recommender Systems</source>
          , RecSys '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>1265</fpage>
          -
          <lpage>1269</lpage>
          . URL: https://doi.org/10.1145/3640457.3687106. doi:
          <volume>10</volume>
          . 1145/3640457.3687106.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>F.</given-names>
            <surname>Betello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Siciliano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Silvestri</surname>
          </string-name>
          ,
          <article-title>Finite rank-biased overlap (frbo): A new measure for stability in sequential recommender systems</article-title>
          ,
          <source>in: Proc. of the 14th Italian Information Retrieval Workshop</source>
          , volume
          <volume>3802</volume>
          ,
          <year>2024</year>
          , pp.
          <fpage>78</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ghahramani</surname>
          </string-name>
          ,
          <article-title>Dropout as a bayesian approximation: Representing model uncertainty in deep learning</article-title>
          ,
          <source>in: international conference on machine learning, PMLR</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1050</fpage>
          -
          <lpage>1059</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ma</surname>
          </string-name>
          , C. Chen,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Liu, S. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>A survey on dropout methods and experimental verification in recommendation</article-title>
          ,
          <source>arXiv preprint arXiv:2204</source>
          .
          <year>02027</year>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vinagre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Jorge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gama</surname>
          </string-name>
          ,
          <article-title>Online bagging for recommender systems</article-title>
          ,
          <source>Expert Systems</source>
          <volume>35</volume>
          (
          <year>2018</year>
          )
          <article-title>e12303</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>D.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Krishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. D.</given-names>
            <surname>Hofman</surname>
          </string-name>
          , T. Jebara,
          <article-title>Variational autoencoders for collaborative filtering</article-title>
          ,
          <source>in: Proceedings of the 2018 world wide web conference</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>689</fpage>
          -
          <lpage>698</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Miao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guan</surname>
          </string-name>
          , Pdgan:
          <article-title>Adversarial learning for personalized diversitypromoting recommendation</article-title>
          .,
          <source>in: IJCAI</source>
          , volume
          <volume>19</volume>
          ,
          <year>2019</year>
          , pp.
          <fpage>3870</fpage>
          -
          <lpage>3876</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , R. Liu,
          <article-title>Adversarial training for session-based item recommendations</article-title>
          ,
          <source>in: 2020 IEEE 9th Joint International Information Technology and Artificial Intelligence Conference (ITAIC)</source>
          , volume
          <volume>9</volume>
          , IEEE,
          <year>2020</year>
          , pp.
          <fpage>1162</fpage>
          -
          <lpage>1168</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>S.-Y.</given-names>
            <surname>Ihm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-E.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-H.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nasridinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-H.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <article-title>A technique of recursive reliability-based missing data imputation for collaborative filtering</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>11</volume>
          (
          <year>2021</year>
          )
          <fpage>3719</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>Efective collaborative ifltering approaches based on missing data imputation</article-title>
          , in: 2009 Fifth International Joint Conference on INC,
          <article-title>IMS and IDC</article-title>
          , IEEE,
          <year>2009</year>
          , pp.
          <fpage>534</fpage>
          -
          <lpage>537</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Drori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sathiamoorthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gilmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Yi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <article-title>Improving training stability for multitask ranking models in recommender systems</article-title>
          ,
          <source>arXiv preprint arXiv:2302.09178</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Personalized top-n sequential recommendation via convolutional sequence embedding</article-title>
          ,
          <source>in: Proceedings of the eleventh ACM international conference on web search and data mining</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>565</fpage>
          -
          <lpage>573</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>J. Y.</given-names>
            <surname>Chin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. Cong,</surname>
          </string-name>
          <article-title>The datasets dilemma: How much do we really know about recommendation datasets?</article-title>
          ,
          <source>in: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>141</fpage>
          -
          <lpage>149</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [47]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-C.</given-names>
            <surname>Haw</surname>
          </string-name>
          , K.-W. Ng,
          <article-title>Deep learning basedrecommendation system: An overview on models, datasets, evaluation metrics, and future trends</article-title>
          ,
          <source>in: Proceedings of the 2019 2nd International Conference on Computational Intelligence and Intelligent Systems</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>6</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [48]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Harper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Konstan</surname>
          </string-name>
          ,
          <article-title>The movielens datasets: History and context, Acm transactions on interactive intelligent systems (tiis) 5 (</article-title>
          <year>2015</year>
          )
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [49]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. McAuley</surname>
          </string-name>
          ,
          <article-title>Justifying recommendations using distantly-labeled reviews and fine-grained aspects</article-title>
          ,
          <source>in: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>188</fpage>
          -
          <lpage>197</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [50]
          <string-name>
            <given-names>P.</given-names>
            <surname>Jaccard</surname>
          </string-name>
          ,
          <article-title>The distribution of the flora in the alpine zone</article-title>
          . 1, New phytologist
          <volume>11</volume>
          (
          <year>1912</year>
          )
          <fpage>37</fpage>
          -
          <lpage>50</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>