<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>P. Seshadri);</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>the Need for Task Discernment</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shahrzad Shashaani</string-name>
          <email>shahrzad.shashaani@tuwien.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavan Seshadri</string-name>
          <email>pavanseshadri1@acm.org</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Peter Knees</string-name>
          <email>peter.knees@tuwien.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Sequential Recommendation, Music Recommendation, Contrastive Learning</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>TU Wien, Faculty of Informatics</institution>
          ,
          <addr-line>Vienna</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>With the availability of music streaming platforms, listening behavior has seen fundamental changes in the past two decades, going from mere consumption of and recommendation within personal collections to an exploration of massive catalogs. As part of this trend, collaborative filtering algorithms that exploit consumption data, user feedback, and, most recently, the sequential order of music consumption, have become indispensable. In prior work, it has been shown that the incorporation of negative feedback (skipped track information) via contrastive learning can be applied to and improve existing sequential recommendation models. In this work, we extend previous findings by investigating two notable aspects of music listening data in detail. First, we analyze popular public datasets used in music recommender systems research (LFM-1k, LFM-2B, and the Music Streaming Sessions Dataset) with respect to the evolution of consumption activity and track skipping behavior, and show strongly deviating patterns based on data creation context. Second, focusing on LFM-2B, we further study the impact of data and skipping information availability on sequential and non-sequential recommendation algorithms over the diferent years available in the data set. We observe deviating model performance using time-based subsets of LFM-2B compared to experiments on the entire dataset. In conclusion, we argue for more careful discernment and understanding of listening tasks and user intents leading to creating datasets, as well as explicitly modeling diferent types of interactions.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Music recommender systems have become central in shaping how users interact with streaming
platforms, significantly influencing music listening and discovery. These systems have evolved over
the past two decades from simple personal collection recommendations to sophisticated tools capable
of navigating vast music catalogs. This evolution has been driven by the increasing availability of
streaming data and the continuous advancement of recommendation algorithms.</p>
      <p>
        Recent studies have highlighted the importance of understanding user behavior to improve the
efectiveness of music recommender systems. For instance, Hidasi et al. (2020) emphasized the role
of contextual information in music recommendation, by modeling the whole session, to achieve more
accurate results [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. While Quadrana et al. (2020) explored how multiple user-item interactions influence
recommendation quality in a sequence-aware recommender system [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Furthermore, research by Wen
et al. (2019) has shown that incorporating user feedback, such as track skips and short plays, can
significantly enhance recommendation accuracy [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Dai et al. (2024) modeled user attention prediction
as a positive-unlabeled learning problem, where active feedback is treated as positive samples and
passive feedback is treated as unlabeled samples to increase user engagement [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>In parallel, the availability of extensive datasets has provided valuable resources for analyzing user
interactions and behavior on a large scale. For example, Yao et al. (2020) utilized these datasets to
study session-based recommendations, highlighting the need for models that can adapt to evolving
user preferences [5].
(P. Knees)</p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
      <p>Furthermore, incorporating real negative feedback, such as user skips, significantly enhances
recommender systems. Mei et al. (2024) have shown that using explicit negative samples reduces training
time and improves test accuracy [6]. Pan et al. (2023) showed that, in sequential recommendation tasks,
leveraging passive negative feedback, like video skips, provides crucial insights into user preferences [7].
Methods combining positive and passive-negative feedback through sub-interest encoders have
demonstrated superior performance, highlighting the importance of diverse feedback types for improving
accuracy and user satisfaction. In a related direction, we proposed a contrastive learning framework that
directly incorporates skip behavior as informative negative signals. This method improves sequential
music recommendation by aligning user representations with positively preferred tracks and pushing
away tracks associated with negative feedback, thus leveraging fine-grained temporal dynamics of
skips for better modeling [8].</p>
      <p>
        In this work, we aim to identify trends in music consumption and track skipping behavior across
public datasets used in music recommender systems research. By focusing on the LFM-1k, LFM-2B,
and Music Streaming Sessions Dataset (MSSD), we analyze the evolution of these behaviors and their
implications for recommender system design. Focusing on LFM-2B for in-depth evaluation, for each
contained year individually, we explore SASRec [9] as a sequential model, negative feedback enhanced
SASRec [8], and two non-sequential baseline algorithms, Weighted Regularized Matrix Factorization
(WRMF) [10] and Bayesian Personalized Ranking (BPR) [11], using the methodology introduced by
Wen et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to incorporate negative feedback.
      </p>
      <p>This work aims to fill a gap in existing research by focusing on how music listening behavior has
evolved over the past two decades and how these changes impact recommender systems. Unlike previous
studies that have largely ignored this aspect, our study examines trends across public datasets used in
music recommender systems research, specifically looking at consumption activity and track-skipping
behavior. By comparing sequential and non-sequential recommendation algorithms incorporating
negative feedback against feedback-agnostic baselines, we demonstrate the increasing importance of
integrating diferent forms of interaction into recommender models. This underscores the necessity of
understanding listening tasks and user intents to create better datasets and explicitly model various
types of interactions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Datasets</title>
      <p>For this study, we use three real-world music recommendation datasets: the Music Streaming Sessions
Dataset (MSSD) [12] using data from Spotify,1 the LFM-2B dataset [13], and the LFM-1k dataset [14, 15],
both using data from Last.fm.2. These datasets have been instrumental in uncovering patterns in music
consumption and track-skipping behavior, which are critical for refining recommendation algorithms.
2.1. Music Streaming Sessions Dataset
The MSSD contains 160 Million user sessions of 10 to 20 consecutively listened songs (&lt;60 seconds
between listens), which are uniformly sampled from a variety of contexts, such as the user’s personally
selected collections, expertly selected playlists, contextual non-personalized recommendations, and
personalized recommendations. As this dataset is pseudonymized and lacks user labels, we can treat
each session as a new user for recommendation tasks.</p>
      <p>Each listening event contains a skip label from 0-3, with 0 denoting ”no skip” and 1-3 denoting the
length of time before a user skipped a given track. This is defined as ”played very briefly”, ”played
briefly”, and ”played mostly (but not completely)”, respectively for labels 1 to 3. In this study, we would
mainly assume skip label 3 not to be a strong indicator of negative preference for the given session, so
we only assume labels 1 and 2 as true skips.
1https://open.spotify.com
2https://last.fm
The LFM-1k dataset contains 19M discrete listening events for 1000 users, containing time stamps, user
IDs, track IDs, and track names for each event. Therefore, we create implicit sessions, such that for
each user, we consider a sequence of chronological listening events with less than 20 minutes between
any individual event as a “session”. We consider skips to be any prior listening event with less than 30
seconds between its subsequent event. We additionally prune any sessions with fewer than 5 events for
a total of 650K sessions, with around 1M unique tracks. In line with the MSSD, we can discard user
labels and treat each individual session as a new user in recommendation tasks.</p>
      <p>The LFM-2B dataset contains more than 2 billion listening events for 120,322 users and 50,813,373 tracks,
which is collected over 15 years (from 2005 until 2020). The dataset includes demographic details of
users (such as age, country, and gender), metadata related to music (such as artist and track names),
and timestamps indicating the exact time when a user listened to a specific track [ 13]. Similarly to
the other datasets, we discard metadata and demographic information. Similar to the LFM-1k dataset,
we define implicit ”sessions” for each user by grouping together a chronological sequence of listening
events where the time interval between any two consecutive events is less than 20 minutes. We define
skips as any previous listening event followed by another event within a time diference of less than 30
seconds. In addition, we remove user labels and treat each session as a distinct user.</p>
      <p>We applied the same session extraction process to all yearly subsets of the data. However, the
numbers of sessions per year are significantly larger than in the 2020 subset reported earlier in [ 8] —
with over 250 million interactions recorded annually between 2012 and 2014, as shown in Figure 1.
Therefore, after creating the sessions as described earlier, we randomly sampled 100k sessions from
each year. These sampled sessions were then used to evaluate the diferent methods.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset Analysis and Comparison</title>
      <p>For clarity, we note that the following analyses are performed on the complete datasets and not the
evaluation subsets used for impact analysis (sec. 4). Figure 1 illustrates the number of listening events
and skips per year for each dataset separately. It is notable that a significant portion of the Spotify
(MSSD) dataset’s listening events was gathered in 2018, showing an uneven distribution over the
dataset’s collection period. In contrast, the data collection distribution for the LFM-1k and LFM-2B
datasets appears more even within the 5 and 15 years data collection period, respectively. In addition,
the skip percentages are relatively low, showing a maximum skip percentage of 1.5% for LFM-1k and
6.1% for LFM-2B datasets. However, the skip percentage is considerably higher for the Spotify (MSSD)
dataset, which reached ∼ 66% in 2018 and higher in other years when less data has been gathered. The
high skip rate in the data may be related to the diferent types of skip behavior that were represented in
the data [12].</p>
      <p>The overall skip comparison between the three datasets is illustrated in Figure 2. As expected, the
Spotify (MSSD) dataset has a higher skip rate than others in the same time interval. For the LFM-2B
dataset, there is no clear pattern indicating whether the number of skips increases along with the
number of listening events, or if the number of skips increases/decreases in the corresponding years.
However, the overall skip percentage for LFM-2B is lower than LFM-1k within the same time interval.
Regarding the LFM-1k dataset, there appears to be a trend of increasing skips over the years. However,
since the data is presented for only a five-year sequence, it is not conclusive evidence of a continuous
increase. While the overall skip trend for LFM-1k and LFM-2B datasets is similar, there are considerable
diferences between them, particularly shown in the subplot in Figure 2. Notably, for the LFM-2B
dataset, the overall skip rate increased after 2012, despite a decrease in overall listening events. This
observation motivates our current study, in which we focus on examining the LFM-2B dataset from
2012 onwards in more detail.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>Previously, Seshadri et al. conducted experiments on a selection of well-established sequential and
non-sequential recommendation models across diferent datasets to evaluate the impact of integrating
negative feedback [8]. Building on this, we extend the analysis by examining the performance of
selected models across multiple yearly subsets of the LFM-2B dataset. We focus on three representative
models: Bayesian Personalized Ranking (BPR) [11], Weighted Regularized Matrix Factorization (WRMF)
[10], and Self-Attentive Sequential Recommendation (SASRec) [9], where the first two models (BPR,
WRMF) are non-sequential approaches. Each model was selected for its distinct approach to modeling
user preferences—BPR and WRMF rely on matrix factorization techniques, while SASRec employs a
self-attention mechanism for sequential modeling. All models were configured according to the original
implementations and subsequently modifications incorporating negative feedback.</p>
      <p>We aim to investigate how the integration of negative feedback influences model performance across
these diferent approaches and whether temporal variation (i.e., using diferent yearly subsets) leads to
noticeable diferences in results.</p>
      <p>For SASRec, we use the same training strategies as in [8]. This includes modifications to standard
practices to incorporate negative feedback efectively into sequential recommender systems. We use
a sampled softmax approach with negative log-likelihood to handle the extensive item space, which
ranges from 300,000 to 500,000 items. During each training iteration, we sample 1,000 unseen items
to rank against the target items. This approach allows the model to adjust to an expanding subset of
items, thereby enhancing its ability to rank items accurately as training progresses.</p>
      <p>To ensure consistency, we set the sequence length limit to 20 items, splitting longer sessions into
multiple segments. Model embeddings and hidden layers are uniformly set to a dimension of 128. For
SASRec, we use two layers of self-attention with eight attention heads each. Initial parameters are
sampled from a truncated normal distribution with  = 0 and  = 1 within a range of [−0.02, 0.02]. We
optimize the  and  parameters from the set [0.1, 0.2, 0.5, 0.75, 1], choosing  = 1 and  = 0.2 for the
LFM-2B dataset. The ADAM optimizer [16] is used with a learning rate of 0.005.</p>
      <p>
        The training process varies by model. For SASRec, we reserve the final and penultimate items in
each session for testing and validation purposes. For BPR and WRMF, we focus on predicting the next
item in a sequence. We integrate negative feedback into these models, following the methods proposed
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These methods include -BL, which adjusts preference labels based on post-click feedback, and
-NR, which probabilistically samples items across diferent feedback types.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion of Results</title>
      <p>The results of evaluating on year-based subsets are shown in Table 1. As described earlier, we randomly
sampled 100k sessions for each year and calculated the skip percentage in both the sampled subset and
the original dataset. As shown in the table, the overall skip occurrence is preserved.</p>
      <p>In line with previous results we can see that sequential models outperform the non-sequential
baselines consistently. However, we can also observe less consistency regarding the impact of incorporating
negative feedback. While we can see some improvements on the sequential model in the earlier years
(2012–2014) and the last years (2019–2020), negative feedback improves the baseline models in all
subsets, albeit at a much lower performance level and dependent on the strategy chosen. For the
sequential model, there seems to be no clear trend of impact based on the skip ratio of the corresponding
year alone, indicating that diferent years exhibit diferent patterns and that a generalization of results
on individual subsets is not possible.</p>
      <p>To further investigate this indication, we assess the efect of the proportion of skip events relative
to total interactions in a dataset for one of the subsets. The results are shown in Table 2. Changing
the sampling method increases the proportion of skips in our used dataset. Consequently, methods
that incorporate negative feedback (skips) benefit from this change and achieve better performance.
This becomes more pronounced when comparing with our previous work [8], where we applied an
oversampling processing on the 2020 LFM-2B subset that resulted in a much higher skip ratio (around
∼14% compared to 3.86% in our current sampling method) and leading to a much better outcome—
e.g., HR@1 for SASRec (original and negative-feedback versions) reached .190 and .221, resp. in the
oversampled setup, a result which could only be achieved by current sampling process at HR@20. The
two non-sequential models showed a similar pattern, while oversampling led to better results for them,
they were more robust to changes and showed smaller performance diferences in comparison. These
ifndings highlight how important maintaining a higher ratio of skip interactions is for this approach,
once again confirming the conclusions in [ 8].</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>From the results obtained, we can see the potential of incorporating negative feedback, however with a
high sensitivity of algorithms regarding the underlying data. While the efect is more robustly seen
.129
.158
.167
.177
.165
.205
.222
.234
.303
.354
.369
.384
.268
.304
.315
.327
.276
.315
.329
.315
.213
.256
.277
.301
.062
.153
.198
.240
.059
.149
.191
.229
.082
.138
.165
.194
.000
.000
.001
.009
.000
.000
.001
.007
.013
.039
.065
.137
.012
.025
.063
.122
.019
.040
.091
.162
.014
.047
.073
.157
.010
.022
.055
.103
.016
.028
.087
.159
.021
.039
.094
.168</p>
      <p>BPR
.009
.026
.042
.087
.007
.022
.041
.084
.014
.047
.075
.150
.014
.043
.071
.143
.021
.063
.102
.171
.017
.036
.089
.159
.012
.038
.065
.136
.017
.038
.089
.166
.060
.120
.156
.202
-NR
.008
.024
.040
.083
.007
.020
.038
.080
.022
.067
.099
.170
.012
.028
.071
.129
.018
.051
.094
.164
.018
.035
.089
.165
.011
.022
.058
.106
.019
.054
.096
.188
.071
.135
.168
.213
with non-sequential models, their overall performance is limited. For the sequential algorithm, we see
a much higher dependency on the choice of subsets, availability of negative data, and sampling method
due to the contrastive learning approach.</p>
      <p>Other conclusions concern the bias in the data distribution over diferent years represented in
the data—not only in the descriptive analysis and across datasets but also impacting the predictive
capabilities of models. The overall performance on LFM-2B based on the originally provided 2020</p>
      <p>SASRec (100k sampled) # of skips SASRec (20 int./sess. sampled) # of skips
subset, as shown in [8], is not necessarily indicative of individual annual subsets, which by themselves
should represent diferent trends in consumption. The observed trends in the data likely reflect the
dataset creation process in addition to the consumption patterns of these years.</p>
      <p>Extrapolating from these findings also to other datasets, it is inherent that individual datasets represent
only some aspects of the larger picture and overall trends of shifting consumption patterns. The models
learned from a single dataset are therefore limited in terms of validity and generalization. However,
even with the zoo of music recommender datasets available (or partly not available anymore, e.g. [13]),
one can not easily discern listening trends and listening modalities. While certain preferences and
temporal phenomena are only represented in some datasets capturing data from the respective time,
they are strongly linked to the platforms, applications, and recommendation paradigms of that time.
This presents a conundrum as one can not simply “add up” diferent datasets to create a larger and
more “complete” pool. Instead, we first need to understand the individual backgrounds, foci, tasks, and
intents captured in and connected to the individual datasets, before devising strategies to craft a more
holistic picture of music listening preferences—if that is considered a goal worthwhile and a relevant
research question.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This research was funded in whole or in part by the Vienna Science and Technology Fund (WWTF)
[Grant ID: 10.47379/DCDH001]. For open access purposes, the author has applied a CC BY public
copyright license to any author-accepted manuscript version arising from this submission.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[5] H. Yao, J. Hu, W. Xie, Y. Huang, W. Xie, Session-based recommendation model based on multiple
neural networks hybrid extraction feature, in: 2020 IEEE International Conference on Big Data
(Big Data), IEEE, 2020, pp. 5315–5322.
[6] M. J. Mei, O. Bembom, A. F. Ehmann, Negative feedback for music personalization, in: Proceedings
of the 32nd ACM Conference on User Modeling, Adaptation and Personalization, 2024, pp. 195–200.
[7] Y. Pan, C. Gao, J. Chang, Y. Niu, Y. Song, K. Gai, D. Jin, Y. Li, Understanding and modeling
passive-negative feedback for short-video sequential recommendation, in: Proceedings of the 17th
ACM conference on recommender systems, 2023, pp. 540–550.
[8] P. Seshadri, S. Shashaani, P. Knees, Enhancing sequential music recommendation with
negative feedback-informed contrastive learning, in: Proceedings of the 18th ACM Conference on
Recommender Systems, 2024, pp. 1028–1032.
[9] W.-C. Kang, J. McAuley, Self-attentive sequential recommendation, in: 2018 IEEE international
conference on data mining (ICDM), IEEE, 2018, pp. 197–206.
[10] Y. Hu, Y. Koren, C. Volinsky, Collaborative filtering for implicit feedback datasets, in: 2008 Eighth</p>
      <p>IEEE international conference on data mining, Ieee, 2008, pp. 263–272.
[11] S. Rendle, C. Freudenthaler, Z. Gantner, L. Schmidt-Thieme, Bpr: Bayesian personalized ranking
from implicit feedback, in: Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial
Intelligence (UAI 2009), 2009. doi:10.48550/arXiv.1205.2618.
[12] B. Brost, R. Mehrotra, T. Jehan, The music streaming sessions dataset, in: L. Liu, R. W. White,
A. Mantrach, F. Silvestri, J. J. McAuley, R. Baeza-Yates, L. Zia (Eds.), The World Wide Web
Conference, WWW 2019, San Francisco, CA, USA, May 13-17, 2019, ACM, 2019, pp. 2594–2600. URL:
https://doi.org/10.1145/3308558.3313641. doi:10.1145/3308558.3313641.
[13] M. Schedl, S. Brandl, O. Lesota, E. Parada-Cabaleiro, D. Penz, N. Rekabsaz, Lfm-2b: A dataset
of enriched music listening events for recommender systems research and fairness analysis, in:
Proceedings of the 2022 Conference on Human Information Interaction and Retrieval, 2022, pp.
337–341.
[14] Ò. Celma, Music Recommendation and Discovery in the Long Tail, Springer, 2010.
[15] Ò. Celma, lastfm music recommendation dataset, 2022. URL: https://doi.org/10.5281/zenodo.</p>
      <p>6090214. doi:10.5281/zenodo.6090214.
[16] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, in: Y. Bengio, Y. LeCun (Eds.),
3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May
7-9, 2015, Conference Track Proceedings, 2015. URL: http://arxiv.org/abs/1412.6980.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          ,
          <article-title>Session-based recommendations with recurrent neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:1511.06939</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Quadrana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cremonesi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jannach</surname>
          </string-name>
          ,
          <article-title>Sequence-aware recommender systems, ACM computing surveys (CSUR) 51 (</article-title>
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Estrin</surname>
          </string-name>
          ,
          <article-title>Leveraging post-click feedback for content recommendations</article-title>
          ,
          <source>in: Proceedings of the 13th ACM Conference on Recommender Systems</source>
          , RecSys '19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , p.
          <fpage>278</fpage>
          -
          <lpage>286</lpage>
          . URL: https://doi.org/10.1145/ 3298689.3347037. doi:
          <volume>10</volume>
          .1145/3298689.3347037.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Modeling user attention in music recommendation</article-title>
          ,
          <source>in: 2024 IEEE 40th International Conference on Data Engineering (ICDE)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>761</fpage>
          -
          <lpage>774</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>