<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Systems, September</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>mendation Models for Knowledge Work Productivity on the RLK WiC Dataset</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yuuki Tachioka</string-name>
          <email>tachioka.yuki@core.d-itlab.co.jp</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Knowledge Work, Human-Centered Recommender Systems, Productivity Support, Behavioral Prediction, Benchmark Dataset</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Denso IT Laboratory</institution>
          ,
          <addr-line>13F Shintora Yasuda Bldg., 4-3-1 Shimbashi, Minato-ku, Tokyo</addr-line>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>2</volume>
      <fpage>2</fpage>
      <lpage>26</lpage>
      <abstract>
        <p>Improving the productivity of knowledge workers is a growing challenge in human-centered computing. This paper presents a benchmark suite built on the RLKWiC dataset, which captures rich behavioral logs and contextual information from real-world digital work environments. We define six practical tasks, including context detection, activity classification, and sequential prediction of web domains, event titles, and applications, designed to reflect realistic productivity support scenarios. We evaluated baseline models that incorporate event- and session-level behavior, using classification and sequence modeling techniques. The results demonstrate that modeling fine-grained user interactions yields consistent performance improvements across tasks. The proposed benchmark provides a reproducible foundation for building recommender systems that proactively support human intent, task continuity, and productivity. By releasing standardized tasks and code, the benchmark addresses the current lack of reproducible evaluation on RLKWiC. Beyond methodological contributions, these tasks provide building blocks for HR applications such as workplace analytics, training support, and well-being monitoring.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In recent years, improving the productivity of knowledge
workers (KWs) has become a highly significant issue both
socially and economically [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], particularly in the fields
of human-centered computing and recommender systems.
KWs must access various types of information, and their
productivity is influenced by multiple factors such as the
working environment, the psychological state, and the
eficiency of information access [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4, 5</xref>
        ]. Among these, quick
access to appropriate information and tools is especially
critical [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. In practice, knowledge work often follows
certain behavioral patterns, making it possible to anticipate
future tasks or required knowledge [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. More
experienced KWs tend to retrieve information more eficiently,
select tools more accurately, and switch tasks more fluently.
These observations motivate the need for intelligent
systems that can support knowledge work by estimating and
recommending the next relevant action, tool, or information
based on the behavioral history [
        <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
        ].
      </p>
      <p>
        To build such systems, high-quality datasets that capture
real-world KW behavior are essential. However, publicly
available datasets with rich semantic annotations that
relfect realistic workflows remain scarce [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Among the
few, BEHACOM [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and RLKWiC (Real-Life Knowledge
Work in Context)1 [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] are notable. Although BEHACOM
primarily records low-level user actions (e.g., keystrokes,
mouse movements), RLKWiC organizes higher-level
behavioral structures, contexts, sessions, and events, and includes
semantic metadata such as file references, web pages, and
DBpedia entities2 [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>Despite its rich structure, RLKWiC lacks well-defined
benchmark tasks and standardized baselines, which limits
its accessibility and broader use in reproducible research.
To address this gap, we define six practical tasks grounded
RecSys in HR’25: The 5th Workshop on Recommender Systems for Human
Resources, in conjunction with the 19th ACM Conference on Recommender
∗Corresponding author.
0009-0002-0587-2943 (Y. Tachioka)
© 2025 Copyright for this paper by its authors. Use permitted under Creative Commons License
Attribution 4.0 International (CC BY 4.0).</p>
      <sec id="sec-1-1">
        <title>RLKWiC dataset.</title>
        <p>
          Although prior work such as BEHACOM [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] and
RLKWiC [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] has provided valuable behavioral datasets, they
have not established standardized tasks that allow
reproducible comparisons between models. Our work fills this
gap by aligning the six benchmark tasks with established
research trends in context-aware recommender systems [
          <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
          ]
and productivity support tools [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. For example,
sequential prediction tasks are directly related to previous studies
on next-domain prediction [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], URL auto-completion, and
entity-based recommendation [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This positioning
ensures that the benchmark tasks are not arbitrary derivations
from RLKWiC, but validated scenarios grounded in existing
literature and practical needs of knowledge work support.
        </p>
        <p>Section 2 provides an overview of the RLKWiC dataset3,
Section 3 defines the six tasks, Section 4 presents the baseline
models, and Section 5 reports experimental results.
Related Work and Positioning.</p>
      </sec>
      <sec id="sec-1-2">
        <title>To our knowledge, ex</title>
        <p>isting studies that have explicitly used the RLKWiC dataset</p>
      </sec>
      <sec id="sec-1-3">
        <title>3Details are found in Section A in the appendix.</title>
        <p>CEUR</p>
        <p>
          ceur-ws.org
are primarily those conducted by its original authors,
focusing on data collection [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] and entity recommendation [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
In the absence of many independent studies on RLKWiC,
we position our benchmark within the broader context of
research on context-aware recommender systems [
          <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
          ] and
productivity support tools [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. This situates the six
benchmark tasks not only as natural extensions of RLKWiC’s
annotations, but also as representative of validated needs
in the field of knowledge work support. We believe that by
releasing standardized benchmark tasks and code, our work
will facilitate a wider adoption of RLKWiC, enabling future
studies to build on a common foundation.
        </p>
        <p>
          Beyond methodological contributions, the proposed tasks
have direct implications for human resource (HR) systems.
For example, in-context prediction could support workplace
analytics tools that detect interruptions and provide
feedback on focus patterns. The prediction of KWA labels could
enable automated profiling of employees’ work activities to
tailor training or learning support [
          <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
          ]. Assessment of
the relevance of the entity can be integrated into knowledge
management systems to recommend reference materials
that are in line with ongoing tasks [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. The prediction
of application and event title can be applied to intelligent
launchers or proactive assistants that reduce the cognitive
cost of frequent task switching [
          <xref ref-type="bibr" rid="ref21 ref22 ref23">21, 22, 23</xref>
          ]. These
scenarios illustrate how benchmark tasks can serve as building
blocks for HR applications that aim to improve employee
productivity, well-being, and training efectiveness.
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. RLKWiC Database</title>
      <p>The RLKWiC dataset captures diverse knowledge-work
behaviors with rich semantic annotations. RLKWiC employs a
three-layered hierarchical structure to model user behavior:
contexts, sessions, and events. In the highest-level layer, a
context refers to a user-defined unit of work, such as “lectures,”
“thesis writing,” or “trip planning.” This explicit
management allows analysis of context switches and multitasking.
Next, a session represents a coherent block of events within
a context. Each session is labeled as “in-context” or
“outof-context”. In addition, in-context sessions are annotated
with one or more of the 12 KWA labels. In the lowest-level
layer, an event corresponds to a user interaction. Each event
is associated with the following features.</p>
      <p>1. Event (window) title and URL: These are
concatenated into a single text string (e.g., “Quantum
Personalplanung” and “chat.openai.com”).
2. Active application: The name of the active
application used in the session (80 applications in total, e.g.,
“default browser”, “Telegram”).
3. Event cause labels: Categorical labels indicating the
trigger for event transitions (17 types in total
(Table 9)).</p>
    </sec>
    <sec id="sec-3">
      <title>3. Benchmark Tasks</title>
      <p>This section defines the six benchmark tasks derived from
the RLKWiC dataset. Here, we focus on the design and
formulation of each task, including their inputs and outputs
at the conceptual level. Implementation details such as
feature extraction, embeddings, and model architectures are
provided separately in Section 4.</p>
      <sec id="sec-3-1">
        <title>3.1. In-context Prediction</title>
        <p>In real-world work environments, users often experience
interruptions and these out-of-context activities can introduce
noise in knowledge work support systems or user behavior
analysis. Therefore, estimating whether an event is in
context is a critical task. To address this issue, we formulate a
binary classification task that determines whether a given
session is in context. A session consists of a sequence of
events grouped by a temporal window or by explicit user
operations.</p>
        <p>As shown in Table 7, the proportion of events in context
varies between participants in the RLKWiC dataset. For
example, participant p6 shows a particularly low in-context
ratio, indicating frequent out-of-context behavior. Since
tracking start and stop actions were under the participant’s
control, the observed in-context ratio may be overestimated.
Consequently, constructing a robust in-context prediction
model is essential as a foundation for task-aware support
systems. For the in-context prediction task, the input is
a session consisting of a sequence of events, where each
event is associated with three types of feature (title/URL,
cause, application). The session-level representation of this
sequence is then used for classification. The output is a
binary label: 1 if the session is considered in context and 0
otherwise.</p>
        <p>For consistency with the KWA label prediction task
described in Section 3.2, we adopt the same data split strategy
based on a five-fold cross-validation. This ensures that the
evaluation results are comparable between the two
classification tasks.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. KWA Label Prediction</title>
        <p>Each in-context session in the RLKWiC dataset is annotated
with one or more of the 12 KWA labels listed in Table 8.
These labels indicate the type of intellectual work that is
carried out during the session, for example, “Information
search,” “Authoring,” or “Networking.” While the RLKWiC
dataset provides these labels by manually analyzing
participants, such labeling is impractical in real-world applications.
Therefore, in this study, we define a multilabel classification
task to automatically predict which KWA labels apply to a
given in-context session. The task is formulated as a
multilabel classification problem: for each in-context session,
the goal is to predict a binary on/off value for each of the
12 KWA labels. Since a single session may be associated
with multiple labels, a multiclass setting is not suitable, and
a multilabel setting is adopted instead.</p>
        <p>There is a strong class imbalance in KWA label
distributions: some labels are rare. To address this, we adopt
the following experimental settings and evaluation criteria.
Since KWA labels are assigned only to in-context sessions,
both training and evaluation are limited to these sessions.
To mitigate label imbalance, we partition the data using
5-fold cross-validation such that the label distribution is as
uniform as possible across folds.</p>
        <p>The input features for this task are the same as those
described in Section 3.1. The output is a 12-dimensional
binary vector that indicates the on/off status of each KWA
label.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Relevance Estimation of DBpedia</title>
      </sec>
      <sec id="sec-3-4">
        <title>Entities</title>
        <p>
          In the RLKWiC dataset, each work session is annotated with
relevance labels that indicate how strongly the session is
related to specific DBpedia entities [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. This annotation
connects the session context to external knowledge bases,
aiming to enhance context understanding and
knowledgebased recommendation. The relevance is expressed using a
three-level label:
• Irrelevant (0): The suggested entity has no
meaningful connection to the session context.
• Relevant (1): The entity is somewhat related to the
session, but does not fully represent its context.
• Representative (2): The entity is closely aligned with
the session context and strongly represents the
session’s main topic.
        </p>
        <p>
          Bakhshizadeh et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] proposed a method that uses
RDF2Vec to generate knowledge graph embeddings of
DBpedia entities and match them with user history to estimate
relevance scores. In addition, they introduced an online
learning approach that dynamically updates these scores
based on user feedback. However, the classification
performance reported in that study remains limited. Specifically,
the F1 score for the binary task of distinguishing Irrelevant
vs Relevant+Representative (0 vs. (1,2)) was 0.686, while that
for Irrelevant+Relevant vs Representative ((0,1) vs. 2) was
0.444. Compared to a random baseline (F1 = 0.5), the latter
result indicates a particularly weak performance in
identifying representative entities. In this study, we define the task
as a 3-class classification problem (0 vs. 1 vs. 2).
Preliminary analysis showed that online learning with sequential
updates posed challenges to reliable prediction. Therefore,
we instead adopt a cross-validation setup for performance
evaluation.
        </p>
        <p>
          The input to the model is the title of the events along
with the candidate DBpedia entities and the output is the
relevance label: 0 (Irrelevant), 1 (Relevant), or 2
(Representative). All labeled entity-session pairs in the dataset are
included in the evaluation. For consistency and
comparability with previous work [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], in addition to evaluating the
model as a 3-class classifier, we also report additional binary
classification problems: one for 0 vs. (1,2) and another for
(0,1) vs. 2.
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>3.4. Sequential Domain Recommendation</title>
        <p>
          Predicting the next web domain that a user will access based
on their behavior history is a key challenge to
understanding user intent and providing contextual task support [
          <xref ref-type="bibr" rid="ref17 ref8">8, 17</xref>
          ].
If we can anticipate the next domain (e.g., google.com,
mail.yahoo.com, qiita.com) a user is likely to visit, it
provides valuable cues for inferring the type of task (e.g., web
search, email checking, document editing) and the
underlying goal. In this task, we predict the next web domain to
be accessed on the basis of a user’s chronological behavior
log. We focus on domains that appear at least three times
in the dataset, resulting in a total of 376 unique domains as
prediction targets. Practical applications of this task include
automatic URL autocompletion tailored to current tasks,
intent inference in the early stages, and dynamic
presentation of bookmarks or search help. The input consists of
a user’s recent sequence of domain-level interactions (i.e.,
accessed domains per event). The output is the next domain
predicted to be accessed.
        </p>
        <p>For each user, the behavioral history is sorted in
chronological order and split into training, validation, and test sets
using a ratio of 0.8:0.1:0.1. For evaluation, we adopt
common ranking metrics such as Hit Rate (Hit@k), Mean
Reciprocal Rank (MRR), and Normalized Discounted
Cumulative Gain (NDCG), which are widely used in
recommendation tasks to assess top- ranked outputs. The data
split method and metrics will also be used consistently in
subsequent recommendation tasks.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.5. Sequential Event Title Recommendation</title>
        <p>Compared to web domains, event (window) titles ofer a
more fine-grained signal of user activity, as they often
contain explicit information such as search queries, document
titles, or visited page contents. Thus, accurate prediction
of the next event title can enable a more precise inference
of user intent and cognitive state. In this task, we predict
the next event title to appear in a user’s session stream. We
focus on titles that occur at least three times in the
RLKWiC dataset, resulting in a total of 2,651 unique titles as
prediction candidates. Potential applications of this task
include prediction of the next page or query, automatic
display of related documents, and reminder prompts during
task switching. The input is a chronologically ordered
sequence of titles from past events, and the output is the title
predicted to occur next.</p>
      </sec>
      <sec id="sec-3-7">
        <title>3.6. Sequential Application</title>
      </sec>
      <sec id="sec-3-8">
        <title>Recommendation</title>
        <p>Users frequently switch between multiple applications to
complete their tasks. For instance, a programmer may
refer to API documentation in a Web browser while coding,
or a writer may alternate between editing documents and
communicating via chat tools. If such application switches
can be predicted, it becomes possible to proactively assist
users based on their task intent. In this task, we predict
the next application that a user will use, focusing on 64
applications that appear at least three times in the dataset.
The cSpaces application, which is used solely for tracking
purposes, is excluded from the prediction targets. Potential
applications of this task include intelligent shortcut
management, such as dynamically reordering application launch
icons or suggesting a swap-style launcher, thus reducing
the efort required for application switching. The input is a
chronologically ordered sequence of applications used and
the output is the next application predicted to be launched
or used.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Baseline Methods</title>
      <p>For each of the benchmark tasks proposed in Section 3, we
construct reasonable and comparable baseline methods. The
implementation of all baseline models and evaluation scripts
will be made publicly available via a GitHub repository4. For
the in-context prediction and KWA label prediction tasks
described in Sections 3.1 and 3.2, we design Transformer-based
classification models that utilize event-level embeddings, as
detailed in Sections 4.1 and 4.2. For the relevance estimation
of DBpedia entities task discussed in Section 3.3, we propose</p>
      <sec id="sec-4-1">
        <title>4https://github.com/DensoITLab/RLKWiC_benchmark</title>
        <p>a model that constructs a contextual representation from the
preceding sequences of events and estimates the relevance
of the entity via similarity with the corresponding entity
vector, as described in Section 4.3.</p>
        <p>
          We used the all-MiniLM-L6-v2 variant of Sentence-BERT
[
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] for all embedding steps. When the concatenated title
and URL string exceeded the model’s maximum length, we
truncated the input while retaining the most informative
segments (title and domain). Cosine similarity scores were
assigned to categories 0, 1, or 2 using a pairwise
classification layer trained on labeled examples, rather than applying
a fixed threshold.
        </p>
        <p>
          For three types of sequential recommendation tasks
composed of domain, event title, and application prediction
presented in Sections 3.4, 3.5, and 3.6, we build datasets
conforming to the atomic file format used in the RecBole
framework [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], and perform comparative evaluations
using representative sequential recommendation models such
as GRU4Rec, SASRec, and BERT4Rec. More details are
provided in Section 4.4.
        </p>
        <sec id="sec-4-1-1">
          <title>4.1. Event Embedding and Sequence</title>
        </sec>
        <sec id="sec-4-1-2">
          <title>Classification Model</title>
          <p>
            In the RLKWiC dataset, the fundamental unit of user
behavior is defined as an event, each of which is associated with
the following attribute information: title/URL, cause label,
and active application. Based on this event-level
information, we design a Transformer-based session classification
model. The overall architecture is illustrated in Figure 1.
The input consists of three components:
• Title and URL: Concatenated and embedded into a
384-dimensional vector using Sentence-BERT [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ]
(specifically, the all-MiniLM-L6-v2 model).
• Cause label and active application: Both are
onehot encoded and passed through separate fully
connected (dense) layers to obtain 16-dimensional dense
vectors.
          </p>
          <p>These components are concatenated to form a unified
embedding for each event. The sequence of event embeddings
is then fed into a Transformer Encoder in chronological
order. The final session representation is obtained by
averaging the hidden states across all events in the sequence.
At the output layer, the aggregated session representation
is passed through a fully connected layer to perform either:
Binary classification for in-context prediction, or Multi-label
classification for KWA label prediction. This model captures
short-term user intent and interest from event sequences,
integrates them into a session-level representation via the
Transformer, and makes task-specific predictions based on
this session embedding.</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>4.2. Session Embedding and Simple</title>
        </sec>
        <sec id="sec-4-1-4">
          <title>Classification Model</title>
          <p>
            An alternative approach to event-wise modeling is to treat
an entire session as a single input unit and classify it using
static features. In this section, we introduce a simple
sessionbased classification model following this principle. The
overall architecture is illustrated in Figure 2. The input to
the model consists of three types of features:
• Title and URL: All title and URL strings within a
session are concatenated in chronological order and
embedded in a 384-dimensional vector using
SentenceBERT [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ].
• Cause labels and application types: Rather than
using one-hot encodings, we adopt a bag-of-words
(BoW) representation that counts the frequency of
each label or application type occurring within the
session. This results in fixed-length vectors that are
independent of the number of events in the session.
These feature vectors (title embedding, cause BoW, and
application BoW) are concatenated and fed into a fully
connected (dense) layer to predict the target label. This simple
feedforward model does not explicitly consider the order
or structure of events but instead relies on aggregated
statistical features at the session level. Although this model
has the advantage of architectural simplicity and eficient
training, its inability to model sequential dependencies may
limit its performance compared to the Transformer-based
event-wise model introduced in Section 4.1, but it may serve
as a practical baseline in settings with limited data or
computational resources.
          </p>
        </sec>
        <sec id="sec-4-1-5">
          <title>4.3. Event-based Relevance Estimation</title>
        </sec>
        <sec id="sec-4-1-6">
          <title>Model</title>
          <p>In this section, we propose an embedding-based model for
estimating the semantic relevance between a user’s event
history and a candidate DBpedia entity. The overall
architecture is illustrated in Figure 3. Given a pair consisting of
a user’s event history and a DBpedia entity to be evaluated,
the goal is to predict how strongly the entity relates to the
user’s current context (as defined in Section 3.3). The input
to the model is the sequence of events that occurred
immediately before the current session. From each event, the title
text is extracted and embedded in a 384-dimensional
vector using Sentence-BERT. The embeddings from multiple
events are then averaged to generate a fixed-length
vector that represents the user’s contextual intent. In parallel,
the abstract associated with the candidate DBpedia entity
is retrieved and embedded using the same Sentence-BERT
model, resulting in an entity representation vector. The
similarity between the user context vector and the entity
vector is then calculated using cosine similarity5. The model
estimates the relevance level based on the similarity score,
assigning one of three labels: 0 (Irrelevant), 1 (Relevant),
or 2 (Representative). The cosine similarity scores are not
mapped to categories by fixed thresholds, but we employ a
supervised pairwise classification layer that takes the
similarity between two vectors as input and predicts one of the
three labels. This ensures that the mapping from similarity
values to discrete categories is learned from the annotated
data rather than predefined heuristics.</p>
        </sec>
        <sec id="sec-4-1-7">
          <title>4.4. Sequential Recommendation Model</title>
          <p>
            For the three sequential recommendation tasks defined in
Sections 3.4 through 3.6, we construct baseline models using
a variety of established sequential recommendation
methods. To support these tasks, we provide scripts that
automatically generate datasets in the atomic file format 6
used by the RecBole framework [
            <xref ref-type="bibr" rid="ref25">25</xref>
            ], comprising .user,
.item, and .inter files. This setup enables ranking-based
evaluation that utilizes user-level behavioral histories. As
shown in Table 10, we evaluated six representative models
implemented in RecBole in consistent settings for all three
tasks. These models span a diverse range of architectures,
including RNNs, CNNs, Transformers, and attention-based
mechanisms, enabling comparisons of diferent sequence
modeling strategies 7. This benchmark allows us to
quantify the dificulty of behavioral prediction in the context of
knowledge work support, as well as to measure performance
diferences across model types. Furthermore, comparing
architectures provides insight into model design choices
and the efectiveness of diferent feature representations.
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Discussions</title>
      <sec id="sec-5-1">
        <title>5.1. In-context Prediction Results</title>
        <p>Table 1 (event-based model) shows that the event-based
model achieved an average F1 score of 75.8% in a five-fold
cross-validation8, with a good balance between accuracy
and recall. The model also demonstrated stable performance
within the 95% confidence interval. These results suggest
that the model architecture, which sequentially incorporates
event-level information, is efective in predicting whether
a session is in context. This result shows that
determining whether sessions are in context from limited features
remains a non-trivial task.
5Alternatively, a pairwise classification model that directly takes the two
vectors as input and predicts the relevance class can also be considered.
6https://recbole.io/docs/user_guide/data/atomic_files.html
7The details of selected model architectures are shown in Appendix-B.1.
8Further fold-wise results and implementation details are provided in
Appendix-B.2.</p>
        <p>Table 1 (session-based model) shows the results of the
session-based model, which produced an average F1 score
of 68.5%, approximately 7 percentage points lower than
that of the event-based model. This performance gap can be
attributed to the session-based model’s inability to explicitly
model the sequential structure of events, relying instead on
aggregated features such as BoW and average embeddings.
Consequently, it may fail to capture dynamic behavioral
changes within a session. Although the event-based model
exhibited superior accuracy, the session-based model was
more eficient in terms of computation. Specifically, the
training and evaluation time per epoch was approximately
3.6 seconds for the session-based model, compared to 25.7
seconds for the event-based model, resulting in a roughly
7x speed-up.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. KWA Label Prediction Results</title>
        <p>Table 2 (event-based model) shows that the event-based
model achieved an average F1 score of 55.3%. Considering
that this is a multi-label classification task with 12 classes
and substantial class imbalance, the results suggest that the
model has achieved a reasonable level of accuracy. However,
recall (0.6227) is relatively higher than precision (0.5577) and
Jaccard score (0.4309), suggesting that the model tends to
overpredict some labels, leading to less precise but more
inclusive predictions.</p>
        <p>Table 2 (session-based model) shows the results for the
session-based model. In particular, it achieved a slightly
higher average F1 score of 56.1%, and the average recall
reached 79.3%, which is a substantial improvement over the
event-based model. This suggests that the session-based
model is more efective at capturing label co-occurrence
tendencies across the session, possibly due to the use of
BoW representations. As a result, it tends to reduce label
omissions and achieve higher recall.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Entity Relevance Prediction Results</title>
        <p>Table 3 summarizes the performance of the proposed model
in five-fold cross-validation under three classification
settings: two binary classification setups and one three-class
classification.</p>
        <p>In the first setting, we distinguish irrelevant entities (label
= 0) from the rest (labels = 1 or 2). The model achieved a
high average F1 score of 81.9%. This indicates a strong
semantic similarity between the user context derived from the
event history and the knowledge embedding derived from
the entity abstract. These results validate the efectiveness
of sentence transformer-based representations and cosine
similarity scoring.</p>
        <p>
          In the second setting, we isolate representative entities
(label = 2) from the other two categories (labels = 0 and 1).
Here, the model achieved an even higher average F1 score
of 85.9%. This suggests that entities labeled as
“Representative” form semantically distinct clusters in the embedding
space and that the model efectively captures this
distinction. Compared to previous work [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], where this binary
split yielded lower accuracy, our results indicate that
identifying “representativeness” can be learned as a meaningful
evaluation metric.
        </p>
        <p>In the more challenging three-class classification setting,
the model still achieved a solid average F1 score of 67.6%.
However, the relatively wide 95% confidence intervals
sugModel
event-based model
session-based model</p>
        <p>Model
event-based model
session-based model
gest variability across folds, likely due to contextual
ambiguity or subjective diferences in user annotations.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Sequential Web Domain</title>
      </sec>
      <sec id="sec-5-5">
        <title>Recommendation Results</title>
        <p>Table 4 shows the results of the web domain prediction
task (defined in Section 3.4). We compared six
sequential recommendation models implemented in RecBole. In
general, NextItNet achieved the highest prediction
performance in most metrics, including Hit@1 (0.1783), MRR, and
NDCG. Furthermore, NARM outperformed all other models
in terms of Hit@5 and NDCG@5, making it another strong
candidate among baselines. In contrast, SASRec, BERT4Rec,
and GRU4Rec demonstrated relatively lower performance.
SASRec achieved a Hit@1 of only 0.0864, indicating
potential limitations in its ability to capture contextual signals
from short-term histories. NextItNet’s architecture, which
employs dilated convolutions to model long-range
dependencies eficiently, appears particularly well-suited for this
task. Its ability to explicitly and hierarchically represent
broader contexts suggests the efectiveness of CNN-based
models in domain-level prediction. Similarly, NARM
integrates an attention mechanism into a GRU-based sequence
model, allowing it to dynamically combine both short- and
long-term user intent for improved recommendation quality.
However, self-attention-based models, such as SASRec and
BERT4Rec, may be less aligned with the characteristics of
this task. Since web domains are relatively abstract and
categorical compared to concrete item IDs or page titles,
global contextual patterns may be more important than
local sequential dependencies, potentially explaining their
underperformance in this setting.</p>
      </sec>
      <sec id="sec-5-6">
        <title>5.5. Sequential Event Title</title>
      </sec>
      <sec id="sec-5-7">
        <title>Recommendation Results</title>
        <p>the best performance in terms of Hit@5, indicating that its
self-attention mechanism is capable of dynamically
attending to contextually important events in the input history.
In contrast, GRU4Rec showed the lowest performance in
all metrics. This result highlights the limitations of
simple RNNs in handling highly diverse and semantically rich
output spaces such as event titles.</p>
      </sec>
      <sec id="sec-5-8">
        <title>5.6. Sequential Application</title>
      </sec>
      <sec id="sec-5-9">
        <title>Recommendation Results</title>
        <p>Table 6 shows the results of the application prediction task
(defined in Section 3.6). In this task, all models
demonstrated relatively high performance. NARM, NextItNet, and
GRU4Rec emerged as the top performers. Application
prediction likely depends on the short-term context, and NARM
scored the best on all metrics except Hit@1, indicating its
strong efectiveness for this task. NextItNet also
consistently ranked high in all metrics. GRU4Rec achieved the best
Hit@1 score (0.6040). Importantly, all three leading
models (NARM, NextItNet, and GRU4Rec) achieved more than
85% in Hit@5, indicating practical feasibility for
applicationswitching support systems. For example, the top five
predicted applications could be presented as shortcut buttons,
significantly reducing the user’s switching efort.
Application usage is closely tied to user tasks and workflow
structure, and patterns are often stable. Therefore, sequential
models are particularly well suited to this task.</p>
      </sec>
      <sec id="sec-5-10">
        <title>5.7. Limitation</title>
        <p>
          Despite its contribution, our work has several limitations.
First, the RLKWiC dataset was collected from only eight
university students in Germany [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], which restricts the
demographic and occupational diversity of the sample. As
a result, the generalizability of the reported benchmark
performance remains limited. Second, participants had control
over recording start and stop actions, which may have led to
biases in the proportion of in-context versus out-of-context
sessions. Third, the dataset exhibits strong label imbalance,
especially for rare knowledge work activities (Table 8) and
representative DBpedia entities (Table 3), which
complicates model training. Finally, the relatively wide 95%
conifdence intervals observed in several tasks (e.g., in-context
prediction and entity relevance estimation) indicate
vari
        </p>
        <p>NDCG@5</p>
        <p>NDCG@10
ance across folds and highlight the need for larger-scale
datasets. Future work should extend evaluations to
crossuser splits, where the full data of certain participants is
held out, to better assess the generalization capability of
predictive models.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>This study presents a practical approach to building a
benchmark suite designed to support knowledge work.
Leveraging RLKWiC’s rich semantics and multilayered structure,
we defined the following six benchmark tasks: (1) In-context
prediction, (2) KWA label classification, (3) Relevance
estimation with DBpedia entities, (4) Web domain prediction,
(5) Event title prediction, and (6) Application prediction.
For each task, we proposed appropriate baseline models:
event/session-level embedding-based classifiers, a relevance
estimation model for entity matching, and sequential
recommendation models.</p>
      <p>The results indicate that Transformer-based models
operating on event sequences achieved strong performance
for in-context detection and KWA classification. Sentence
embedding-based similarity scoring proved efective for
relevance estimation. Sequential models such as NextItNet
and NARM achieved high accuracy in predicting domains,
events, and applications.</p>
      <p>These benchmark tasks and their results provide a solid
foundation for future research in knowledge work support
and behavioral prediction systems. We hope that this work
serves as a standard reference point.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author used
ChatGPT-4 and writefull in order to: Grammar and spelling
check. After using these services, the author reviewed and
0.8553
0.8421
0.8654
0.7623
0.8609
0.8811
0.2438
0.2237
0.1271
0.2535
0.2503
0.2521</p>
      <p>MRR@10</p>
      <p>NDCG@5</p>
      <p>NDCG@10</p>
      <p>Recall@5</p>
      <p>Recall@10</p>
      <p>NDCG@5</p>
      <p>NDCG@10</p>
    </sec>
    <sec id="sec-8">
      <title>Appendix</title>
    </sec>
    <sec id="sec-9">
      <title>A. Details of RLKWiC Database</title>
      <p>The RLKWiC dataset is a highly valuable resource that
captures diverse knowledge-work behaviors in real-world
knowledge-work environments with rich semantic
annotations. The dataset was collected over approximately two
months, from the end of May to early July 2023, from eight
students (aged 23 to 35) at the University of
KaiserslauternLandau in Germany. Data collection was carried out
using two tools: cSpaces, which allows participants to
explicitly manage their work contexts, and the User Activity
Tracker, which automatically records user interaction logs.</p>
      <p>As a result, a wide range of information was recorded in
detail, including active window titles, applications used,
clipboard contents, file operations, browsing history, and
context switches. For privacy protection, participants had
full control over recording, deleting collected data, and
applying anonymization.</p>
      <p>Table 7 provides aggregated statistics for the RLKWiC
dataset, showing the number of contexts, sessions, events,
total durations, and in-context ratios per participant. These
ifgures illustrate the variation in working styles and
contexttracking practices. For example, participant p6 has a
significantly lower in-context ratio (25.2%), suggesting frequent
interruptions or less precise context labeling, while others
such as p5 and p7 show very high in-context ratios above
95%. The table also highlights that all but two participants
recorded more than 10,000 minutes of events, ensuring
sufifcient data volume for analysis.</p>
      <p>RLKWiC employs a three-layered hierarchical structure
to model user behavior: contexts, sessions, and events. In the
highest-level layer, a context refers to a user-defined unit of
work, such as “lectures,” “thesis writing,” or “trip planning.”
Through cSpaces, users could flexibly create new contexts
or switch between existing ones depending on their current
task. This explicit management enables analysis of
context switches and multitasking. Next, a session represents
a coherent block of events within a context. Each session
is labeled as “in-context” or “out-of-context”. In-context
sessions are defined as sessions are semantically aligned
with the user’s self-declared current task or objective. In
contrast, out-of-context sessions include unrelated or
interruptive sessions, such as administrative operations or
personal browsing, that are not directly tied to the ongoing
work context.</p>
      <p>In-context sessions are further annotated with one or
more Knowledge Work Activity (KWA) labels. Table 8 lists
the 12 KWA categories (e.g., “Information search,”
“Learning,” “Authoring”) and their frequency across the dataset.</p>
      <p>This annotation enables task-level analysis such as focus
distribution and label co-occurrence across sessions. It also
serves as the target for the KWA label classification tasks in
Section 3.2.</p>
      <p>ID
p1
p2
p3
p4
p5
p6
p7
p8</p>
      <p>In the lowest-level layer, an event corresponds to a user
interaction with timestamp, such as application launches,
window switches, file operations, or clipboard actions. These
ifne-grained logs are crucial for mining behavior patterns,
estimating user focus, and building predictive interaction
models. Each event is associated with the following features.</p>
      <p>1. Event (window) title and URL: These are
concate</p>
      <p>nated into a single text string.
2. Active application: The name of the active
applica</p>
      <p>tion used in the session (80 applications in total).
3. Event cause labels: Categorical labels indicating the
trigger for event transitions (17 types in total), as
detailed in Table 9.</p>
      <p>Table 9 summarizes all cause labels recorded in the dataset
and their frequency. The most frequent cause is “active
window changed” with more than 42,000 occurrences,
relfecting the application or window switch behavior. Other
notable causes include web visits (focused or visible),
context switches, file drops, and tagging operations. These
categorical labels provide rich signals to understand user
intent and trigger conditions in multi-tasking environments.</p>
      <p>In addition to the hierarchical structure, RLKWiC
includes metadata on the documents accessed by users. For
local files and web pages, metadata such as filenames, file
paths, visited URLs, page titles, and access timestamps are
recorded and linked to the corresponding context. This
enables a comprehensive analysis of information-seeking
behavior and reference history. Furthermore, the dataset
is enriched with lexical and semantic features. It includes
bag-of-words and stemmed tokens extracted from
documents and webpages, as well as entity links to DBpedia.</p>
      <p>
        This allows documents to be associated with concepts such
as organizations, locations, or academic topics,
facilitating semantic search, knowledge graph construction, and
entity-based recommendations [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. In summary, RLKWiC
is a uniquely comprehensive dataset that integrates layered
information on work contexts, behavior logs, reference
materials, and semantic structures, ofering a solid foundation
for analysis and support of knowledge work.
42,701 an item was removed from the 407 new context was created 58
context’s activity history
8,682 a folder was rebirthed to a context 281 more context’s activity history 57
(by adding tags) was browsed
7,545 new item was added to the con- 281 an item from the context’s activ- 55
text ity history was opened
855 new tag was added 226 an item was removed from the 53
context
664 a file was dropped on the cSpaces 113 tag was removed from the con- 14
sidebar text
464 an item from the context was 95
opened
      </p>
    </sec>
    <sec id="sec-10">
      <title>B. Supplemental information of experiments</title>
      <sec id="sec-10-1">
        <title>B.1. Brief description of sequential recommendation models</title>
        <p>Table 10 lists the six sequential recommendation models
employed in our benchmark experiments. These models span
a diverse range of architectural paradigms:
Transformerbased (SASRec, BERT4Rec), RNN-based (GRU4Rec, NARM),
CNN-based (NextItNet) and hybrid methods (FPMC) -
allowing for a broad comparison of sequence modeling strategies.
By including this variety, we aim to evaluate how diferent
temporal modeling mechanisms (e.g., self-attention,
recurrent updates, dilated convolutions, or Markov transitions)
impact prediction performance across multiple behavioral
targets (web domains, event titles, and applications). This
diversity also helps identify which model families are best
suited for diferent aspects of knowledge work prediction.</p>
      </sec>
      <sec id="sec-10-2">
        <title>B.2. Details of five-fold cross-validation results</title>
        <p>Tables 11 and 12 detail the fold-wise performance of the two
models used in the in-context prediction task: the
eventbased and session-based classifiers, respectively. Inspecting
these fold-level results, we observe that the event-based
model exhibits relatively stable performance across all folds,
with an accuracy ranging between 0.7481 and 0.8231, and
the F1 score staying within a narrow band of 0.7249 to 0.8035.
This consistency across partitions suggests that the model
generalizes well and is not overly sensitive to variations
in the training/test splits. In contrast, the session-based
model shows a greater degree of variability. For example,
fold2 yields substantially lower accuracy (0.6794) and F1
score (0.6272), whereas fold5 shows much stronger
performance (Accuracy = 0.7538, F1 = 0.7252). This implies that the
session-based model is more afected by the distribution of
features across folds, probably due to its reliance on coarse
aggregate features rather than sequential structure. The
fold-level breakdown provides insight into the robustness
and sensitivity of each model under diferent data partitions,
complementing the averaged results presented in the main
text.</p>
        <p>Tables 13 and 14 present the fold-wise performance
results for the KWA label prediction task using the event-based
and session-based models, respectively. In the case of the
event-based model, the performance remains relatively
stable across folds, with F1 scores ranging from 0.5093 (fold3)
to 0.5961 (fold2). This modest variation suggests that the
model consistently captures key patterns in event sequences
for multilabel classification, although some folds (e.g., fold3)
may sufer from limited label diversity or skewed
distributions. The session-based model, while achieving a slightly
higher mean F1 score overall, shows much larger variability
between folds. In particular, fold4 achieves an F1 score of
0.6041 with a very high recall (0.9162), whereas fold1 drops
significantly to 0.4865. This discrepancy indicates that the
session-based model is more sensitive to the distribution
of co-occurring labels across folds. Its reliance on
aggregated bag-of-words representations may lead to overfitting
or undergeneralization depending on the composition of
the validation set. These fold-level diferences highlight the
challenges of multilabel prediction under label imbalance
and interlabel dependencies, and point to the need for
stratified or label-aware data partitioning in future experiments.</p>
        <p>Table 15 summarizes the fold-wise evaluation results for
the DBpedia entity relevance prediction task under three
classification settings: binary (0 vs. (1,2)), binary ((0,1) vs.
2) and three-class (0 vs. 1 vs. 2). Across all settings, the
fold-level breakdown reveals meaningful diferences in task
dificulty and model consistency.</p>
        <p>• In the setting 0 vs. (1,2), the F1 scores are relatively
stable across folds (ranging from 0.7571 to 0.8737),
indicating that the model can reliably distinguish
irrelevant entities from those with some relevance.
The highest fold4 score suggests that this partition
had particularly clean or separable training
examples.
• In the setting more challenging (0,1) vs. 2, the F1
score varies more widely, from 0.7812 (fold3) to
0.9203 (fold4), which implies that identifying
“representative” entities is more sensitive to the
composition of the fold. Folds with fewer strongly
representative entities may hinder classifier calibration.
• The three-class classification setting exhibits the
largest performance fluctuation between folds,
with F1 scores ranging from 0.5579 (fold3) to
0.8101 (fold4). This variability reflects the
increased ambiguity in distinguishing relevant but
non-representative entities (class 1) from the other
two classes, especially when user annotations are
subjective or unevenly distributed.</p>
        <p>These fold-wise results emphasize the inherent dificulty of
ifne-grained entity relevance classification and suggest that
future work may benefit from fold stratification with respect
to entity-type distributions or additional regularization to
reduce variability.</p>
        <p>Accuracy</p>
        <p>Precision
0.3376
0.4451
0.4676
0.4666
0.3619
0.5143
0.5960
0.6228
0.4894
0.4175</p>
        <p>Recall
0.6254
0.7656
0.8000
0.9162
0.8553</p>
        <p>F1-Score
0.4865
0.5898
0.6123
0.6041
0.5132</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Davenport</surname>
          </string-name>
          ,
          <article-title>Thinking for a living: How to get better performance and results from knowledge workers</article-title>
          , Harvard Business Press, Brighton, Massachusetts, USA,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Drucker</surname>
          </string-name>
          ,
          <article-title>Management Challenges for the 21st Century</article-title>
          , HarperBusiness, New York,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Palvalin</surname>
          </string-name>
          ,
          <article-title>What matters for knowledge work productivity?</article-title>
          ,
          <source>Employee Relations</source>
          <volume>41</volume>
          (
          <year>2019</year>
          )
          <fpage>209</fpage>
          -
          <lpage>227</lpage>
          . URL: https://doi.org/10.1108/ER-04-2017-0091. doi:
          <volume>10</volume>
          . 1108/ER-04-2017-0091.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Züger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Corley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Meyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Fritz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Shepherd</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Augustine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Francis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kraft</surname>
          </string-name>
          , W. Snipes,
          <article-title>Reducing interruptions at work: A large-scale field study of flowlight</article-title>
          ,
          <source>in: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, CHI '17</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2017</year>
          , pp.
          <fpage>61</fpage>
          -
          <lpage>72</lpage>
          . URL: https://doi.org/10.1145/ 3025453.3025662. doi:
          <volume>10</volume>
          .1145/3025453.3025662.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Mark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gudith</surname>
          </string-name>
          , U. Klocke,
          <article-title>The cost of interrupted work: more speed and stress</article-title>
          ,
          <source>in: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI '08</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2008</year>
          , pp.
          <fpage>107</fpage>
          -
          <lpage>110</lpage>
          . URL: https://doi.org/10.1145/1357054.1357072. doi:
          <volume>10</volume>
          . 1145/1357054.1357072.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Geng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Knowledge recommendation for workplace learning: a system design and evaluation perspective</article-title>
          ,
          <source>Internet Research</source>
          <volume>30</volume>
          (
          <year>2020</year>
          )
          <fpage>243</fpage>
          -
          <lpage>261</lpage>
          . URL: https: //doi.org/10.1108/INTR-07-2018-0336. doi:
          <volume>10</volume>
          .1108/ INTR-07-2018-0336.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>E.</given-names>
            <surname>Horvitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kadie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Paek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hovel</surname>
          </string-name>
          ,
          <article-title>Models of attention in computing and communication: from principles to applications, Commun</article-title>
          . ACM
          <volume>46</volume>
          (
          <year>2003</year>
          )
          <fpage>52</fpage>
          -
          <lpage>59</lpage>
          . URL: https://doi.org/10.1145/636772.636798. doi:
          <volume>10</volume>
          . 1145/636772.636798.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Adomavicius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ricci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tuzhilin</surname>
          </string-name>
          ,
          <article-title>Context-aware recommender systems</article-title>
          ,
          <source>AI</source>
          Magazine
          <volume>32</volume>
          (
          <year>2011</year>
          )
          <fpage>67</fpage>
          -
          <lpage>80</lpage>
          . URL: https://ojs.aaai.org/ aimagazine/index.php/aimagazine/article/view/2364. doi:
          <volume>10</volume>
          .1609/aimag.v32i3.
          <fpage>2364</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Mateos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bellogin</surname>
          </string-name>
          ,
          <article-title>A systematic literature review of recent advances on context-aware recommender systems</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          <volume>58</volume>
          (
          <year>2024</year>
          )
          <article-title>20</article-title>
          . URL: https://doi.org/10.1007/s10462-024-10939-4. doi:
          <volume>10</volume>
          .1007/s10462-024-10939-4.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Teevan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Dumais</surname>
          </string-name>
          , E. Horvitz,
          <article-title>Personalizing search via automated analysis of interests and activities</article-title>
          ,
          <source>in: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '05,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2005</year>
          , pp.
          <fpage>449</fpage>
          -
          <lpage>456</lpage>
          . URL: https://doi.org/10.1145/ 1076034.1076111. doi:
          <volume>10</volume>
          .1145/1076034.1076111.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Chamun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pinheiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jornada</surname>
          </string-name>
          , J.
          <string-name>
            <surname>a. B. S. de Oliveira</surname>
          </string-name>
          , I. Manssour,
          <article-title>Extracting web content for personalized presentation</article-title>
          ,
          <source>in: Proceedings of the 2014 ACM Symposium on Document Engineering</source>
          , DocEng '14,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2014</year>
          , pp.
          <fpage>157</fpage>
          -
          <lpage>164</lpage>
          . URL: https://doi. org/10.1145/2644866.2644871. doi:
          <volume>10</volume>
          .1145/2644866. 2644871.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Tanaka</surname>
          </string-name>
          ,
          <article-title>Web knowledge extraction for improving search</article-title>
          ,
          <source>in: Proceedings of the 2nd International Conference on Ubiquitous Information Management and Communication</source>
          , ICUIMC '08,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2008</year>
          , pp.
          <fpage>140</fpage>
          -
          <lpage>145</lpage>
          . URL: https://doi.org/10.1145/1352793.1352823. doi:
          <volume>10</volume>
          . 1145/1352793.1352823.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>M. M. Kamel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Gil-Solla</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ramos-Carber</surname>
          </string-name>
          ,
          <article-title>Tasks recommendation in crowdsourcing based on workers' implicit profiles and performance history</article-title>
          ,
          <source>in: Proceedings of the 9th International Conference on Software and Information Engineering</source>
          , ICSIE '20,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          , pp.
          <fpage>51</fpage>
          -
          <lpage>55</lpage>
          . URL: https://doi.org/10.1145/3436829. 3436834. doi:
          <volume>10</volume>
          .1145/3436829.3436834.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bakhshizadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Jilek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schröder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Maus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dengel</surname>
          </string-name>
          ,
          <article-title>Data collection of real-life knowledge work in context: The RLKWiC dataset</article-title>
          , in: S. Li (Ed.),
          <source>Information Management</source>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>277</fpage>
          -
          <lpage>290</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>P. M. Sanchez Sanchez</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jorquera Valero</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Zago</surname>
            ,
            <given-names>A. Huertas</given-names>
          </string-name>
          <string-name>
            <surname>Celdran</surname>
            ,
            <given-names>L. Fernandez</given-names>
          </string-name>
          <string-name>
            <surname>Maimo</surname>
            ,
            <given-names>E. Lopez</given-names>
          </string-name>
          <string-name>
            <surname>Bernal</surname>
            ,
            <given-names>S. Lopez</given-names>
          </string-name>
          <string-name>
            <surname>Bernal</surname>
            ,
            <given-names>J. Martinez</given-names>
          </string-name>
          <string-name>
            <surname>Valverde</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Nespoli</surname>
            ,
            <given-names>J. Pastor</given-names>
          </string-name>
          <string-name>
            <surname>Galindo</surname>
          </string-name>
          ,
          <string-name>
            <surname>Angel L. Perales Gomez</surname>
            ,
            <given-names>M. Gil</given-names>
          </string-name>
          <string-name>
            <surname>Perez</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <article-title>Martinez Perez, BEHACOM - a dataset modelling users' behaviour in computers, Data in Brief 31 (</article-title>
          <year>2020</year>
          )
          <article-title>105767</article-title>
          . URL: https://www.sciencedirect.com/ science/article/pii/S2352340920306612. doi:https:// doi.org/10.1016/j.dib.
          <year>2020</year>
          .
          <volume>105767</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bakhshizadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Maus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dengel</surname>
          </string-name>
          ,
          <article-title>Context-based entity recommendation for knowledge workers: Establishing a benchmark on real-life data</article-title>
          ,
          <source>in: Proceedings of the 18th ACM Conference on Recommender Systems</source>
          , RecSys '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , pp.
          <fpage>654</fpage>
          -
          <lpage>659</lpage>
          . URL: https://doi.org/10.1145/3640457.3688068. doi:
          <volume>10</volume>
          . 1145/3640457.3688068.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>W.</given-names>
            <surname>Hussein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. F.</given-names>
            <surname>Gharib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Ismail</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. G.-H. M. Mostafa</surname>
          </string-name>
          ,
          <article-title>A user-concept matrix clustering algorithm for eficient next page prediction</article-title>
          ,
          <source>Int. J. Knowl. Web Intell</source>
          .
          <volume>5</volume>
          (
          <year>2016</year>
          )
          <fpage>208</fpage>
          -
          <lpage>229</lpage>
          . URL: https://doi. org/10.1504/IJKWI.
          <year>2016</year>
          .
          <volume>078718</volume>
          . doi:
          <volume>10</volume>
          .1504/IJKWI.
          <year>2016</year>
          .
          <volume>078718</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Granitzer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Rath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kröll</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Seifert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ipsmiller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Devaurs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Weber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lindstaedt</surname>
          </string-name>
          ,
          <article-title>Machine learning based work task classification</article-title>
          ,
          <source>Journal of Digital Information Management</source>
          <volume>7</volume>
          (
          <year>2009</year>
          )
          <fpage>306</fpage>
          -
          <lpage>313</lpage>
          . URL: https://hal.science/hal-00872101.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          ,
          <article-title>Work Analysis in the Knowledge Economy</article-title>
          , Palgrave Macmillan,
          <year>2019</year>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>319</fpage>
          -94448-7.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Jacucci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Daee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Vuong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Andolina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Klouche</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sjöberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ruotsalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kaski</surname>
          </string-name>
          ,
          <article-title>Entity recommendation for everyday digital tasks</article-title>
          ,
          <source>ACM Trans. Comput.- Hum. Interact</source>
          .
          <volume>28</volume>
          (
          <year>2021</year>
          ). URL: https://doi.org/10.1145/ 3458919. doi:
          <volume>10</volume>
          .1145/3458919.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kersten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. C.</given-names>
            <surname>Murphy</surname>
          </string-name>
          ,
          <article-title>Reducing friction for knowledge workers with task context</article-title>
          ,
          <source>AI Mag</source>
          .
          <volume>36</volume>
          (
          <year>2015</year>
          )
          <fpage>33</fpage>
          -
          <lpage>41</lpage>
          . URL: https://doi.org/10.1609/aimag.v36i2.2581. doi:
          <volume>10</volume>
          .1609/aimag.v36i2.
          <fpage>2581</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Aliannejadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Crestani</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Context-aware target apps selection and recommendation for enhancing personal mobile assistants</article-title>
          ,
          <source>ACM Trans. Inf. Syst</source>
          .
          <volume>39</volume>
          (
          <year>2021</year>
          ). URL: https://doi.org/10.1145/ 3447678. doi:
          <volume>10</volume>
          .1145/3447678.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Chis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hughes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sycara</surname>
          </string-name>
          ,
          <article-title>The cognitive load - productivity tradeof in task switching</article-title>
          ,
          <source>Proceedings of the Human Factors and Ergonomics Society Annual Meeting</source>
          <volume>67</volume>
          (
          <year>2023</year>
          )
          <fpage>666</fpage>
          -
          <lpage>671</lpage>
          . URL: https: //doi.org/10.1177/21695067231193677. doi:
          <volume>10</volume>
          .1177/ 21695067231193677.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERT-networks</article-title>
          , in: K. Inui,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          Wan (Eds.),
          <source>Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>3982</fpage>
          -
          <lpage>3992</lpage>
          . URL: https://aclanthology.org/D19-1410/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          - 1410.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Bian,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          , RecBole
          <volume>2</volume>
          .
          <article-title>0: Towards a more up-to-date recommendation library</article-title>
          ,
          <source>in: Proceedings of the 31st ACM International Conference on Information &amp; Knowledge Management, CIKM '22</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          , pp.
          <fpage>4722</fpage>
          -
          <lpage>4726</lpage>
          . URL: https://doi.org/10.1145/3511808.3557680. doi:
          <volume>10</volume>
          . 1145/3511808.3557680.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>W.-C. Kang</surname>
            ,
            <given-names>J. McAuley</given-names>
          </string-name>
          ,
          <article-title>Self-attentive sequential recommendation</article-title>
          ,
          <source>in: 2018 IEEE International Conference on Data Mining (ICDM)</source>
          ,
          <source>IEEE Computer Society</source>
          , Los Alamitos, CA, USA,
          <year>2018</year>
          , pp.
          <fpage>197</fpage>
          -
          <lpage>206</lpage>
          . URL: https://doi.ieeecomputersociety.
          <source>org/ 10</source>
          .1109/ICDM.
          <year>2018</year>
          .
          <volume>00035</volume>
          . doi:
          <volume>10</volume>
          .1109/ICDM.
          <year>2018</year>
          .
          <volume>00035</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sun</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ou</surname>
          </string-name>
          , P. Jiang,
          <article-title>BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer</article-title>
          ,
          <source>in: Proceedings of the 28th ACM International Conference on Information and Knowledge Management</source>
          , CIKM '19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , pp.
          <fpage>1441</fpage>
          -
          <lpage>1450</lpage>
          . URL: https://doi. org/10.1145/3357384.3357895. doi:
          <volume>10</volume>
          .1145/3357384. 3357895.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>B.</given-names>
            <surname>Hidasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Baltrunas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tikk</surname>
          </string-name>
          ,
          <article-title>Session-based recommendations with recurrent neural networks</article-title>
          ,
          <source>arXiv preprint arXiv:1511.06939</source>
          ,
          <year>2016</year>
          . URL: https://arxiv.org/abs/1511.06939. arXiv:
          <volume>1511</volume>
          .
          <fpage>06939</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rendle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Freudenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Schmidt-Thieme</surname>
          </string-name>
          ,
          <article-title>Factorizing personalized Markov chains for next-basket recommendation</article-title>
          ,
          <source>in: Proceedings of the 19th International Conference on World Wide Web, WWW '10</source>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2010</year>
          , pp.
          <fpage>811</fpage>
          -
          <lpage>820</lpage>
          . URL: https://doi.org/10.1145/ 1772690.1772773. doi:
          <volume>10</volume>
          .1145/1772690.1772773.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>F.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karatzoglou</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Arapakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Jose</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>A simple convolutional generative network for next item recommendation</article-title>
          ,
          <source>in: Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining</source>
          , WSDM '19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , pp.
          <fpage>582</fpage>
          -
          <lpage>590</lpage>
          . URL: https://doi.org/10.1145/3289600.3290975. doi:
          <volume>10</volume>
          . 1145/3289600.3290975.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lian</surname>
          </string-name>
          , J. Ma,
          <article-title>Neural attentive session-based recommendation</article-title>
          ,
          <source>in: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management</source>
          , CIKM '17,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2017</year>
          , pp.
          <fpage>1419</fpage>
          -
          <lpage>1428</lpage>
          . URL: https://doi.org/10.1145/3132847. 3132926. doi:
          <volume>10</volume>
          .1145/3132847.3132926.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>