=Paper=
{{Paper
|id=Vol-3178/CIRCLE_2022_paper_10
|storemode=property
|title=Accounting for User’s Knowledge and Search Goals in Information Retrieval Evaluation - Extended Abstract
|pdfUrl=https://ceur-ws.org/Vol-3178/CIRCLE_2022_paper_10.pdf
|volume=Vol-3178
|authors=Dima El-Zein,Célia da-Costa-Pereira
|dblpUrl=https://dblp.org/rec/conf/circle/ZeinP22
}}
==Accounting for User’s Knowledge and Search Goals in Information Retrieval Evaluation - Extended Abstract==
Accounting for User’s Knowledge and Search Goals in
Information Retrieval Evaluation - Extended Abstract
Dima El-Zein, Célia da-Costa-Pereira
Université Côte d’Azur, Labortatoire I3S, CNRS, UMR 7271, France
Accounting for the user’s cognitive aspects in the information retrieval field is still considered a challenge
up until our days. Knowing that recent frameworks are trying to fill this gap, the bigger challenge
remains to evaluate those frameworks and to measure the results’ relevance in view of the user cognition.
The majority of existing evaluation measures often consider isolated document-query environments.
Traditional evaluation measures, for example, precision and recall, are not suitable to evaluate the quality
of such IR algorithms. Goffman et al. recognised that the relevance of a document must be determined
with respect to the documents appearing before it while Boyce et al. claimed that the change a document
makes in the knowledge state must be reflected in the choice of document for the second position. The
few measures that account for the user’s cognitive aspects when evaluating the “relevance" of a result or
ranking are limited to one search session, one query, or one search goal. The evaluation metric proposed
by Clarke et al. for example systematically rewarded novelty and diversity; however, only one interaction
session was considered. Also, the existing evaluation methods do not consider that the user can submit
different queries. They neither track nor update the user’s knowledge. Filling this gap is the main aim of
our work.
We present our ongoing work on an evaluation framework for cognitive information retrieval
systems that considers the user’s previous knowledge and his/her search goal. The user’s knowledge is a
dynamic representation that changes after reading any result. Similarly, the progress towards the goal is
re-assessed. The mentioned representations are not limited to one query or one search session but rather
constructed throughout all the user-system interactions. This paper focuses on learning-related search
tasks where the user’s need (or goal) is to learn about some topic 𝑇 . We adopted the user’s learning
model named vocabulary learning, which takes place at the lower level of Bloom’s taxonomy. During a
vocabulary learning task, in an information retrieval context, users interact with the search system to
acquire keywords related to a topic T.
Search and Learning Goals: This model considers a user’s information need as achieved when
he/she learns about the keywords related to 𝑇 . The learning goal is represented as a weighted set
of keywords 𝑉 𝐾 𝑇 = {𝑣𝑘1 , . . . , 𝑣𝑘𝑚𝑇 }, which we call vocabulary keyword need; 𝑚𝑇 is the number
of keywords representing the need. For example, the need to know about the topic “deep learning"
might involve learning about machine learning, programming, and calculus. The keywords needed
to learn a topic could be extracted implicitly from the query expressing the user’s need, or explicitly
from the search task text. A search task, for example, could be a school assignment, or a set of
questions to answer. Once the keywords to be learned are defined, the number of occurrences the
user has to read must also be decided. We adopt the hypothesis that to satisfy a search need or
learning goal, any user should read a given number of occurrences of those keywords. We define the
corresponding distribution vector vocabulary occurrence need 𝑉 𝑂𝑇 = {𝑣𝑜1 , . . . , 𝑣𝑜𝑚𝑇 }, (𝑣𝑜𝑖 ∈ N),
noting how many instances of every keyword of 𝑉 𝐾 a user has to read to achieve the learning goal
𝑇 . To calculate the number of instances to be read, we suppose 𝑡𝑛 is the total number of keywords
the user had to read, then its distribution among the keywords will be proportional to the weight
𝑤𝑖 of each keyword in a corpus. We consider in this paper a personalised approach that accounts
for the user’s previous knowledge. Indeed, the user might already have some previous knowledge
about those keywords, for example, having already read them in another document in the previous session.
User’s Knowledge The user’s knowledge is represented by the set of keywords he/she is knowl-
edgeable about, stored in a knowledge base 𝐾𝐵 = {𝑘1 , . . . , 𝑘𝑝 }. The amount of the user’s knowledge
about a word is measured by the cumulative count of the number of times the user has seen the word
before; this measure is stored in 𝐾𝑂 = {𝑘𝑜1 , . . . , 𝑘𝑜𝑝 }. We also assume that the user’s knowledge
of a keyword monotonically increases with each instance of it that the user reads. We made the hy-
pothesis that users acquire their knowledge from the documents they read. When the user reads a
document 𝑑 = {𝑑𝑘1 , . . . , 𝑑𝑘𝑞 }, the keywords 𝑑𝑘𝑖 contained in the document are taken into account in
the user’s knowledge - added to the user’s knowledge base KB. The keywords occurrence is represented
by 𝑑𝑜 = {𝑑𝑘𝑜1 , . . . , 𝑑𝑘𝑜𝑞 }. Several methods can be used to extract the keywords from the documents,
for example, RAKE—Rapid Automatic Keyword Extraction Algorithm, TF-IDF, or KeyBERT. The user’s
previous knowledge will impact the number of occurrences he/she still has to read. Notice that, in
order to solve the “cold start problem”, we assume that the user’s knowledge base is empty at the very
first session, i.e. we suppose that the user did not read any document before the first search session
(𝐾𝐵 = ∅). However, unlike in the literature, we relax this hypothesis for the subsequent search sessions
and suppose that the user can have some previous knowledge acquired in the previous sessions. In
this case, we have that 𝐾𝐵 ̸= ∅ i.e. 𝐾𝐵 = {𝑘1 , . . . , 𝑘𝑝 }. The i𝑡ℎ element of the user’s vocabulary
occurrence need, must consider the previous user’s knowledge. The vocabulary need at the beginning of
one session is calculated using the following equation, with 𝑣𝑘𝑖 = 𝑘𝑗 .
if 𝑘𝑜𝑗 = 0;
⎧
⎨ 𝑣𝑜𝑖 ,
𝑣𝑜𝑢𝑖 = 0, if 𝑘𝑜𝑗 > 𝑣𝑜𝑖 ; (1)
𝑣𝑜𝑖 − 𝑘𝑜𝑗 , otherwise .
⎩
The intuitive meaning of this formula is that if the user does not have any previous knowledge then the
amount of information he/she has still to read corresponds to the original content of the vocabulary
occurrence need i.e. 𝑉 𝑂𝑇 = 𝑉 𝑂𝑢𝑇 . On the contrary, if the user has already some previous knowledge
that is relevant to the information need, the number of occurrences he/she has still to read should
decrease with respect to 𝑉 𝑂𝑇 .
The Evaluation Framework We present the features that our cognitive evaluation frameworks
will take into account:
(I) Measuring the gain brought by one document to a user. Let 𝐺𝑎𝑖𝑛𝑉 𝑂 (𝑑) denote the learning gain
brought by a document 𝑑 to a user 𝑢 (user 𝑢 is represented here by his need 𝑉 𝑂). It is considered
to be the sum of gains brought by the keywords inside it 𝑔𝑎𝑖𝑛𝑣𝑜 (𝑑𝑘𝑖 ). Calculating the gain of every
document will allow determining the most “relevant” document or the one to be returned first. In the
ideal situation, the document should contain the exact number of occurrences needed, i.e. ∀𝑖, 𝑗 such that
𝑣𝑘𝑖 = 𝑑𝑘𝑗 , we should have 𝑣𝑜𝑖 = 𝑘𝑜𝑗 . In this case, indeed, the gain would be equal to its maximum
value. When the number of occurrences in the document is different (greater or lower) from the number
of occurrences needed, which is very probable, a gain function must be applied. The ongoing work is
studying the relevance of a document with respect to the user’s need. The proportion of 𝑑𝑘𝑖 and 𝑣𝑘𝑗
will be taken into account. The results will allow us to come up with the gain formula accounting for the
conditions above.
(II) Measuring the Gain brought by a document at rank r. Applying the gain function to each document
allows us to determine the one to be ranked first. For the second and for every subsequent document,
the gain measure should be reapplied with the user’s updated knowledge state. The documents scoring
higher will be ranked first. More precisely, before calculating the gain brought by a document, the 𝑉 𝑂𝑢
must be updated according to the information proposed in the document before it (at rank r-1). For
example, the knowledge state of a user after reading document 𝑎 at rank 𝑟 = 1, gets updated by adding
all the keywords of document 𝑎 and their occurrences. We can define the gain at rank 𝑟, 𝐺𝑎𝑖𝑛[𝑟], as the
gain provided by the document at rank 𝑟: 𝐺𝑎𝑖𝑛[𝑟] = 𝐺𝑎𝑖𝑛𝑉 𝑂 (𝑑).
(2)
(III) Calculating the cumulative
∑︀ and discounted Gains We calculate the cumulative gain vector at rank
𝑟, 𝐶𝐺[𝑟], as follows: 𝐶𝐺[𝑟] = 𝑟𝑗=1 𝐺𝑎𝑖𝑛[𝑗]
In order to take into consideration the rank of the documents, before computing the cumulative gain
vector, a discount may be applied at each rank to penalize documents lower in the ranking. We use the
classical discount: 𝑙𝑜𝑔2 (1 + 𝑟) to calculate the discounted cumulative gain at rank 𝑟. Finally, we normalize
the calculated DCG by the ideal discounted cumulative gain vector. 𝐷𝐶𝐺′ : 𝑛 − 𝐷𝐶𝐺 = 𝐷𝐶𝐺 𝐷𝐶𝐺
′.
We discussed in this paper an ongoing work of an evaluation framework for retrieval algorithms
that account for the user’s cognition, especially the user’s search goal and knowledge. The framework
penalises the redundancy of information not only with respect to the previously proposed documents,
but also to the user’s knowledge, and helps in the direction of achieving a search goal. In our framework,
the user’s search need is represented by a defined number of occurrences of keywords the user needs to
read. A document is considered relevant as long as it contains some needed keywords. After that point,
we penalize for redundancy. We also consider the “changing characteristics” of the cognitive aspects. We
can say that the to be proposed measure will be a first step toward evaluating the ranking and retrieval
algorithms with respect to the user’s cognition.
CIRCLE 2022 - Joint Conference of the Information Retrieval Communities in Europe, July 4-7, 2022, Samatan, France
*
Corresponding author.
$ elzein@i3s.unice.fr (D. El-Zein); Celia.DA-COSTA-PEREIRA@univ-cotedazur.fr (C. da-Costa-Pereira)
https://www.i3s.unice.fr/~elzein/ (D. El-Zein); https://www.i3s.unice.fr/~cpereira/ (C. da-Costa-Pereira)
0000-0003-4156-1237 (D. El-Zein); 0000-0001-6278-7740 (C. da-Costa-Pereira)
© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
Workshop
Proceedings
http://ceur-ws.org
ISSN 1613-0073
CEUR Workshop Proceedings (CEUR-WS.org)