<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>with User-Centric Agents</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Saber Zerhoudi</string-name>
          <email>saber.zerhoudi@uni-passau.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Granitzer</string-name>
          <email>michael.granitzer@uni-passau.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>User interactions, Retrieval-Augmented Generation (RAG), Personalized Information Retrieval, Multi-Agent RAG</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>In this study</institution>
          ,
          <addr-line>we present PersonaRAG, an innovative</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Passau</institution>
          ,
          <addr-line>Passau</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>diferent LLM architectures</institution>
          ,
          <addr-line>such as Llama 3 70b and Mix-</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>Large Language Models (LLMs) struggle with generating reliable outputs due to outdated knowledge and hallucinations. RetrievalAugmented Generation (RAG) models address this by enhancing LLMs with external knowledge, but often fail to personalize the retrieval process. This paper introduces PersonaRAG, a novel framework incorporating user-centric agents to adapt retrieval and generation based on real-time user data and interactions. Evaluated across various question answering datasets, PersonaRAG demonstrates superiority over baseline models, providing tailored answers to user needs. The results suggest promising directions for user-adapted information retrieval systems. Findings and resources are available at https://github.com/padas-lab-de/ir-rag-sigir24-persona-rag.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Large Language Models (LLMs) such as GPT-4 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and
LLaMA 3 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] have significantly advanced the field of natural
language processing (NLP) by demonstrating impressive
performance across various tasks and exhibiting emergent
abilities that push the boundaries of artificial intelligence [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
However, these models face challenges such as generating
unreliable outputs due to issues like hallucination and
outdated parametric memories [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Retrieval-Augmented Generation (RAG) models have
shown promise in addressing these issues by integrating
externally retrieved information to support more efective
performance on complex, knowledge-intensive tasks [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Despite these advancements, the deployment of RAG systems
within broader AI frameworks continues to face significant
challenges, particularly in handling noise and irrelevance
in retrieved data [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        A key limitation of existing RAG systems is their inability
to adapt outputs to users’ specific informational and
contextual needs. Personalized techniques in information retrieval,
such as adaptive retrieval based on user interaction data and
context-aware strategies, are increasingly recognized as
essential for enhancing user interaction and satisfaction [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ].
These methods aim to refine the retrieval process
dynamically, tailoring it more closely to individual user profiles
and situational contexts [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        The integration of agent-based systems with
personalized RAG architectures presents a compelling avenue for
research. Such systems utilize a multi-agent framework
to simulate complex, adaptive interactions tailored to
userspecific requirements [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. By embedding intelligent,
useroriented agents within the RAG framework, these systems
can evolve into more sophisticated tools that not only
retrieve relevant information but also align it closely with the
user’s specific preferences and contexts in real-time.
Importantly, the personalization strategy employed in these
systems is fully transparent to the user, ensuring that the
user is aware of how their information is being used to tailor
the results.
      </p>
      <p>Information Retrieval’s Role in RAG Systems (IR-RAG) workshop at SIGIR,
(M. Granitzer)</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related</title>
    </sec>
    <sec id="sec-3">
      <title>Work</title>
      <p>
        Retrieval-Augmented Generation (RAG) systems have
emerged as a significant advancement in natural language
processing and machine learning, enhancing language
models by integrating external knowledge bases to improve
performance across various tasks, such as question answering,
dialog understanding, and code generation [
        <xref ref-type="bibr" rid="ref13 ref6">6, 13</xref>
        ]. These
systems employ dense retrievers to pull relevant
information, which the language model then uses to generate
responses. However, the development of RAG systems and
their integration within broader artificial intelligence
frameworks is an ongoing area of research, with several challenges
and opportunities for improvement.
      </p>
      <p>Recent developments in RAG systems have focused on
reifning these models to better handle the noise and irrelevant
information often retrieved during the process. Xu et al.
CEUR</p>
      <p>
        ceur-ws.org
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] addressed this issue by employing natural language
inference models to select pertinent sentences, thereby
enhancing the RAG’s robustness. Additionally, advancements
have been made in adaptively retrieving information, with
systems like those proposed by Jiang et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] dynamically
fetching passages that are most likely to improve generation
accuracy.
      </p>
      <p>
        Despite these improvements, RAG systems still face
limitations, particularly in adapting their output to the user’s
specific profile, such as their information needs or
intellectual knowledge. This limitation stems from the current
design of most RAG systems, which do not typically
incorporate user context or personalized information retrieval
strategies [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Consequently, there exists a gap between
the general efectiveness of RAG systems and their
applicability in personalized user experiences, where context and
individual user preferences play a crucial role.
      </p>
      <p>
        Personalization in information retrieval is increasingly
recognized as essential for enhancing user interaction and
satisfaction [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Techniques such as user profiling,
contextaware retrieval, and adaptive feedback mechanisms are
commonly employed to tailor search results to individual users’
needs. For instance, Jeong et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] proposed adaptive
retrieval strategies that dynamically adjust the retrieval
process based on the complexity of the query and the user’s
historical interaction data. These personalized approaches
not only improve user satisfaction but also increase the
eficiency of information retrieval by reducing the time users
spend sifting through irrelevant information.
      </p>
      <p>
        The integration of personalized techniques with
agentbased systems provides a promising pathway to augment
the capabilities of RAG systems. Agent-based systems,
particularly in the form of LLM-Based Multi-Agent
Frameworks [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], enable the simulation of complex interactions
that can lead to more nuanced and contextually
appropriate outputs. By incorporating multi-agent systems into
RAG frameworks, there is potential for developing more
robust and adaptive retrieval mechanisms that can handle
a broader range of queries and generate more accurate
responses, closely tailored to the specific needs and contexts
of individual users.
      </p>
      <p>In conclusion, while significant progress has been made
in enhancing the efectiveness and personalization of RAG
systems, ongoing research is crucial to address their existing
limitations and expand their applications. The integration of
personalized information retrieval and agent-based
enhancements represents a promising avenue for further enhancing
the adaptability and accuracy of RAG systems, potentially
leading to intelligent information retrieval tailored to the
specific needs of users.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Methodology</title>
      <p>In this section, we present the methodology underlying our
PersonaRAG approach, which aims to enhance the ability
of Language Large Models (LLMs) to actively engage with,
understand, and leverage user profile information for
personalized content generation. We begin by discussing the
fundamental concepts of Retrieval-Augmented Generation
(RAG) models (Section 3.1) and then introduce our
PersonaRAG technique, which encourages LLMs to actively
assimilate knowledge from live search sessions (Section
3.2).</p>
      <sec id="sec-4-1">
        <title>3.1. Fundamentals of Retrieval-Augmented</title>
      </sec>
      <sec id="sec-4-2">
        <title>Generation (RAG) Models</title>
        <p>
          State-of-the-art RAG models, as described in previous
studies [
          <xref ref-type="bibr" rid="ref19 ref20 ref21">19, 20, 21</xref>
          ], employ retrieval systems to identify a set
of passages  = { 1, … ,   } when given a query q. These
passages are intended to enhance the generative
capabilities of LLMs by providing them with contextually relevant
information.
        </p>
        <p>
          Early versions of RAG models typically employ a
traditional retrieval-generation framework, in which the
retrieved data set  = { 1, … ,   } is directly fed into LLMs
to generate responses to the query  . However, these
passages often contain irrelevant information, and the direct
utilization approach in RAG has been shown to restrict the
potential benefits of the RAG framework [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. This
limitation has sparked further discussion on how to improve
LLMs by integrating retrieval results and outputs generated
by the models themselves [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>3.2. PersonaRAG: RAG with User-Centric</title>
      </sec>
      <sec id="sec-4-4">
        <title>Agents</title>
        <p>Drawing from the principles of adaptive learning and
usercentered design, we develop a new PersonaRAG architecture
to enable IR systems to dynamically learn from and adapt
to user behavior in real-time. As shown in Figure 2,
PersonaRAG introduces a three-step pipeline: retrieval, user
interaction analysis, and cognitive dynamic adaptation.
Unlike traditional IR models that statically respond to queries,
PersonaRAG focuses on leveraging live user data to
continually refine its understanding and responses without the
need for manual retraining.</p>
        <sec id="sec-4-4-1">
          <title>3.2.1. User Interaction Analysis</title>
          <p>
            To understand user behavior from live interactions,
PersonaRAG treats the IR system as a cognitive structure capable of
receiving, interpreting, and acting upon user feedback [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ].
Mimicking human learning behaviors, we establish four
distinct agents within the system dedicated to analyzing
user interactions from diferent perspectives: engagement
tracking, preference analysis, context understanding, and
feedback integration. These agents’ roles are detailed in
Section 3.2.2.
          </p>
        </sec>
        <sec id="sec-4-4-2">
          <title>3.2.2. Cognitive Dynamic Adaptation</title>
          <p>Following adaptive learning principles, we employ a
dynamic adaptation mechanism to assist the IR system in
utilizing real-time user data for continuous improvement. This
mechanism facilitates the integration of insights gained
from User Interaction Analysis into the system’s retrieval
processes. Specifically, we prompt the system to adjust its
query responses based on an initial understanding of the
user’s needs and refine these responses as more user data
becomes available. This approach not only personalizes the
search results but also helps in correcting any misalignments
or errors in real-time.</p>
          <p>
            PersonaRAG employs a highly specialized agent
architecture, with each agent focusing on a specific aspect of the
information retrieval process. All agents utilize in-context
learning, i.e., prompting, to perform their designated tasks.
This role specialization allows for the eficient
decomposition of complex user queries into manageable tasks [
            <xref ref-type="bibr" rid="ref25">25</xref>
            ].
To foster this, we engage the IR system as five specialized
agents to analyze user interactions based on retrieved data.
At present, the focus is on the functionality and interaction
of these agents rather than their individual performance
metrics.
          </p>
          <p>
            User Profile Agent This component manages and
updates user profile data, incorporating historical user
interactions and preferences [
            <xref ref-type="bibr" rid="ref26 ref27">26, 27</xref>
            ]. It monitors how users
interact with search results, such as click-through rates and
navigation paths. The User Profile Agent helps the system
understand what captures user interest and leads to deeper
engagement, enabling personalized search experiences.
Contextual Retrieval Agent This agent is responsible
for the initial retrieval of documents based on the user’s
current query. It accesses both a traditional search index
and a more dynamic context-aware system that can
consider broader aspects of the query environment. It utilizes
user profile data to modify and refine search queries or to
prioritize search results. For instance, if a user consistently
engages more with certain types of documents or topics,
the retrieval agent can boost those document types in the
search results, ensuring that the most relevant information
is presented to the user.
          </p>
          <p>Live Session Agent This agent analyzes the current
session in real-time, observing user actions such as clicks, time
spent on documents, modifications to the query, and any
feedback provided. It creates a session-specific context
model that captures the user’s immediate needs and
interests. The real-time data collected by this agent is used to
adjust the ongoing session, potentially re-ranking search
results or suggesting new queries based on the user’s
behavior and preferences. Additionally, the Live Session Agent
updates the user profile with new insights gleaned from
the session, allowing for a more personalized and eficient
search experience in future interactions.</p>
          <p>Document Ranking Agent This agent is responsible
for re-ranking the documents retrieved by the Contextual
Retrieval Agent. It integrates insights from both the User
Profile Agent and the Live Session Agent to score and order
the documents more efectively. By considering the user’s
historical preferences and their current session behavior,
the Document Ranking Agent ensures that the most
relevant and valuable documents are presented to the user in
a prioritized manner. This agent continuously adapts its
ranking algorithms based on the feedback received from the
user and the insights provided by the other agents in the
system.</p>
          <p>Feedback Agent This agent gathers implicit and explicit
feedback during and after user interactions. Implicit
feedback includes behavioral data like time spent on documents,
click counts, and navigation patterns. Explicit feedback
involves direct user input on document relevance and
quality, collected through ratings, surveys, or comments. The
agent uses this information to train and refine models for
other agents, particularly the Document Ranking Agent.
This process enhances the system’s ability to anticipate user
needs and deliver relevant documents based on accumulated
feedback and insights.</p>
          <p>By dynamically integrating insights from the User
Proifle Agent, Contextual Retrieval Agent, Live Session Agent,
Document Ranking Agent, and Feedback Agent into the IR
processes, PersonaRAG not only adapts to immediate user
needs but also evolves over time to better anticipate and
meet user expectations. This multi-agent approach enables
PersonaRAG to embody a truly adaptive and user-focused
information retrieval system, leveraging specialized agents
to analyze user interactions from diferent behavioral
perspectives and deliver highly personalized and contextually
relevant search experiences. The inclusion of the Document
Ranking Agent ensures that the most pertinent documents
are identified and presented to users, further enhancing the
system’s ability to efectively satisfy user information needs.</p>
        </sec>
      </sec>
      <sec id="sec-4-5">
        <title>3.3. PersonaRAG Operational Workflow</title>
        <p>
          The PersonaRAG framework employs a structured
worklfow that allows for sequential and parallel processing of
tasks, ensuring clarity and consistency in communication
between agents through well-defined data structures and
protocols [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. The process involves the User Profile Agent,
Contextual Retrieval Agent, Live Session Agent, Document
Ranking Agent, and Feedback Agent working together to
refine search queries, prioritize relevant results, and
improve document scoring and re-ranking based on user
proifle, session-specific contexts, and feedback.
        </p>
        <p>
          PersonaRAG’s modular design allows for flexibility in the
system setup, enabling researchers to focus on the most
relevant aspects of the user’s profile, session, and feedback data.
Agents work collaboratively by utilizing content from the
Global Message Pool, which serves as a central hub for
interagent communication [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], eliminating ineficiencies and
enabling agents to access or update information as required.
        </p>
        <p>
          The Feedback Agent collects and analyzes implicit and
explicit user feedback to generate insights into the
efectiveness of retrieval strategies and document relevance. This
feedback is used to make dynamic adjustments to the
system, refining retrieval methods and altering the weighting of
user profile factors. Through this iterative process,
PersonaRAG continuously adapts and improves its performance,
enhancing the accuracy and user satisfaction of the retrieval
results [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Experimental Setups</title>
      <p>In this section, we present the experimental setup employed
in our study, including the datasets, baseline models,
evaluation metrics, and implementation details. We also provide
an overview of the prompts used in our experiments.</p>
      <sec id="sec-5-1">
        <title>4.1. Datasets</title>
        <p>
          Our experiments are conducted on three widely used
singlehop benchmark datasets in the field of Information Retrieval
(IR): NaturalQuestions (NQ) [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ], TriviaQA [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ], and
WebQuestions (WebQ) [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ]. NQ is a well-known dataset in
Natural Language Understanding (NLU), consisting of
structured questions and corresponding Wikipedia pages
annotated with long and short answers. TriviaQA comprises
question-answer pairs collected from trivia and quiz-league
websites, while WebQ consists of questions selected using
the Google Suggest API, with answers being entities in
Freebase.
        </p>
        <p>Table 1 summarizes the datasets used in our initial study.
Due to the high cost of using language models and the large
number of API calls required, we randomly sampled 500
questions from each raw dataset to create more manageable
subsets for our experiments. While this sampling approach
limits the scope of our study, it allows us to conduct an
initial investigation into the performance of diferent RAG
systems on these datasets. We acknowledge that future
work with larger sample sizes and more comprehensive
experiments will be necessary to draw definitive conclusions.
Nonetheless, we believe this preliminary study provides
valuable insights into the relative strengths and weaknesses
of the tested RAG approaches.</p>
      </sec>
      <sec id="sec-5-2">
        <title>4.2. Models</title>
        <p>
          We compare PersonaRAG with several baseline models,
including prompt learning and RAG models. The prompt
templates used in user interaction analysis and dynamic
adaptation are presented in Section 4.4. Initially, the
questionanswering (QA) instruction is fed to ChatGPT to conduct
8,757
8,837
2,032
the vanilla answer generation model. Following the work of
Wei et al. [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ], the Chain-of-Thought model is implemented,
which generates question rationale results to produce the
ifnal results. Additionally, the Guideline model serves as
a baseline, generating problem-solving steps and guiding
Language Models (LLMs) to generate the answer.
        </p>
        <p>
          For the RAG-based baselines, two models are
implemented: vanilla RAG and Chain-of-Thought, which include
utilizing raw retrieved passages (CoT with Passage) and
refining the passages as notes (CoT with Note). The vanilla
RAG model directly feeds the top-ranked passages to the
LLM. The Chain-of-Note model [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] is also implemented,
which refines and summarizes the retrieved passages for
generation. Inspired by Self-RAG Asai et al. [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ], the
SelfRerank model is conducted, which filters out unrelated
contents without fine-tuning LLMs.
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>4.3. Evaluation Metrics</title>
        <p>When evaluating adaptive models, it is crucial to consider
both task performance and user-centric adaptability
simultaneously, along with their trade-ofs. Therefore, the results
are reported using diferent metrics, some of which measure
efectiveness and others measure eficiency.</p>
        <p>
          For efectiveness, accuracy is used, following the standard
evaluation protocol in the field of Information Retrieval
(IR) [
          <xref ref-type="bibr" rid="ref34 ref35 ref36">35, 36, 34</xref>
          ]. Accuracy assesses whether the predicted
answer contains the ground-truth answer. Both the outputs
of the Language Learning Model (LLM) and golden answers
are converted to lowercase, and string matching (StringEM)
is performed between each golden answer and the model
prediction to calculate accuracy.
        </p>
        <p>To evaluate user-centric adaptability, the BLEU-2 score is
measured to assess the text similarity between diferent RAG
and baseline setups and how well the generated answers
resemble each other. This metric provides insights into
the system’s ability to generate consistent and coherent
responses across various configurations. Additionally, the
average sentence length and the average number of syllables
of the answers from diferent RAG setups are reported as
a post-hoc analysis. These measures validate whether the
RAG system efectively adjusts its responses based on user
knowledge levels, ensuring that the generated answers are
tailored to the user’s understanding and expertise.</p>
        <p>Combining these evaluation strategies provides a
comprehensive view of both the efectiveness and user-centric
adaptability of the RAG system. The accuracy metric
ensures that the system generates correct answers, while the
BLEU-2 score and post-hoc analysis of sentence length and
syllable count confirm the system’s ability to adapt to user
knowledge levels. As the understanding of user needs and
system capabilities evolves, it is essential to continuously
refine these metrics to maintain the RAG system’s
efectiveness in delivering personalized, context-aware responses
that cater to the diverse requirements of users in the field</p>
      </sec>
      <sec id="sec-5-4">
        <title>4.4. Implementation Details</title>
        <p>
          For a fair comparison and following the work of Mallen et al.
[
          <xref ref-type="bibr" rid="ref35">35</xref>
          ] and Trivedi et al. [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ], the same retriever, a term-based
sparse retrieval model known as BM25 [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ], is used across all
diferent models. The retrieval model is implemented using
the OpenMatch toolkit [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ]. For the external document
corpus, the KILT-Wikipedia corpus preprocessed by Petroni
et al. [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ] is used, and the top-k relevant documents are
retrieved.
        </p>
        <p>
          Regarding the LLMs used to generate answers, the Llama
3 model instruct (ref) with 70b parameters, Mixture of
Experts (MoE) 8x7b (ref), and the GPT-3.5 model
(gpt-3.5turbo-0125) are employed. For the retrieval-augmented
LLM design, the implementation details from Trivedi et al.
[
          <xref ref-type="bibr" rid="ref37">37</xref>
          ] are followed, which include input prompts,
instructions, and the number of test samples for evaluation (e.g.,
500 samples per dataset).
        </p>
      </sec>
      <sec id="sec-5-5">
        <title>4.5. Prompts Used in PersonaRAG</title>
        <p>This subsection presents the prompt templates employed in
the construction of the PersonaRAG model. The prompts
utilized in the User Interaction Analysis and Cognitive
Dynamic Adaptation components are detailed below. The
prompt templates used by the baseline models are available
in the project repository 1. In the templates, {question}
represents the input question, {global_memory} the Global
Message Pool, while {passages} denotes the retrieved
passages. Additionally, {cot_answer} is populated with the
output generated by the Chain-of-Thought model.</p>
        <p>The placeholder {user_profile_answer} is filled
with the response produced by the User Profile agent
model. Respectively, {contextual_answer}
corresponds to the Contextual Retrieval agent model,
{live_session_answer} to the Live Session agent
model, {document_ranking_answer} to the Document
Ranking agent model, and {feedback_answer} to the
Feedback agent model.</p>
        <sec id="sec-5-5-1">
          <title>4.5.1. Prompts Used in User Interaction Analysis</title>
          <p>User Profile Agent</p>
          <p>Your task is to help the User Profile Agent
improve its understanding of user preferences
based on ranked document lists and the shared
global memory pool.
Contextual Retrieval Agent</p>
          <p>You are a search technology expert guiding the
Contextual Retrieval Agent to deliver
contextaware document retrieval.</p>
          <p>Live Session Agent</p>
          <p>Your expertise in session analysis is required
to assist the Live Session Agent in dynamically
adjusting results.</p>
          <p>Document Ranking Agent</p>
          <p>Your task is to help the Document Ranking Agent
prioritize documents for better ranking.
Feedback Agent</p>
          <p>You are an expert in feedback collection and
analysis, guiding the Feedback Agent to gather
and utilize user insights.</p>
          <p>Question: {question}
Passages: {passages}
Global Memory: {global_memory}
Task Description:
Using the retrieved passages and global memory
pool, identify methods for collecting implicit
and explicit user feedback. Suggest ways to
refine feedback mechanisms to align with user
preferences, such as ratings, surveys, or
behavioral data. Your recommendations should
guide the Feedback Agent in updating other
agents' models for more personalized and
relevant results.</p>
          <p>Global Message Pool</p>
          <p>You are responsible for maintaining and
enriching the Global Message Pool, serving
as a central hub for inter-agent communication.
Question: {question}
Agent Responses: {agent_responses}
Existing Global Memory: {global_memory}
Task Description:
Using the responses from individual agents
and the existing global memory, consolidate
key insights into a shared repository.</p>
          <p>Your goal is to organize a comprehensive
message pool that includes agent-specific
findings, historical user preferences,
sessionspecific behaviors, search queries, and user
feedback. This structure should provide
all agents with meaningful data points and
strategic recommendations, reducing redundant
communication and improving the system's overall
efficiency.</p>
        </sec>
        <sec id="sec-5-5-2">
          <title>4.5.2. Prompts Used in Cognitive Dynamic</title>
        </sec>
        <sec id="sec-5-5-3">
          <title>Adaptation</title>
          <p>Chain-of-Thought</p>
          <p>To solve the problem, Please think and reason
step by step, then answer.</p>
          <p>Question: {question}
Passages: {passages}
Reasoning process:
1. Read the given question and passages to
gather relevant information.
2. Write reading notes summarizing the key
points from these passages.
3. Discuss the relevance of the given question
and passages.
4. If some passages are relevant to the given
question, provide a brief answer based on the
passages.
5. If no passage is relevant, directly provide
the answer without considering the passages.</p>
          <p>Answer:
Cognitive Agent</p>
          <p>Your task is to help the Cognitive Agent
enhance its understanding of user insights
to continuously improve the system's responses.
Question: {question}
Initial Response: {cot_answer}</p>
        </sec>
        <sec id="sec-5-5-4">
          <title>Method</title>
        </sec>
        <sec id="sec-5-5-5">
          <title>Setting</title>
          <p>w/o RAG
vanillaRAG
Self-Refined
PersonaRAG
gpt-3.5-turbo-0125
Guideline
Chain-of-Thought (CoT)
Chain-of-Note (CoN)
Self-Rerank (SR)</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Experimental Results and</title>
    </sec>
    <sec id="sec-7">
      <title>Analyses</title>
      <p>In this section, we show the overall experimental results
and ofer in-depth analyses of our method.</p>
      <sec id="sec-7-1">
        <title>5.1. Main Results</title>
        <p>crucial role in eficiently extracting the necessary
information regarding the user’s information need to achieve these
improvements.</p>
        <p>Furthermore, on the WebQ dataset, PersonaRAG achieved
accuracy scores of 63.46% and 67.50% using Top-3 and Top-5
passages, respectively, surpassing the vanillaRAG model by
25% and 17.36%, and nearly all other baseline models
(except for Chain-of-Thought using Top-5, which performed
equally). On the NQ dataset, PersonaRAG maintained
similarly robust performance with scores of 49.02% and 48.78%,
outperforming all baselines (except for Chain-of-Thought
and Self-Rerank (SR) using Top-5). This pattern was
further validated by experiments on other datasets, with
results showing that PersonaRAG consistently outperforms
conventional RAG models with the capability of providing
an answer tailored to the user’s interaction and
information need. The comprehensive understanding it provides
contributes to the generation of accurate and user-centric
answers across various question complexities.</p>
      </sec>
      <sec id="sec-7-2">
        <title>5.2. Comparative Analysis of RAG</title>
      </sec>
      <sec id="sec-7-3">
        <title>Configurations</title>
        <p>Further experiments explored PersonaRAG’s adaptive
capabilities (Figure 3). BLEU-2 scores compared outputs
from Chain-of-Note (consistently best outside PersonaRAG)
with other methods. PersonaRAG showed higher
similarity scores, indicating its ability to generate responses that
address user needs rather than just summarizing input.
Additionally, PersonaRAG provides personalized answers
tailored to user profiles, extending beyond mere information
provision.</p>
        <p>The Chain-of-Note approach demonstrated comparable
performance to the Chain-of-Thought approach, implying
that both techniques efectively extract pertinent
information from the retrieved passages and adapt it to align with
the user’s information need.</p>
        <p>In contrast, vanillaGPT and vanillaRAG outputs difered
significantly from the Chain-of-Note approach, indicating
that counterfactual cognition often leads to diverse
outcomes rather than focusing solely on query-relevant
content. This suggests LLMs can construct knowledge from
multiple perspectives and customize responses based on
user understanding.</p>
        <p>Post-hoc analyses of average sentence length and syllable
count across RAG configurations provided insights into the
system’s ability to adapt responses to user comprehension
levels. These observations highlight PersonaRAG’s
capacity to synthesize knowledge from various perspectives and
tailor responses to diferent levels of user expertise.
(a) Text Similarity for Top-3 Passages</p>
      </sec>
      <sec id="sec-7-4">
        <title>5.3. Analysis on Generalization Ability</title>
        <p>This experiment evaluates the quality of knowledge
construction using diferent large language models (LLMs). As
illustrated in Table 3, the PersonaRAG outcomes are used
to prompt open-source LLMs, specifically LLaMA3-70B and
MoE-8x7b, to generate accurate answers.</p>
        <p>Compared to LLMs without retrieval-augmented
generation (w/o RAG), vanilla RAG and Chain-of-Note often
exhibit lower performance. This result suggests that retrieved
passages can act as noise, adversely afecting model
performance even after refinement through note generation. One
primary reason for this behavior is that both LLaMA3-70B
and MoE-8x7b struggle to eficiently analyze and identify
relevant knowledge due to limitations in their processing
capacities.</p>
        <p>In contrast, the PersonaRAG method provides notable
performance improvements: over 8% for LLaMA3-70B and
more than 10% for MoE-8x7b across all datasets,
underscoring its efectiveness. The PersonaRAG methodology
distinguishes itself from the Chain-of-Note approach by ofering a
cognitive framework that connects retrieved passages with
prior knowledge. This framework models the instructor’s
(GPT-3.5) reasoning process, guiding the student models
(LLaMA3-70B and MoE-8x7b) to better understand
knowledge retrieved from passages. The results demonstrate that
the LLMs are capable of selecting appropriate passages to
build more accurate responses, highlighting the benefits of
the PersonaRAG approach for improving generalization.
Finally, we randomly sample one case in Table to
demonstrate the efectiveness of PersonaRAG.</p>
        <p>The user interaction analysis mechanism efectively
generates comprehensive results by integrating foundational
and advanced insights from user data. Retrieved
passages provide critical clues for answering questions, while
agent analyses summarize and illustrate the applicability
of external information to user queries. The cognitive
dynamic adaptation module refines initial chain-of-thought
responses using these insights, generating accurate answers.
For example, including knowledge about the ”theft of the
Mona Lisa in 1911,” ”Vincenzo Peruggia,” and ”Florence”
enhances the reasoning process’s precision and detail. This
demonstrates PersonaRAG’s efectiveness in helping IR
agents combine external knowledge with intrinsic user data
to produce well-informed responses.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>6. Conclusion</title>
      <p>This paper proposes PersonaRAG, which constructs the
retrieval-augmentation architecture incorporating user
interaction analysis and cognitive dynamic adaptation.
PersonaRAG builds the user interaction agents and dynamic
cognitive mechanisms to facilitate the understanding of user
needs and interests and enhance the system capabilities to
deliver personalized, context-aware responses with the
intrinsic cognition of LLMs.</p>
      <p>Furthermore, PersonaRAG demonstrates efectiveness
in leveraging external knowledge and adapting responses
based on user profiles, knowledge levels, and information
needs to support LLMs in generation tasks without
finetuning. However, this approach requires multiple calls to the
LLM’s API, which can introduce additional time latency and
increase API calling costs when addressing questions. The
process involves constructing the initial Chain-of-Thought,
processing the User Interaction Agents results, and
executing the Cognitive Dynamic Adaptation to generate the final
answer. Furthermore, the inputs to LLMs in this approach
tend to be lengthy due to the inclusion of extensive retrieved
passages and the incorporation of user needs, interests, and
profile construction results. These factors can impact the
efifciency and cost-efectiveness of the PersonaRAG approach
in practical applications of Information Retrieval (IR)
systems.</p>
      <p>Future research will aim to optimize the process by
reducing API calls and developing concise representations of user
profiles and retrieved information without compromising
response quality. We also plan to explore more user-centric
agents to better capture writing styles and characteristics
of RAG users/searchers. This will enhance the system’s
ability to understand and adapt to individual preferences,
improving personalization and relevance in IR tasks.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>This work has received funding from the European Union’s
Horizon Europe research and innovation program under
grant agreement No 101070014 (OpenWebSearch.EU, https:
//doi.org/10.3030/101070014).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Pan</surname>
          </string-name>
          , K. Ma,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Chain-</surname>
          </string-name>
          of-note:
          <article-title>Enhancing robustness in retrievalaugmented language models</article-title>
          ,
          <source>CoRR abs/2311</source>
          .09210 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2311.09210. doi:
          <volume>10</volume>
          .48550/ARXIV.2311.09210.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] OpenAI, GPT-4
          <source>technical report, CoRR abs/2303</source>
          .08774 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv. 2303.08774. doi:
          <volume>10</volume>
          .48550/ARXIV.2303.08774. arXiv:
          <volume>2303</volume>
          .
          <fpage>08774</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lavril</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Izacard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Martinet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lachaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lacroix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Rozière</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Hambro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Azhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          , E. Grave, G. Lample,
          <article-title>Llama: Open and eficient foundation language models</article-title>
          ,
          <source>CoRR abs/2302</source>
          .13971 (
          <year>2023</year>
          ). URL: https://doi. org/10.48550/arXiv.2302.13971. doi:
          <volume>10</volume>
          .48550/ARXIV. 2302.13971. arXiv:
          <volume>2302</volume>
          .
          <fpage>13971</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            , E. Sigler,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Litwin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Chess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Berner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McCandlish</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language models are few-shot learners</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems</source>
          <year>2020</year>
          ,
          <article-title>NeurIPS 2020</article-title>
          , December 6-
          <issue>12</issue>
          ,
          <year>2020</year>
          , virtual,
          <year>2020</year>
          . URL: https://proceedings.neurips.cc/paper/2020/hash/ 1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cahyawijaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wilie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lovenia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Do</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fung</surname>
          </string-name>
          ,
          <article-title>A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity</article-title>
          , in: J. C. Park,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Arase</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wijaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Purwarianti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Krisnadhi</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd</source>
          <article-title>Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics</article-title>
          ,
          <source>IJCNLP 2023 -Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          , Nusa Dua, Bali, November 1 -
          <issue>4</issue>
          ,
          <year>2023</year>
          , Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>675</fpage>
          -
          <lpage>718</lpage>
          . URL: https://doi.org/10.18653/v1/
          <year>2023</year>
          .ijcnlp-main.
          <volume>45</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2023</year>
          .IJCNLP-MAIN.
          <year>45</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P. S. H.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Küttler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <article-title>Retrievalaugmented generation for knowledge-intensive NLP tasks</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems</source>
          <year>2020</year>
          ,
          <article-title>NeurIPS 2020</article-title>
          , December 6-
          <issue>12</issue>
          ,
          <year>2020</year>
          , virtual,
          <year>2020</year>
          . URL: https://proceedings.neurips.cc/paper/2020/hash/ 6b493230205f780e1bc26945df7481e5-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Sun,
          <article-title>Benchmarking large language models in retrieval-augmented generation</article-title>
          , in: M. J.
          <string-name>
            <surname>Wooldridge</surname>
            ,
            <given-names>J. G.</given-names>
          </string-name>
          <string-name>
            <surname>Dy</surname>
          </string-name>
          , S. Natarajan (Eds.),
          <source>Thirty-Eighth AAAI Conference on Artificial Intelligence</source>
          ,
          <source>AAAI 2024, Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, February 20-27</source>
          ,
          <year>2024</year>
          , Vancouver, Canada, AAAI Press,
          <year>2024</year>
          , pp.
          <fpage>17754</fpage>
          -
          <lpage>17762</lpage>
          . URL: https://doi.org/10.1609/aaai.v38i16. 29728. doi:
          <volume>10</volume>
          .1609/AAAI.V38I16.29728.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Teevan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Dumais</surname>
          </string-name>
          , E. Horvitz,
          <article-title>Personalizing search via automated analysis of interests and activities</article-title>
          ,
          <source>SIGIR Forum 51</source>
          (
          <year>2017</year>
          )
          <fpage>10</fpage>
          -
          <lpage>17</lpage>
          . URL: https://doi. org/10.1145/3190580.3190582. doi:
          <volume>10</volume>
          .1145/3190580. 3190582.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sugiyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hatano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yoshikawa</surname>
          </string-name>
          ,
          <article-title>Adaptive web search based on user profile constructed without any efort from users</article-title>
          , in: S. I. Feldman,
          <string-name>
            <given-names>M.</given-names>
            <surname>Uretsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Najork</surname>
          </string-name>
          , C. E. Wills (Eds.),
          <source>Proceedings of the 13th international conference on World Wide Web, WWW</source>
          <year>2004</year>
          , New York, NY, USA, May
          <volume>17</volume>
          -20,
          <year>2004</year>
          , ACM,
          <year>2004</year>
          , pp.
          <fpage>675</fpage>
          -
          <lpage>684</lpage>
          . URL: https://doi.org/10.1145/988672. 988764. doi:
          <volume>10</volume>
          .1145/988672.988764.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Adomavicius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mobasher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ricci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tuzhilin</surname>
          </string-name>
          ,
          <article-title>Context-aware recommender systems</article-title>
          ,
          <source>AI Mag</source>
          .
          <volume>32</volume>
          (
          <year>2011</year>
          )
          <fpage>67</fpage>
          -
          <lpage>80</lpage>
          . URL: https://doi.org/10.1609/aimag.v32i3. 2364. doi:
          <volume>10</volume>
          .1609/AIMAG.V32I3.2364.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>M. J. Wooldridge</surname>
          </string-name>
          , An Introduction to MultiAgent Systems, Second Edition, Wiley,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A. Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sablayrolles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Savary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bamford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Chaplot</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. de Las Casas</surname>
            ,
            <given-names>E. B.</given-names>
          </string-name>
          <string-name>
            <surname>Hanna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Bressand</surname>
            ,
            <given-names>G.</given-names>
            Lengyel, G. Bour, G.
          </string-name>
          <string-name>
            <surname>Lample</surname>
            ,
            <given-names>L. R.</given-names>
          </string-name>
          <string-name>
            <surname>Lavaud</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Saulnier</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Lachaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Stock</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Antoniak</surname>
            ,
            <given-names>T. L.</given-names>
          </string-name>
          <string-name>
            <surname>Scao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Gervet</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>W. E.</given-names>
          </string-name>
          <string-name>
            <surname>Sayed</surname>
          </string-name>
          , Mixtral of experts,
          <source>CoRR abs/2401</source>
          .04088 (
          <year>2024</year>
          ). URL: https://doi.org/10.48550/arXiv.2401.04088. doi:
          <volume>10</volume>
          .48550/ARXIV.2401.04088.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Choi,</surname>
          </string-name>
          <article-title>RECOMP: improving retrievalaugmented lms with compression and selective augmentation</article-title>
          ,
          <source>CoRR abs/2310</source>
          .04408 (
          <year>2023</year>
          ). URL: https: //doi.org/10.48550/arXiv.2310.04408. doi:
          <volume>10</volume>
          .48550/ ARXIV.2310.04408.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. F.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Liu</surname>
          </string-name>
          , J. DwivediYu,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Callan</surname>
          </string-name>
          , G. Neubig,
          <article-title>Active retrieval augmented generation</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP</source>
          <year>2023</year>
          , Singapore, December 6-
          <issue>10</issue>
          ,
          <year>2023</year>
          , Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>7969</fpage>
          -
          <lpage>7992</lpage>
          . URL: https://doi.org/ 10.18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>495</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2023</year>
          .EMNLP-MAIN.
          <year>495</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Embedding-based query language models</article-title>
          , in: B.
          <string-name>
            <surname>Carterette</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Fang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Lalmas</surname>
          </string-name>
          , J. Nie (Eds.),
          <source>Proceedings of the 2016 ACM on International Conference on the Theory of Information Retrieval</source>
          ,
          <string-name>
            <surname>ICTIR</surname>
          </string-name>
          <year>2016</year>
          , Newark, DE, USA,
          <source>September 12- 6</source>
          ,
          <year>2016</year>
          , ACM,
          <year>2016</year>
          , pp.
          <fpage>147</fpage>
          -
          <lpage>156</lpage>
          . URL: https://doi. org/10.1145/2970398.2970405. doi:
          <volume>10</volume>
          .1145/2970398. 2970405.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Ghorab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. O'Connor</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Wade</surname>
          </string-name>
          ,
          <article-title>Personalised information retrieval: survey and classification, User Model</article-title>
          .
          <source>User Adapt. Interact</source>
          .
          <volume>23</volume>
          (
          <year>2013</year>
          )
          <fpage>381</fpage>
          -
          <lpage>443</lpage>
          . URL: https://doi.org/10.1007/s11257-012-9124-1. doi:
          <volume>10</volume>
          .1007/S11257-012-9124-1.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jeong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Park</surname>
          </string-name>
          , Adaptive-rag:
          <article-title>Learning to adapt retrieval-augmented large language models through question complexity</article-title>
          ,
          <source>CoRR abs/2403</source>
          .14403 (
          <year>2024</year>
          ). URL: https://doi. org/10.48550/arXiv.2403.14403. doi:
          <volume>10</volume>
          .48550/ARXIV. 2403.14403.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Sun, Metaagents:
          <article-title>Simulating interactions of human behaviors for llm-based taskoriented coordination via collaborative generative agents</article-title>
          ,
          <source>CoRR abs/2310</source>
          .06500 (
          <year>2023</year>
          ). URL: https://doi. org/10.48550/arXiv.2310.06500. doi:
          <volume>10</volume>
          .48550/ARXIV. 2310.06500.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Retrievalaugmented generation for large language models: A survey</article-title>
          ,
          <source>CoRR abs/2312</source>
          .10997 (
          <year>2023</year>
          ). URL: https://doi. org/10.48550/arXiv.2312.10997. doi:
          <volume>10</volume>
          .48550/ARXIV. 2312.10997.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <article-title>A survey on retrieval-augmented text generation for large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2404.10981</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Siriwardhana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weerasekera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kaluarachchi</surname>
          </string-name>
          , E. Wen,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nanayakkara</surname>
          </string-name>
          ,
          <article-title>Improving the domain adaptation of retrieval augmented generation (RAG) models for open domain question answering</article-title>
          ,
          <source>Trans. Assoc. Comput. Linguistics</source>
          <volume>11</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          . URL: https://transacl.org/ojs/index.php/ tacl/article/view/4029.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Sun,
          <article-title>Benchmarking large language models in retrieval-augmented generation</article-title>
          , in: M. J.
          <string-name>
            <surname>Wooldridge</surname>
            ,
            <given-names>J. G.</given-names>
          </string-name>
          <string-name>
            <surname>Dy</surname>
          </string-name>
          , S. Natarajan (Eds.),
          <source>Thirty-Eighth AAAI Conference on Artificial Intelligence</source>
          ,
          <source>AAAI 2024, Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, February 20-27</source>
          ,
          <year>2024</year>
          , Vancouver, Canada, AAAI Press,
          <year>2024</year>
          , pp.
          <fpage>17754</fpage>
          -
          <lpage>17762</lpage>
          . URL: https://doi.org/10.1609/aaai.v38i16. 29728. doi:
          <volume>10</volume>
          .1609/AAAI.V38I16.29728.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>K.</given-names>
            <surname>Wu</surname>
          </string-name>
          , E. Wu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zou</surname>
          </string-name>
          ,
          <article-title>How faithful are rag models? quantifying the tug-of-war between rag and llms' internal prior</article-title>
          ,
          <source>arXiv preprint arXiv:2404.10198</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Atkinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Shifrin</surname>
          </string-name>
          ,
          <article-title>Human memory: A proposed system and its control processes</article-title>
          , in: K. W. Spence, J. T. Spence (Eds.),
          <source>Psychology of Learning and Motivation</source>
          , volume
          <volume>2</volume>
          of Psychology of Learning and Motivation, Elsevier,
          <year>1968</year>
          , pp.
          <fpage>89</fpage>
          -
          <lpage>195</lpage>
          . URL: https:// doi.org/10.1016/s0079-
          <volume>7421</volume>
          (
          <issue>08</issue>
          )
          <fpage>60422</fpage>
          -
          <lpage>3</lpage>
          . doi:
          <volume>10</volume>
          .1016/ S0079-
          <volume>7421</volume>
          (
          <issue>08</issue>
          )
          <fpage>60422</fpage>
          -
          <lpage>3</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Semantic web-based information retrieval models: a systematic survey</article-title>
          ,
          <source>in: Data Science and Analytics: 5th International Conference on Recent Developments in Science</source>
          , Engineering and Technology,
          <string-name>
            <surname>REDSET</surname>
          </string-name>
          <year>2019</year>
          , Gurugram, India,
          <source>November 15-16</source>
          ,
          <year>2019</year>
          ,
          <string-name>
            <given-names>Revised</given-names>
            <surname>Selected</surname>
          </string-name>
          <string-name>
            <surname>Papers</surname>
          </string-name>
          ,
          <source>Part II 5</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>204</fpage>
          -
          <lpage>222</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kacem</surname>
          </string-name>
          ,
          <article-title>Personalized Information Retrieval based on Time-Sensitive User Profile. (Recherche d'Information Personalisée basée sur un Profil Utilisateur Sensible au Temps)</article-title>
          ,
          <source>Ph.D. thesis</source>
          , Paul Sabatier University, Toulouse, France,
          <year>2017</year>
          . URL: https://tel. archives-ouvertes.fr/tel-01707423.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>A multi-agent framework for context-aware dynamic user profiling for web personalization</article-title>
          ,
          <source>in: Software Engineering: Proceedings of CSI 2015</source>
          , Springer,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          , Y. Cheng, J.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>S. K. S.</given-names>
          </string-name>
          <string-name>
            <surname>Yau</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Ran</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
          </string-name>
          , Metagpt:
          <article-title>Meta programming for multiagent collaborative framework</article-title>
          ,
          <source>CoRR abs/2308</source>
          .00352 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2308.00352. doi:
          <volume>10</volume>
          .48550/ARXIV.2308.00352.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Limbu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Connor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pears</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. G.</surname>
          </string-name>
          <article-title>MacDonell, Contextual relevance feedback in web information retrieval</article-title>
          , in: I. Ruthven (Ed.),
          <source>Proceedings of the 1st International Conference on Information Interaction in Context, IIiX</source>
          <year>2006</year>
          , Copenhagen, Denmark,
          <source>October 18-20</source>
          ,
          <year>2006</year>
          , ACM,
          <year>2006</year>
          , pp.
          <fpage>138</fpage>
          -
          <lpage>143</lpage>
          . URL: https://doi. org/10.1145/1164820.1164848. doi:
          <volume>10</volume>
          .1145/1164820. 1164848.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kwiatkowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Palomaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Redfield</surname>
          </string-name>
          , M. Collins,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>Parikh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Alberti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Epstein</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kelcey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Petrov</surname>
          </string-name>
          ,
          <article-title>Natural questions: a benchmark for question answering research</article-title>
          ,
          <source>Trans. Assoc. Comput. Linguistics</source>
          <volume>7</volume>
          (
          <year>2019</year>
          )
          <fpage>452</fpage>
          -
          <lpage>466</lpage>
          . URL: https://doi.org/10.1162/tacl_a_00276. doi:
          <volume>10</volume>
          .1162/TACL\_A\_
          <volume>00276</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Weld</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <article-title>Triviaqa: A large scale distantly supervised challenge dataset for reading comprehension</article-title>
          , in: R. Barzilay, M. Kan (Eds.),
          <article-title>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</article-title>
          ,
          <string-name>
            <surname>ACL</surname>
          </string-name>
          <year>2017</year>
          , Vancouver, Canada,
          <source>July 30 - August 4</source>
          , Volume
          <volume>1</volume>
          :
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1601</fpage>
          -
          <lpage>1611</lpage>
          . URL: https://doi.org/10. 18653/v1/
          <fpage>P17</fpage>
          -1147. doi:
          <volume>10</volume>
          .18653/V1/P17-1147.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>Berant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Frostig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Semantic parsing on freebase from question-answer pairs</article-title>
          ,
          <source>in: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, EMNLP</source>
          <year>2013</year>
          ,
          <volume>18</volume>
          -21
          <source>October</source>
          <year>2013</year>
          , Grand Hyatt Seattle, Seattle, Washington, USA,
          <article-title>A meeting of SIGDAT, a Special Interest Group of the ACL</article-title>
          , ACL,
          <year>2013</year>
          , pp.
          <fpage>1533</fpage>
          -
          <lpage>1544</lpage>
          . URL: https://aclanthology.org/D13-1160/.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Schuurmans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ichter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <article-title>Chain-of-thought prompting elicits reasoning in large language models</article-title>
          , in: S. Koyejo,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Belgrave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Oh (Eds.),
          <source>Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems</source>
          <year>2022</year>
          , NeurIPS
          <year>2022</year>
          , New Orleans, LA, USA, November 28 - December 9,
          <year>2022</year>
          ,
          <year>2022</year>
          . URL: http://papers.nips.cc/paper_files/paper/2022/hash/ 9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference. html.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>A.</given-names>
            <surname>Asai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hajishirzi</surname>
          </string-name>
          ,
          <article-title>Self-rag: Learning to retrieve, generate, and critique through self-reflection</article-title>
          ,
          <source>CoRR abs/2310</source>
          .11511 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2310.11511. doi:
          <volume>10</volume>
          .48550/ARXIV.2310.11511.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mallen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Asai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. Das</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Khashabi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Hajishirzi</surname>
          </string-name>
          ,
          <article-title>When not to trust language models: Investigating efectiveness of parametric and nonparametric memories</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>ACL</source>
          <year>2023</year>
          , Toronto, Canada, July 9-
          <issue>14</issue>
          ,
          <year>2023</year>
          , Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>9802</fpage>
          -
          <lpage>9822</lpage>
          . URL: https:// doi.org/10.18653/v1/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>546</volume>
          . doi:
          <volume>10</volume>
          .18653/ V1/
          <year>2023</year>
          .
          <article-title>ACL-LONG</article-title>
          .
          <year>546</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>J.</given-names>
            <surname>Baek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jeong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Hwang</surname>
          </string-name>
          ,
          <article-title>Knowledge-augmented language model verification</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP</source>
          <year>2023</year>
          , Singapore, December 6-
          <issue>10</issue>
          ,
          <year>2023</year>
          , Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>1720</fpage>
          -
          <lpage>1736</lpage>
          . URL: https://doi.org/ 10.18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>107</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2023</year>
          .EMNLP-MAIN.
          <year>107</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>H.</given-names>
            <surname>Trivedi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Balasubramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Khot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sabharwal</surname>
          </string-name>
          ,
          <article-title>Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>ACL</source>
          <year>2023</year>
          , Toronto, Canada, July 9-
          <issue>14</issue>
          ,
          <year>2023</year>
          , Association for Computational Linguistics,
          <year>2023</year>
          , pp.
          <fpage>10014</fpage>
          -
          <lpage>10037</lpage>
          . URL: https://doi.org/10.18653/v1/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>557</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2023</year>
          .
          <article-title>ACL-LONG</article-title>
          .
          <year>557</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Walker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jones</surname>
          </string-name>
          , M. HancockBeaulieu, M. Gatford, Okapi at TREC-3, in: D. K. Harman (Ed.),
          <source>Proceedings of The Third Text REtrieval Conference</source>
          , TREC 1994, Gaithersburg, Maryland, USA, November 2-
          <issue>4</issue>
          ,
          <year>1994</year>
          , volume
          <volume>500</volume>
          -225 of NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>1994</year>
          , pp.
          <fpage>109</fpage>
          -
          <lpage>126</lpage>
          . URL: http://trec.nist.gov/pubs/trec3/papers/city.ps.gz.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          , Openmatch-v2:
          <article-title>An all-in-one multi-modality plm-based information retrieval toolkit</article-title>
          , in: H.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>W. E.</given-names>
          </string-name>
          <string-name>
            <surname>Duh</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>M. P.</given-names>
          </string-name>
          <string-name>
            <surname>Kato</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Poblete (Eds.),
          <source>Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>SIGIR</surname>
          </string-name>
          <year>2023</year>
          , Taipei, Taiwan,
          <source>July 23-27</source>
          ,
          <year>2023</year>
          , ACM,
          <year>2023</year>
          , pp.
          <fpage>3160</fpage>
          -
          <lpage>3164</lpage>
          . URL: https://doi.org/10.1145/3539618. 3591813. doi:
          <volume>10</volume>
          .1145/3539618.3591813.
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S. H.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yazdani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Thorne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Maillard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Plachouras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          , S. Riedel,
          <article-title>KILT: a benchmark for knowledge intensive language tasks</article-title>
          , in: K.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Hakkani-Tür</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Beltagy</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cotterell</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          , Y. Zhou (Eds.),
          <source>Proceedings of the</source>
          <year>2021</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online</article-title>
          , June 6-11,
          <year>2021</year>
          , Association for Computational Linguistics,
          <year>2021</year>
          , pp.
          <fpage>2523</fpage>
          -
          <lpage>2544</lpage>
          . URL: https://doi.org/ 10.18653/v1/
          <year>2021</year>
          .naacl-main.
          <volume>200</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2021</year>
          .NAACL-MAIN.
          <year>200</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>