<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>N. Othman, R. Faiz, K. Smaïli, Enhancing question retrieval in community question answer-
ing using word embeddings, Procedia Computer Science</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1108/IJCS-03-2019-0011</article-id>
      <title-group>
        <article-title>Answering⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vraj Patel</string-name>
          <email>P@1</email>
          <email>vrajp2108@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Palak Vanpariya</string-name>
          <email>palakvanpariya13@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kandarp Gajjar</string-name>
          <email>kandarp.gajjar.460@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Himani Trivedi</string-name>
          <email>himani_ce@ldrp.ac.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Personalized Information Retrieval, Community Question Answering, SE-PQA Dataset, Personalized Ranking</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>LDRP Institute of Technology &amp; Research</institution>
          ,
          <addr-line>Kadi Sarva Vishwavidyalaya, Gandhinagar, Gujarat</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>159</volume>
      <issue>2019</issue>
      <fpage>485</fpage>
      <lpage>494</lpage>
      <abstract>
        <p>In the current informational age, personalized information retrieval (PIR) has proved to be useful in addressing the problem of information overload. We introduce a new Personal Information Retrieval architecture utilizing the SE-PQA (Stack Exchange - Personalized Question Answering) dataset and a community question answering task model. Our approach leverages rich user relationship level social features and social interactivity data contained in the SE-PQA, which spans over 1 million questions and 2 million answers across various Stack Exchange communities. The proposed model integrates three critical components: initial BM25 retrieval, MiniLM semantic reranking, and user-specific ranking through the TAG model, combining the strengths of traditional information retrieval, eficient language models, and personalized ranking. Extensive experimentation with both the Base SE-PQA and Personalized SE-PQA datasets demonstrates the eficacy of this methodology, with significant improvements in performance metrics. On the Personalized SE-PQA dataset, which incorporates user-selected of 0.465, and MAP@100 of 0.428. These results suggest that incorporating both traditional and neural approaches, along with user-specific features, can contribute to more efective personalized Community Question Answering (CQA) systems, while demonstrating the potential of SE-PQA data in developing and evaluating Personalized Information Retrieval (PIR) frameworks.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The exponential surge in online information necessitates increasingly personalized retrieval
mechanisms to efectively cater to users’ diverse needs. Community Question Answering (CQA) platforms,
exemplified by StackExchange, face unique challenges in balancing the specificity of user queries
with the collective expertise within their specialized domains. Enhancing personalization in these
environments holds the potential to significantly elevate user experience by delivering search results
that align more closely with individual preferences, expertise levels, and interaction histories.</p>
      <p>
        The recent advancement in PIR has been restricted, since there has not been suficient availability of
larger, real-world datasets that reflect the richness of user behavior and content relevance. The release
of the SE-PQA dataset StackExchange - Personalized Question Answering [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] has provided researchers
with a source to develop and test PIR models against the context of CQA. This dataset has over 1 million
questions and 2 million answers from diferent StackExchange communities with rich user-level features
and social interaction data, thus making it amenable to building complex personalization techniques.
      </p>
      <p>Traditional approaches to PIR in the context of CQA platforms have primarily relied on basic
user profiling techniques or rudimentary forms of collaborative filtering. While these methodologies
demonstrate some efectiveness, they are inherently limited in their capacity to capture the complex,
multi-faceted nature of user preferences and the dynamic social interactions characteristic of CQA
environments. Furthermore, the application of advanced language models coupled with sophisticated
(H. Trivedi)
∗Corresponding author.
†These authors contributed equally.</p>
      <p>CEUR</p>
      <p>ceur-ws.org
personalization techniques remains an under-explored avenue of research within the CQA domain,
presenting significant opportunities for enhancing the accuracy and relevance of information retrieval
in these platforms.</p>
      <p>The rest of this paper is structured as follows: Section 2 provides a comprehensive review of related
work in PIR and CQA, highlighting current limitations and research gaps. Section 3 delineates our
implementation, detailing the integration of diverse personalization components. We present and
discuss our results and analysis in Section 4, including comparisons with baseline methods and ablation
studies. Section 5 concludes the paper by succinctly summarizing our key findings and proposing
potential avenues for future research in this rapidly evolving field. Finally, Section 6 encompasses the
references cited throughout this work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Research in Community Question Answering (CQA) began with retrieval challenges. Othman et
al. [2] addressed the word mismatch problem using word embeddings and k-means clustering to
achieve semantic similarity in question retrieval. However, issues like out-of-vocabulary (OOV) words,
high computational overhead, and biases in embeddings limited its scalability to larger datasets and
multilingual contexts.</p>
      <p>Yang et al. [3] explored expert recommendation systems, proposing a framework to classify
recommendation methods. While this eased limitations in earlier retrieval systems, it introduced new
challenges such as sparse data and inconclusive results, highlighting the need for advanced
personalization techniques and accurate profiling mechanisms.</p>
      <p>Building on these gaps, Zhang et al. [4] proposed a personalized chatbot model that learns implicit
user profiles from dialogue histories, mitigating sparse data issues. However, it introduced noise-related
challenges in extended dialogues, emphasizing the trade-of between performance and eficiency.</p>
      <p>
        Further advancements were made in ”SE-PQA: Personalized Community Question Answering” [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
which introduced multi-domain user interaction features to enhance personalization on large-scale
datasets. Although promising, the study relied on a simple user model and called for more extensive
experimentation to unlock its full potential across diverse domains.
      </p>
      <p>Recently, the use of Large Language Models (LLMs) has been explored as a transformative step
in CQA. A study titled ”Large Language Models and Future of Information Retrieval” [5] addressed
persistent challenges like OOVs, scalability, and intent understanding. Despite their capabilities, LLMs
bring concerns about bias and ethical considerations, necessitating careful deployment in real-world
applications.</p>
      <p>Ongoing research in Personalized Information Retrieval (PIR) continues to push the boundaries of
the field. The ’Overview of the PIR Track at FIRE 2024’ [ 6] provides a comprehensive evaluation of
state-of-the-art personalized retrieval techniques, while highlighting the collaborative eforts that have
driven advancements in balancing precision, efectiveness, and user relevance. These developments,
discussed in the proceedings of FIRE 2024 [7], underscore the growing importance of refining retrieval
systems to meet the demands of dynamic and evolving online environments.</p>
      <p>These studies collectively illustrate the evolution of CQA systems—from word-embedding techniques
to sophisticated LLM-powered solutions—reflecting continuous progress toward enhancing scalability,
personalization, and user-centricity.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Model</title>
      <p>In our approach for personalized information retrieval, we make use of three main models: BM25 as
the primary model for retrieval, MiniLM for reranking, and the TAG model for personalized ranking.
BM25 can be regarded to be the retrieval model at the first step of ranking potential answers to the
queries raised by the users based on their relevance. It considers query term frequencies in documents
and the overall importance of the documents in the entire dataset so that the most relevant answers are
surfaced first.</p>
      <p>Next, we employ MiniLM as a neural re-ranker. This model tunes the initial output of BM25 in terms
of the context and semantics of a question and its answer. Deep analysis of language patterns and
meanings residing in them with the help of MiniLM ensures that answers provided to users are not
only relevant but also contextual.</p>
      <p>Lastly, we add the TAG model to refine the rank. This assigns scores to the right answer according
to the user’s past behaviors and interests, which are depicted through tags. At the moment of query
submission, the TAG model calculates the tags associated with the user’s past questions and matches
them against the tags associated with the answers. If that result is coming from a user who has the
same interest as us, it gains more points, which in turn enhances the personalization aspect of our
retrieval system.</p>
      <p>Together, the three models-BM25, MiniLM, and TAG form a more robust framework that improve
on relevance as well as personalization of answers provided. That actually addresses the problem of
information overload in community question-answering platforms and allows users to get answers
tailored uniquely to their preferences and context.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <p>We performed experiments on the SE-PQA dataset to evaluate our proposed model. It combines three
components: BM25 for initial retrieval, MiniLM for semantic re-ranking, and the TAG model for
personalized ranking. The model aims to combine the strengths of traditional information retrieval,
eficient language models, and user-specific personalization.</p>
      <p>We evaluated our model on two variants of the SE-PQA dataset: the Base SE-PQA and the Personalized
SE-PQA. Table 1 presents the results for the cQA task on the Base SE-PQA dataset.</p>
      <sec id="sec-4-1">
        <title>Recall@100 MAP@100 0.325 0.184</title>
        <p>0.110
0.349
0.361
0.378
0.359
0.237
0.160
0.383
0.398
0.414
0.615
0.615
0.615
0.615
0.615
0.615
0.320
0.199
0.131
0.342
0.355
0.370</p>
        <p>The results in Table 1 show that the combined method, BM25 + MiniLM + TAG, outperforms individual
models and simpler combinations on all metrics. It achieves significant improvements of 15.8% for
P@1 and 15.6% for MAP@100 compared to the BM25 baseline. MiniLM proves efective at increasing
semantic understanding for both queries and answers, serving as a considerably lighter alternative to
larger language models.</p>
        <p>To further validate our approach, we evaluated the models on the Personalized SE-PQA dataset,
which incorporates richer user-level features. Table 2 presents these results.</p>
        <p>As shown in Table 2, the results on the Personalized SE-PQA dataset are even more pronounced. Our
integrated model achieves a P@1 score of 0.339, an NDCG@10 score of 0.465, and a MAP@100 score of
0.428. These findings highlight the importance of personalization in CQA systems and demonstrate the
efectiveness of the proposed integrated approach, which leverages the semantic understanding and
eficiency of MiniLM.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Recall@100 MAP@100 0.353 0.209</title>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>Our work on personalized information retrieval for Community Question Answering systems using the
SE-PQA dataset gives good results. In fact, we notice that the combination of BM25, MiniLM, and the
TAG model performs better than all individual approaches, as well as any simpler combination. This
could be seen particularly in the Personalized SE-PQA dataset, where our best model BM25 + MiniLM +
TAG achieved the highest overall scores: P@1=0.339, NDCG@10=0.465, and MAP@100=0.428. The
potential of this proposed combination lies in its ability to exploit the advantages that are rooted within
each of the diferent areas: the benefits of the traditional information retrieval algorithms (BM25),
eficient language modeling (MiniLM), and user-specific personalization (TAG). The addition of MiniLM
proved particularly useful, as it yielded very strong semantic understanding without the computational
burdens that often accompany heavier language models.</p>
      <p>While these results are promising, there is still quite a lot of room to improve and explore in this
area. Future work may further push the personalization aspects toward even higher levels, perhaps
introducing more dynamic user behavior modeling or even advanced few-shot learning techniques.
The investigation can also be conducted with even more eficient language models or making the
existing approach optimized for real-time applications. As these CQA sites are developed, so must be
our perceptions of retrieval and our methods of information retrieval, continually seeking to balance
precision, efectiveness, and user relevance in ever more complex and diverse online communities.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT-4 for text polishing (Activity: Editing).
Additionally, Quillbot was used for grammar correction (Activity: Checking). After using these tools,
the author(s) reviewed and edited the content as needed and take full responsibility for the publication’s
content.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We express our sincere gratitude to the Department of Computer Engineering, LDRP Institute of
Technology and Research, afiliated with Kadi Sarva Vishwavidyalaya (KSV), for their continuous
support, guidance, and encouragement throughout this work. We also thank the University of
MilanoBicocca for providing the results that made this paper possible.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kasela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Braga</surname>
          </string-name>
          , G. Pasi,
          <string-name>
            <given-names>R.</given-names>
            <surname>Perego</surname>
          </string-name>
          ,
          <article-title>Se-pqa: Personalized community question answering</article-title>
          ,
          <source>in: Companion Proceedings of the ACM Web Conference</source>
          <year>2024</year>
          , volume
          <volume>9</volume>
          <source>of WWW '24</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <year>2024</year>
          , pp.
          <fpage>1095</fpage>
          -
          <lpage>1098</lpage>
          . URL: http://dx.doi.org/10.1145/3589335.3651445. doi:
          <volume>10</volume>
          .1145/3589335.3651445.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>