<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Resolving duplicates in Large Multiple-Choice Questions Repositories</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Valentina Albano</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Donatella Firmani</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luigi Laura</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jerin George Mathew</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anna Lucia Paoletti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Irene Torrente</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dip. Funzione Pubblica, Corso Vittorio Emanuele II</institution>
          ,
          <addr-line>116, 00186 Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Formez</institution>
          ,
          <addr-line>Viale Marx, 15, 00137 Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Sapienza University</institution>
          ,
          <addr-line>Piazzale Aldo Moro, 5, 00185 Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Uninettuno University</institution>
          ,
          <addr-line>Corso Vittorio Emanuele II, 39, 00186 Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <fpage>1</fpage>
      <lpage>1</lpage>
      <abstract>
        <p>Multiple-choice questions (MCQs) are commonly used in educational assessments and professional certification examinations. However, managing vast collections of MCQs presents numerous challenges, including maintaining their quality and relevance. A notable issue in such repositories is the occurrence of conceptually identical questions presented in varied forms. These duplicates, while diferent in wording, fail to enhance the value of the repository. In this extended abstract, we present our approach for identifying and handling potential duplicate questions in large MCQ databases. Our proposed method involves three primary stages: initial pre-processing of MCQs, calculation of similarity based on Natural Language Processing (NLP) techniques, and a graph-based method for exploring these similarities.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;multiple-choice questions</kwd>
        <kwd>entity resolution</kwd>
        <kwd>record linkage</kwd>
        <kwd>graph communities</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Multiple-Choice Questions (MCQs) are widely utilized for knowledge assessment across various
domains, from university admissions and job evaluations to self-assessment and entertainment,
including popular game shows and mobile gaming apps. Large-scale standardized tests typically
feature MCQs with four response options: one correct answer and three distractors.</p>
      <p>
        Academic research primarily focuses on the efectiveness of MCQs as evaluation tools.
Azevedo, Oliveria, and Damas Beites’ study [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is exemplary in exploring methods for fair
student assessments through MCQ analysis. Learning Analytics, as defined in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], involves the
comprehensive measurement and analysis of learner data to optimize educational environments.
Furthermore, the laborious nature of manual MCQ creation has led researchers to the
development of automatic generation techniques. These range from using resources like WordNet and
shallow parsing to advanced methods involving ontologies and deep neural networks [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>Unlike previous research, our approach focuses on the maintenance of large Multiple-Choice
Questions (MCQs) repositories with a data quality perspective. Specifically, we focus on the
identification of conceptually similar questions within these repositories. Our method can
be used by MCQs database administrators to identify overly similar questions and verifying
question coherence with syllabus areas and coverage. Our intuition is rooted in the believe that
redundant or less coherent questions can be accumulated in MCQs repositories over time – e.g.,
upon addition and merging operation – which can dilute their quality and efectiveness.</p>
      <p>The primary feature of our method is the utilization of recent Natural Language Processing
(NLP) techniques in an entirely unsupervised environment. These advanced NLP methods,
based on the popular Transformer architecture, are capable of identifying a substantial number
of relevant matches. However, they also carry the risk of generating false positives, where
questions might be inaccurately classified as similar. To mitigate this issue of false positives,
our approach incorporates an innovative graph exploration technique. This technique focuses
on identifying candidate matches from questions that are part of densely connected graph
communities. By doing so, our method not only minimizes the likelihood of false positives but
can also uncover the underlying structure of the MCQs repository.</p>
      <p>
        Outline. Section 2 describes the main components of our approach and a case study from a
large-scale training project by the Italian Government. Section 3 summarizes our experimental
results in the main application scenarios. Section 4 describes related works and Section 5
presents conclusive remarks. For more detailed insights, please refer to the recent journal paper
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and to the conference paper in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] presenting an earlier version of our approach.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Overview</title>
      <p>In this section, we outline our proposed workflow for managing large Multiple Choice Questions
(MCQs) repositories, which includes the following four key steps:
1. Similarity Computation: Compute similarity scores between all pairs of questions in
an unsupervised manner, without the need for a labeled dataset.
2. Threshold Definition : Establish a similarity threshold,  , to preliminarily distinguish
between similar and dissimilar questions. This threshold is adjustable, allowing interactive
user selection.
3. Graph Construction: Construct a graph,  , where nodes represent questions, and
edges connect questions with similarity scores above the threshold,  .
4. Graph Exploration: Employ graph visual analysis, focusing on graph communities to
identify clusters of potentially similar questions.</p>
      <p>Case study. We demonstrate our approach using the MCQs database from the Competenze
Digitali program.1 The program aims to provide non-IT public employees with personalized
e-learning training in basic digital skills. Key elements include (i) a Syllabus outlining minimum
digital skills required for public employees, (ii) an online platform for skill gap analysis, training
course definition, and delivery, and (iii) a catalog of quality training developed in collaboration
1Database access is restricted, but the new program is publicly available at https://www.syllabus.gov.it.
Question: What is the definition of “cookies”?
A. Small files that store information about online browsing on the user’s computer or device.
B. Small files that track and collect user’s online behaviors for targeted advertising purposes.
C. Tokens generated by web applications to authenticate user sessions and enable personalized
features.</p>
      <p>D. Privacy settings that allow users to control the information shared with websites they visit.
bchain#1bchain#2netsec#1netsec#2email#1email#2db#1 db#2 ai#1 ai#2
(a)
(b)
with major public and private entities. The MCQ dataset of the Competenze Digitali program
comprises 798 Italian language questions, each presenting four candidate answers, of which
only one is correct. Every question corresponds to a particular syllabus, which groups together
questions related to the same topic (e.g. computer networks). In total, there are 11 distinct
syllabi in the dataset. We provide a sample question in Figure 1.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Applications</title>
      <sec id="sec-3-1">
        <title>3.1. Identifying redundant questions</title>
        <p>
          We describe a method for identifying overly similar questions in the Competenze Digitali
MCQs database. In our experiment, we tested four multilingual models from the Sentence
Transformers library [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] – specifically fine-tuned with Italian data – that are ST-Roberta,
ST-DistilUSE, ST-MiniLM, and ST-MPNet. We assessed their performance on a subset of
our MCQ dataset, which included manually selected question pairs related to the same syllabus
topic, and identified ST-MPNet as the best-performing model. The resulting similarity matrix
from ST-MPNet, illustrated in Figure 2a, displays a pattern of darker 2 × 2 squares along the
diagonal, indicating accurate semantic matches of pairs of most similar questions.
        </p>
        <p>A critical consideration in our approach is whether to compute similarity using only the
question text, the question with the correct answer, or the question with all answer options.
Our analysis suggests that including all answer options provides the most accurate results,
particularly for questions with identical wording, such as those in our repository that simply
ask, “Which of the following statements is false?”.</p>
        <p>
          When analyzing the similarity values, we found a spectrum ranging from completely unrelated
questions (similarity 0) to those that were suspiciously similar (similarity values between 0.98
and 0.99). To address the intermediate range of similarities, we implemented a graph-based
method by setting a threshold  = 0.7. We examined the graph 0.7, consisting of 718 nodes and
3, 836 edges, and applied the Clauset-Newman-Moore algorithm [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] for community detection.2
The communities identified, as depicted in Figure 2b, are often granular enough to enable
manual inspection and identification of similar question groups. For cases requiring additional
analysis, node-centrality methods such as [
          <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
          ] can be employed to enhance visual exploration.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Verifying syllabus coherence</title>
        <p>Using the previously described similarity techniques and a structured syllabus, we can eficiently
assess the alignment of questions with specific syllabus areas. For instance, in the Competenze
Digitali program, the syllabus comprises five main areas and 11 sub-areas. Each question in our
database is uniquely identified by the area and sub-area numbers it corresponds to.</p>
        <p>This structure allows us to analyze the similarity between each question and its designated
sub-area within the syllabus. In our dataset we have 798 questions and 11 sub-areas, so we
computed 798 · 11 = 8778 values of similarity. In Table 3 we report the head and the tail of the
values; the first items are the similarity score of the first question in the dataset against the first
sub-areas of the Syllabus. It is easy to see that the largest score, among the ones shown in the
table, is exactly for the sub-area of the Syllabus it belongs, i.e., Q-1.1.1.1-1 belongs to S1.1. The
same happens for the last question, i.e., Q-5.2.3.5-8 belongs to S5.2. About half of the questions
in our dataset showed the highest similarity with their respective syllabus sub-areas.</p>
        <p>
          2The results of a baseline approach based on connected components instead of communities are available in [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>This outcome highlights the efectiveness of our approach in ensuring that questions are
relevant and aligned with specific curriculum areas. By examining these similarity scores,
educators and administrators of the MCQs repository can identify syllabus areas needing more
focus or refinement and detect potential redundancies in their question sets.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Related Works</title>
      <p>
        The exploration of duplicate question detection has been investigated in the Q&amp;A domain,
aiming to eficiently answer queries by linking to similar, previously answered questions in
Q&amp;A forums. Notable recent works include those of Li et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], who focused on medical Q&amp;A
platforms, and Kamienski et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], who concentrated on game development forums. Both
studies developed deep-learning systems for recognizing similar questions. Li et al. trained a
Long-Short Term Memory (LSTM) neural network on pairs of questions, aligning semantically
similar queries within the LSTM’s vector space. Kamienski et al. combined large pre-trained
deep learning models with supervised techniques, using features from models like MPNet to
train a supervised model that predicts similarity scores between sentence pairs.
      </p>
      <p>
        In the field of learning analytics, there has been limited investigation into identifying
duplicate questions. A notable study is presented in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], which developed a machine learning-based
system for managing large question paper databases. It trains an XGBoost model [13] on
manually-selected features like structural attributes and word embeddings, using labeled
duplicate question pairs from Quora to identify semantically similar English sentence pairs.
      </p>
      <p>Unlike these methods, our study leverages pre-trained large language models in an
unsupervised manner, without requiring ground-truth labels. Additionally, our approach incorporates a
graph construction phase to facilitate the identification of duplicates.</p>
      <p>Other works. The general problem of identifying duplicate records in databases is known
in literature as Entity Resolution. Data management applications typically use supervised
methods [14, 15] and external knowledge bases [16] to learn vector representations of records
and attributes. In this paper, we focus instead on unsupervised similarity computation and leave
the final decision to a human expert, in the spirit of oracle-based approaches such as [17, 18].</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusive remarks</title>
      <p>We describe our approach for the maintenance of large Multiple-Choice Questions (MCQs)
repositories, based on our experience with a large-scale training project by the Italian
Government. Our method employs Natural Language Processing (NLP) techniques, particularly
Transformer architectures, to semantically detect similar pairs of questions in MCQs repositories,
aiming to go beyond traditional word co-occurrence analysis while acknowledging the potential
for false positives. To address these false positives, we incorporate a graph exploration strategy
that focuses on community structures within the similarity graph, thus enhancing the precision
of similarity assessments by analyzing relationships within the same community.</p>
      <p>Future works include exploring algorithms for eficiently merge multiple MCQ repositories
and developing automated tools for aligning questions upon syllabus updates, thereby reducing
the manual efort required for repository maintenance.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>The authors have been partially supported by SEED PNR Project “FLOWER” “Frontiers in
Linking records: knOWledge graphs, Explainability and tempoRal data” and Sapienza Research
Project B83C22007180001 “Trustworthy Technologies for Augmenting Knowledge Graphs”.
system, in: 2019 IEEE Tenth International Conference on Technology for Education (T4E),
2019, pp. 266–267. doi:10.1109/T4E.2019.00067.
[13] T. Chen, C. Guestrin, Xgboost: A scalable tree boosting system, in: Proceedings of the
22nd acm sigkdd international conference on knowledge discovery and data mining, 2016,
pp. 785–794. doi:10.1145/2939672.2939785.
[14] U. Brunner, K. Stockinger, Entity matching with transformer architectures-a step forward
in data integration, in: EDBT 2020, 2020. doi:10.5441/002/edbt.2020.58.
[15] M. Ebraheem, S. Thirumuruganathan, S. R. Joty, M. Ouzzani, N. Tang, Distributed
representations of tuples for entity resolution, Proc. VLDB Endow. 11 (2018) 1454–1467. URL: http:
//www.vldb.org/pvldb/vol11/p1454-ebraheem.pdf. doi:10.14778/3236187.3236198.
[16] A. S. Andreou, D. Firmani, J. G. Mathew, M. Mecella, M. Pingos, Using knowledge graphs for
record linkage: Challenges and opportunities, in: M. Ruiz, P. Sofer (Eds.), Advanced
Information Systems Engineering Workshops - CAiSE 2023 International Workshops, Zaragoza,
Spain, June 12-16, 2023, Proceedings, volume 482 of Lecture Notes in Business Information
Processing, Springer, 2023, pp. 145–151. URL: https://doi.org/10.1007/978-3-031-34985-0_15.
doi:10.1007/978-3-031-34985-0\_15.
[17] D. Firmani, S. Galhotra, B. Saha, D. Srivastava, Robust entity resolution using a crowdoracle,
IEEE Data Eng. Bull. 41 (2018) 91–103. URL: http://sites.computer.org/debull/A18june/p91.
pdf.
[18] S. Galhotra, D. Firmani, B. Saha, D. Srivastava, Eficient and efective er with progressive
blocking, The VLDB Journal 30 (2021) 537–557. doi:10.1007/s00778-021-00656-7.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Azevedo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. P.</given-names>
            <surname>Oliveira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Damas</surname>
          </string-name>
          <string-name>
            <surname>Beites</surname>
          </string-name>
          ,
          <article-title>Using learning analytics to evaluate the quality of multiple-choice questions: A perspective with classical test theory and item response theory</article-title>
          ,
          <source>The International Journal of Information and Learning Technology</source>
          <volume>36</volume>
          (
          <year>2019</year>
          )
          <fpage>322</fpage>
          -
          <lpage>341</lpage>
          . doi:
          <volume>10</volume>
          .1108/IJILT-02-2019-0023.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Long</surname>
          </string-name>
          , G. Siemens,
          <string-name>
            <given-names>G.</given-names>
            <surname>Conole</surname>
          </string-name>
          , D. Gasevic (Eds.),
          <source>Proceedings of the 1st International Conference on Learning Analytics and Knowledge</source>
          ,
          <string-name>
            <surname>LAK</surname>
          </string-name>
          <year>2011</year>
          ,
          <article-title>Banf</article-title>
          ,
          <string-name>
            <surname>AB</surname>
          </string-name>
          , Canada, February 27 - March 01,
          <year>2011</year>
          , ACM,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mitkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. Le</given-names>
            <surname>An</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Karamanis</surname>
          </string-name>
          ,
          <article-title>A computer-aided environment for generating multiple-choice test items</article-title>
          ,
          <source>Natural language engineering 12</source>
          (
          <year>2006</year>
          )
          <fpage>177</fpage>
          -
          <lpage>194</lpage>
          . doi:
          <volume>10</volume>
          . 1017/S1351324906004177.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Albano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Firmani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Laura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. G.</given-names>
            <surname>Mathew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Paoletti</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Torrente</surname>
          </string-name>
          ,
          <article-title>Nlp-based management of large multiple-choice test item repositories</article-title>
          ,
          <source>Journal of Learning Analytics</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>V.</given-names>
            <surname>Albano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Firmani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Laura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Paoletti</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Torrente</surname>
          </string-name>
          ,
          <article-title>Managing large multiple-choice test item repositories</article-title>
          ,
          <source>in: Proceedings of the 26 International Conference Information Visualisation (IV22)</source>
          ,
          <year>2022</year>
          . doi:
          <volume>10</volume>
          .1109/IV56949.
          <year>2022</year>
          .
          <volume>00054</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1908</year>
          .10084. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1908</year>
          .
          <volume>10084</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Clauset</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E. J.</given-names>
            <surname>Newman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <article-title>Finding community structure in very large networks</article-title>
          ,
          <source>Physical Review E</source>
          <volume>70</volume>
          (
          <year>2004</year>
          ). doi:
          <volume>10</volume>
          .1103/physreve.70.066111.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ausiello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Firmani</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Laura,</surname>
          </string-name>
          <article-title>The (betweenness) centrality of critical nodes and network cores</article-title>
          ,
          <source>in: 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC)</source>
          , IEEE,
          <year>2013</year>
          , pp.
          <fpage>90</fpage>
          -
          <lpage>95</lpage>
          . doi:
          <volume>10</volume>
          .1109/IWCMC.
          <year>2013</year>
          .
          <volume>6583540</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ausiello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Firmani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Laura</surname>
          </string-name>
          ,
          <article-title>Real-time monitoring of undirected networks: Articulation points, bridges, and connected and biconnected components</article-title>
          ,
          <source>Networks</source>
          <volume>59</volume>
          (
          <year>2012</year>
          )
          <fpage>275</fpage>
          -
          <lpage>288</lpage>
          . doi:
          <volume>10</volume>
          .1002/net.21450.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , W. Fan,
          <article-title>Finding similar medical questions from question answering websites</article-title>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1810</year>
          .
          <volume>05983</volume>
          . arXiv:
          <year>1810</year>
          .05983.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kamienski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hindle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-P.</given-names>
            <surname>Bezemer</surname>
          </string-name>
          ,
          <article-title>Analyzing techniques for duplicate question detection on q&amp;a websites for game developers</article-title>
          ,
          <source>Empirical Software Engineering</source>
          <volume>28</volume>
          (
          <year>2023</year>
          )
          <article-title>17</article-title>
          . doi:
          <volume>10</volume>
          .1007/s10664-022-10256-w.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Duplicate question management and answer verification</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>