<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Comparison of Recom mender Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Robin Stenzel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Max Lübbering</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bilge Ulusay</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniel Uedelhoven</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rafet Sifa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fraunhofer IAIS</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Germany</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Allocation</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>LWDA'22: Lernen</institution>
          ,
          <addr-line>Wissen, Daten, Analysen</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Recommender systems, Expert Finding, Community Question Answering</institution>
          ,
          <addr-line>Latent Dirichlet</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Community question answering platforms like Stackoverflow are among the most popular interactive environments on the Internet for individuals to share knowledge. Finding experts to answer questions is one of those platforms' major challenges. To this end, we compare SBERTRec and LDA-Rec, two recommender system algorithms which are based on the state-of-the-art transformer architecture and well-established probabilistic topic modeling algorithm Latent Dirichlet Allocation, respectively. Our results show that SBERT-Rec significantly outperforms LDA-Rec in terms of average rank score. While SBERT-Rec excels in an open-world scenario with no presumptions about the underlying subjects of the corpus, LDA-Rec carves out distinct and human interpretable topics inside a niche closed-world corpus. Finally, we provide a novel metric for expert matching evaluation that supports partial experts/non-experts annotations.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        With free online encyclopedias like Wikipedia and search engines, internet users can
access any bit of information, regardless of time or location. From the start of the
Internet, people used this new freedom to find like-minded people to share knowledge and
ideas. Community-driven question and answer websites like Stackoverflow have emerged
in recent years, matching questioners with domain experts. Many such niche platforms
exist providing a question-and-answer system with an embedded voting mechanism [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>Since these websites are entirely community-regulated, there is no inherent distinction
between experts in specific fields allowing anyone to answer. Thus, users are equally
presented with new questions covering a wide range of topics regardless of personal
expertise. This raises the central question of this work: Can we match questions to
experts solely based on a platform’s historical question and answer data without any
true expert annotations?</p>
      <p>
        E-commerce websites are faced with a similar problem to increase the conversion rate
and leverage recommender systems to suggest products to users matching their interests.
Amazon and Netflix are two well-known examples that recommend products and movies
based on a user’s search or purchase history [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
      </p>
      <p>This paper examines two diferent recommender systems (RS) for question-to-expert
matching: 1) Topic modeling based RS and 2) transformer-embedding based RS. While
the former approach yields interpretable closed-world topics, the latter’s embeddings are
more expressive due to the incorporation of world knowledge, thus leading to superior
recommendation performance. Finally, we propose a new metric for recommender systems
that supports partially annotated data.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>We look at the existing research from three perspectives: recommender systems, statistical
topic models, and expert finding.</p>
      <sec id="sec-2-1">
        <title>2.1. Expert Finding and Statistical Topic Models</title>
        <p>
          Popular online knowledge-sharing communities like Quora and Stackoverflow have become
well-known platforms for finding people with specific knowledge in academic and
nonacademic fields. The goal of these platforms is to present the most relevant experts to a
user searching for a term. LDA topic modeling [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] is one of the methods to achieve this
goal. Using LDA, the relevant topics from a corpus are identified and used to establish
an association among each user and their field of experts [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. LDA topic modeling is
widely used in expert community question answering [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], expert identification [ 9],
and matching companies to news articles [10].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Recommender Systems</title>
        <p>A recommender system is a crucial component of e-commerce, marketing, and social
media platforms that predicts what consumers find interesting. Moreover, recommender
systems are increasingly used in expert finding [ 11] and, more specifically, in finding
experts in software development for design decision making [12].</p>
        <p>A typical recommender system runs on one of the three fundamental engines:
contentbased systems, hybrid filtering, and collaborative filtering-based systems. Content-based
ifltering [ 13] recommends products or services based on item similarity and previous
online activity. This filter avoids a cold start for new items by not relying on other users’
comments [14],[15]. Unlike content-based filtering, collaborative filtering [ 16] suggests
people based on shared interests. Lastly, hybrid filtering methods combine these two
approaches [17].</p>
        <p>By using natural language processing techniques such as BERT [18], recommender
systems can provide more relevant suggestions to users. Several studies have used BERT
in collaborative filtering [ 19] and other recommender systems, significantly improving
recommendations [20],[21]. Furthermore, SBERT [22], a more computational eficient
LDA
(a) LDA-Rec
SBERT
(b) SBERT-Rec
cosine
sim
variant of BERT optimized for sentence similarity estimation, has been used to match
jobs and job seekers [23].</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Approach</title>
      <p>We propose two diferent recommender system approaches, namely LDA-Rec and
SBERTRec, as shown in Fig. 1. LDA-Rec utilizes LDA’s probabilistic topic modeling to represent
questions and users within the topic space. The dissimilarity of question and user pairs
is estimated via Kullback-Leibler divergence. The SBERT-Rec recommender system
comprises a SBERT model for question/user representation and a subsequent cosine
similarity estimation module. LDA-Rec difers fundamentally from SBERT-Rec in the
way the representations are estimated. The LDA model makes a closed-world assumption
learning granular topics within a corpus, whereas the SBERT model incorporates world
knowledge, making it less focused on a particular subject. Further, LDA provides
interpretable corpus-inherent topics, providing valuable insights into the subjects discussed
on these platforms.</p>
      <sec id="sec-3-1">
        <title>3.1. Latent Dirichlet Allocation</title>
        <p>
          Latent Dirichlet Allocation (LDA) is a generative probabilistic model for topic modeling
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. It is a hierarchical Bayesian model that estimates the probability distributions of
topics appearing in a document and of words associated with those topics. For each topic
the model iteratively ranks the vocabulary to maximize their likelihood using variational
methods and the EM algorithm [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Sentence BERT</title>
        <p>Sentence BERT (SBERT) is a fine-tuned BERT model optimized for text similarity
estimation. As shown by [24, 22], the original BERT architecture provides
state-ofthe-art results on semantic textual similarity (STL). This setup, however, requires
both sentences as input for similarity estimation, leading to considerable computational
Mathematics
Users
Questions
Answers</p>
        <p>84,896
1,371,686
1,770,606</p>
        <p>245,701
1,806,605
3,891,942
ineficiencies. [ 25] points out that determining the most similar sentence pairs out of
10,000 sentences requires 50,000,000 inferences. This computational overhead also renders
our recommender system infeasible since our objective is to match questions to potential
experts from a set of more than 10,000 users.</p>
        <p>SBERT provides a solution to this problem in two ways as illustrated in Fig. 2: a) The
embeddings are fine-tuned in a Siamese network setup, which yields embeddings better
capturing a sentence’s semantics. b) The cosine similarity measure is decoupled from the
network itself. The two adaptations allow computing the embeddings only once and then
calculating the sentence similarity ofline. The authors of [ 25] has shown that this more
eficient setup provides state-of-the-art results, making it a superior approach.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>4.1. Evaluation Approach</title>
        <p>While the two datasets represent a typical recommender system setting for expert
matching, they impose complicated challenges on recommender system evaluation. This
is because only a tiny fraction of experts respond to a question and the remaining
unknown experts are indistinguishable from non-experts.</p>
        <p>Unknown experts can be expected to achieve high similarity scores, rendering
recommender metrics such as precision@k and recall@k with their constant  infeasible.
Similarly, the order-aware mean reciprocal rank metric only considers the top-ranked
true expert, which greatly depends on the unknown ratio of true experts vs. unknown
experts, and disregards the remaining true experts.</p>
        <p>To this end, we propose the average rank metric
(, , ) =
{
1, if sim(, ) &gt;
0, otherwise</p>
        <p>sim(, )
rel rank(, ) =
average rank() =</p>
        <p>1 ∑ (, , )
| \{}| ∈ \{}
1 ∑ rel rank(, ),
|  | ∈ 
(1)
(2)
(3)
which averages the relative rank of every true expert  ∈   of question  among
all remaining users  ∈  \{} based on similarity function sim(, ) , irrespective of true
experts, non-experts, and unknown experts. Analogous to AUROC in outlier detection
setting [27, 28], this metric can be interpreted as the probability of a random true expert
being ranked higher than a randomly sampled user. As a result, the average rank metric
is independent of dataset size and considers all true experts within the evaluation.</p>
        <p>LDA-Rec and SBERT-Rec are evaluated with respect to their relative rank over all
questions in a dataset. In case of KBL divergence, we compute its inverse so a higher
value corresponds to a higher similarity.</p>
        <p>(a) CS
(b) Math</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Datasets</title>
        <p>We trained and evaluated our models on two community question answering datasets.
The first one covers mathematics and contains questions and answers from Math Stack
Exchange1, while the second covers computer science and software engineering from
Stack Overflow 2.The Internet Archive3 has a data dump from both sites’ posts. The
datasets are large-scale, well-known, and contain a wide range of topics, questions, and
users. They are an excellent fit for our research question since many learning platforms
are designed similarly.</p>
        <p>We applied multiple preprocessing steps to the datasets. First, we created questions
by concatenating the titles with the corresponding main texts and removed users with
less than 50 answers to assure expert representations of high quality. Then we split each
question’s text into its constituent words while converting each character into lower case
and applying lemmatization. Moreover, we filtered out special characters, stop words,
and one-letter words.</p>
        <p>The key idea is to represent a user by all the questions he answered and measure its
similarity to the representation of the question at hand. We assume a user to be a true
expert for all questions that he answered. Likewise, we consider a user that answered
similar questions before but not this one as an unknown expert and everyone else as
non-experts. Note that unknown experts are indistinguishable from non-experts within
the dataset.</p>
        <p>Table 1 shows the final number of questions, answers, and users in each dataset.
(a) LDA on CS</p>
        <p>(b) LDA on Math
(c) SBERT on CS
(d) SBERT on Math</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Results</title>
        <p>The LDA model has been trained for 500 iterations and 100 passes. For the Math and CS
datasets, we determined the number of topics to be 12 and 14, respectively, as supported
by the topic coherence development, shown in Fig. 3. For the SBERT-Rec, we used the
pre-trained model without further fine-tuning.</p>
        <p>As shown in Table 2, SBERT-Rec significantly outperforms LDA-Rec on both Math
and CS datasets by 7 and 13 percentage points, respectively. Further, LDA only improves
slightly on the CS dataset compared to the Math dataset, whereas SBERT’s enhances
7 percentage points. As a topic in CS can be arbitrary, e.g., due to the advent of new
technologies every year, it can be assumed that the CS dataset comprises more distinct
topics than the Math dataset. This is in line with the increase in average rank scores on
the CS dataset for both models.</p>
        <p>1https://math.stackexchange.com, Math Stack Exchange
2https://stackoverflow.com, Stack Overflow
3https://archive.org/details/stackexchange, The Internet Archive</p>
        <p>Id Name Top four words
1 General request use, like, would, want
2 Code new, class, string, public
3 Backend application, server, run, error
4 File handling file, c, x, b
5 User input name, value, type, form
6 Java script function, page, text, html
7 Databases table, query, database, id
8 Websites http, com, xml, url
9 Authentification user, view, self, model
10 Java webservice java, org, service, web
11 HTML div, td, px, width
12 Android development android, layout, parent, id
13 .net web service event, asp, control, net
14 Images list, image, item, li</p>
        <p>Additionally to the average rank scores, we plotted the density histograms of
expert/question similarities and non-expert/question similarities, as displayed in Fig. 4. On
both datasets, the similarity histograms are better separated, compliant with
SBERTRec’s higher average rank scores.</p>
        <p>While SBERT-Rec can be regarded as a black-box model, each topic learned by LDA
is the vocabulary ranked by topic relevancy and thus interpretable. As shown in Table 3
and Table 4, the set of learned topics within CS is more distinct than the topics within
the Math dataset, explaining the superior scores on the CS dataset. Nevertheless, for
LDA-Rec, the average rank score diference between the two datasets is less significant
in comparison to SBERT-Rec. This pinpoints LDA’s advantage as a closed-set topic
modeling algorithm to learn niche corpora.</p>
        <p>In conclusion, we have empirically shown that both methods are practical recommender
systems tailored for orthogonal settings. While SBERT-Rec performs best in an
openworld scenario without presumptions on the corpus-inherent topics, LDA-Rec can carve
out clear topics within a niche closed-world corpus.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study presents two recommender systems for expert finding in community question
answering platforms like Stack Overflow, using LDA topic modeling and the
transformerbaed SBERT model. We test our approach on large-scale datasets from Stack Overflow
and Math Stack Exchange, demonstrating its efectiveness and delivering high-quality
expert matching results. Our results reveal that SBERT-Rec outscored LDA-Rec on
both datasets based on the average rank score. While SBERT-Rec performs better in
an open-world scenario with no presumptions about the corpus’s underlying subjects,
LDA-Rec finds unique topics inside a particular closed-world corpus. In terms of future
research, it would be interesting to employ the auto-regressive language model Generative
Pre-trained Transformer 3 (GPT-3), which incorporates even more world knowledge. We
will also investigate the efect of tags on the representation of the questions.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>In parts, the authors of this work were funded by the Federal Ministry of Education
and Research of Germany. The authors would also like to thank the Daniel Jung Media
GmbH for their insightful input.
International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, 2015,
pp. 176–185. doi:10.1109/CyberC.2015.87.
[9] R. Chi, L. Wang, Expert identification based on dynamic lda topic model, 2018, pp. 881–888.</p>
      <p>doi:10.1109/DSC.2018.00141.
[10] M. Lübbering, J. Kunkel, P. Farrell, What company does my news article refer to? tackling multiclass
problems with topic modeling (2019).
[11] H. Chen, A. G. O. II, C. L. Giles, Expertseer: a keyphrase based expert recommender for digital
libraries, CoRR abs/1511.02058 (2015). URL: http://arxiv.org/abs/1511.02058. arXiv:1511.02058.
[12] M. Bhat, K. Shumaiev, K. Koch, U. Hohenstein, A. Biesdorf, F. Matthes, An expert recommendation
system for design decision making: Who should be involved in making a design decision?, in: 2018
IEEE International Conference on Software Architecture (ICSA), 2018, pp. 85–8509. doi:10.1109/
ICSA.2018.00018.
[13] M. J. Pazzani, D. Billsus, Content-based recommendation systems, in: The Adaptive Web, 2007.
[14] S. Zahoor, Addressing cold start problem in recommendation systems with collaborative filtering
and reverse collaborative filtering, International Journal of Computer Sciences and Engineering 6
(2018) 211–214. doi:10.26438/ijcse/v6i4.211214.
[15] J. Bobadilla, F. Ortega, A. Hernando, J. Bernal, A collaborative filtering approach to mitigate the
new user cold start problem, Knowl. Based Syst. 26 (2012) 225–238.
[16] J. B. Schafer, D. Frankowski, J. Herlocker, S. Sen, Collaborative filtering recommender systems,
2007.
[17] R. Burke, Hybrid recommender systems: Survey and experiments, User Modeling and User-Adapted</p>
      <p>Interaction 12 (2002). doi:10.1023/A:1021240730564.
[18] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional transformers
for language understanding, CoRR abs/1810.04805 (2018). URL: http://arxiv.org/abs/1810.04805.
arXiv:1810.04805.
[19] T. Wang, Y. Fu, Item-based collaborative filtering with bert, 2020, pp. 54–58. doi: 10.18653/v1/
2020.ecnlp-1.8.
[20] G. Cenikj, S. Gievska, Boosting recommender systems with advanced embedding models, 2020, pp.</p>
      <p>385–389. doi:10.1145/3366424.3383300.
[21] Z. Qiu, X. Wu, J. Gao, W. Fan, U-bert: Pre-training user representations for improved
recommendation, Proceedings of the AAAI Conference on Artificial Intelligence 35 (2021) 4320–4327. URL:
https://ojs.aaai.org/index.php/AAAI/article/view/16557.
[22] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, V. Stoyanov,</p>
      <p>Roberta: A robustly optimized bert pretraining approach, arXiv preprint arXiv:1907.11692 (2019).
[23] D. Lavi, V. Medentsiy, D. Graus, consultantbert: Fine-tuned siamese sentence-bert for
matching jobs and job seekers, CoRR abs/2109.06501 (2021). URL: https://arxiv.org/abs/2109.06501.
arXiv:2109.06501.
[24] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional transformers
for language understanding, arXiv preprint arXiv:1810.04805 (2018).
[25] N. Reimers, I. Gurevych, Sentence-bert: Sentence embeddings using siamese bert-networks, arXiv
preprint arXiv:1908.10084 (2019).
[26] F. Schrof, D. Kalenichenko, J. Philbin, Facenet: A unified embedding for face recognition and
clustering, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2015).</p>
      <p>URL: http://dx.doi.org/10.1109/CVPR.2015.7298682. doi:10.1109/cvpr.2015.7298682.
[27] M. Lübbering, M. Gebauer, R. Ramamurthy, C. Bauckhage, R. Sifa, Decoupling autoencoders for
robust one-vs-rest classification, in: 2021 IEEE 8th International Conference on Data Science and
Advanced Analytics (DSAA), IEEE, 2021, pp. 1–10.
[28] M. Lübbering, M. Gebauer, R. Ramamurthy, C. Bauckhage, R. Sifa, Bounding open space risk
with decoupling autoencoders in open set recognition, International Journal of Data Science and
Analytics (2022) 1–23.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Treude</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barzilay</surname>
          </string-name>
          , M.
          <article-title>-A. Storey, How do programmers ask and answer questions on the web?(nier track)</article-title>
          ,
          <source>in: Proceedings of the 33rd international conference on software engineering</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>804</fpage>
          -
          <lpage>807</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Linden, Two decades of recommender systems at amazon</article-title>
          . com,
          <source>Ieee internet computing 21</source>
          (
          <year>2017</year>
          )
          <fpage>12</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Gomez-Uribe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hunt</surname>
          </string-name>
          ,
          <article-title>The netflix recommender system: Algorithms, business value, and innovation</article-title>
          ,
          <source>ACM Transactions on Management Information Systems (TMIS) 6</source>
          (
          <issue>2015</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Blei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. I. Jordan</surname>
          </string-name>
          , Latent dirichlet allocation,
          <source>the Journal of machine Learning research 3</source>
          (
          <year>2003</year>
          )
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Momtazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <article-title>Topic modeling for expert finding using latent dirichlet allocation</article-title>
          ,
          <source>Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</source>
          <volume>3</volume>
          (
          <year>2013</year>
          ). doi:
          <volume>10</volume>
          .1002/widm.1102.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Riahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zolaktaf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shafiei</surname>
          </string-name>
          , E. Milios,
          <article-title>Finding expert users in community question answering</article-title>
          ,
          <source>WWW'12 - Proceedings of the 21st Annual Conference on World Wide Web Companion</source>
          (
          <year>2012</year>
          ). doi:
          <volume>10</volume>
          .1145/2187980.2188202.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Predicting best answerers for new questions: An approach leveraging distributed representations of words in community question answering</article-title>
          ,
          <source>in: 2015 Ninth International Conference on Frontier of Computer Science and Technology</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>18</lpage>
          . doi:
          <volume>10</volume>
          .1109/FCST.
          <year>2015</year>
          .
          <volume>56</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. LI</surname>
          </string-name>
          ,
          <article-title>A hybrid model for experts finding in community question answering</article-title>
          ,
          <source>in: 2015</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>