<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Recommendation of Academic Collaborators: A Methodology Incorporating Word Embedding and Network Embedding</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Xiaowen Xi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ying Guo</string-name>
          <email>guoying_cupl@126.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Weiyu Duan</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Archives of Chinese Academy of Sciences</institution>
          ,
          <addr-line>Beijing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>China University of Political Science and Law</institution>
          ,
          <addr-line>Beijing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Fruitful academic collaborations have become increasingly more important for solving scientific problems, participating in research projects, and improving productivity. As such, frameworks for recommending suitable collaborators are attracting extensive attention from scholars. In an effort to improve on the current solutions, we developed an approach that produces recommendations with better precision, recall, and accuracy. Our strategy is to leverage the benefits of the two most common similarity indicators for collaborator recommendation - research interests and co-authorship network topology into one unified framework. A Word2Vec model creates word embeddings of research interests, which solves the problem of calculating similarity solely based on co-occurrence, not context, while a Node2Vec model automatically extracts and learns the topological features of a co-authorship network, moving beyond just local features to capture global network features as well. The two similarity measures are then fused with CombMNZ resulting in a ranked list of recommended collaborators for the target scholar(s). The workings of the framework and its benefits are demonstrated through a case study on academics in the field of intelligent driving and a comparison with the two most commonly used baselines: Random Walk with Restart (RWR) and Latent Dirichlet Allocation (LDA). Our framework should be of benefit for academics, research centers, and private-enterprise R&amp;D managers to find partners, so as to achieve the ultimate goal of completing research projects, solving scientific problems, and promoting discipline development and progress.</p>
      </abstract>
      <kwd-group>
        <kwd>Academic Collaborator Recommendation</kwd>
        <kwd>Research Interest</kwd>
        <kwd>Network Topology</kwd>
        <kwd>Word Embedding</kwd>
        <kwd>Network Embedding</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        The complexity and diversity of academic activities are ever-expanding, yet, with
each new breakthrough, the intersections between disciplines are becoming more and
more obvious. While not without its advantages, the increasing level of crossover
necessary to find comprehensive solutions is making scientific research more difficult and
often beyond the capabilities of a single researcher or a research institution[
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Hence,
academic cooperation has gradually become the modus operandi for conducting
research. As Guns and Rousseau state[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], academic cooperation helps to improve the
efficiency of scientific inquiry and research output[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In this vein, scholars and scientists,
just like the professionals in any sector, typically aspire to collaborate with the
highestlevel researchers in their field possible. Often, the aim is to establish a joint research
team to exchange knowledge, share resources and, hopefully, use the power of new and
different perspectives to generate thinking greater than the sum of its parts. The ultimate
goal, of course, is to successfully complete research projects, find high-quality solutions
to scientific problems with greater efficiency, and to contribute to the development and
progress of the entire field.
      </p>
      <p>
        However, accomplishing all these objectives depends on identifying advantageous
collaborators in the first place. Thus, recommendation frameworks for screening
potential scientific collaborators have been a topic of intense focus for some time. Existing
systems generally fall into one of two categories: those that recommend collaborators
based on similar research interests[
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ] and those that explore co-authorship networks[
        <xref ref-type="bibr" rid="ref7 ref8">7,
8</xref>
        ]. Frameworks based on similar research interests are typically built around text mining
techniques that extract topics via keywords, subject terms, or labels. Recommendations
are then calculated based on co-occurrence indexes or the like. The problem is that
these types of indicators cannot capture the context in which the topic was mentioned,
and so cannot factor that information into a recommendation. This can easily result in
an inaccurate representation of a scholar’s research interests and, thus, inappropriate
recommendations. The frameworks designed to explore co-authorship networks
generally base their similarity calculations on the network’s topological features, using
indicators like the Common Neighbor Index (CN), the Restart Random Walk Index (RWR),
the Local Path Index (LP), and so on. The problem with these approaches is that, first,
each recommendation problem requires its own custom-refined indicator or set of
indicators, and designing the ‘perfect’ indicator is a task that requires a great deal of finesse.
Second, current topological indicators only capture local features, such as direct paths
or common neighbors or others. They do not typically provide “big picture”
information about the entire network.
      </p>
      <p>With the aim of addressing these concerns, we developed a novel framework for
recommending academic collaborators that leverages both types of indicators through
word and network embedding. There are three main steps to the process: 1) extract the
research interests of scholars from a corpus of articles with the Word2Vec model, then
calculate the similarity between scholars’ interests in terms of cosine distance; 2)
construct a co-authorship network, then extract and calculate the similarities between
topological features with the Node2Vec model; and 3) integrate the results of both
similarity measures using the CombMNZ method to produce a ranked list of
recommendations for the target scholar. Additionally, to verify the effectiveness and efficiency of
our framework, we conducted an empirical analysis on the field of intelligent driving
and compared the results to the two most common recommendation approaches for
finding potential collaborators used today. The results show our recommendations have
higher accuracy, recall rate and F1. The approach can be applied to a range of
fields/sectors/industries with little to no modifications. Academics, research centers, and
privateenterprise R&amp;D managers should find the insights and recommendations provided by
our system highly useful.
2</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORKS</title>
      <p>
        In early studies on general recommender systems, the scholar’s research interest was
mainly captured by extracting salient terms and phrases from the dataset. The next main
advancement came with feature weighting through techniques such as Term
FrequencyInverse Document Frequency (TF-IDF)[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. However, due to the ambiguity of natural
language (synonyms, polysemy, etc.), comparing scholars based on keywords does not
always accurately reflect the actual similarity of their academic interests. Topic models
are credited with solving some of these ambiguity problems with language and also for
raising general interest in feature extraction[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. However, topic models treat documents
as a bag-of-words and assume that words occur independently. Without context, the
recommendations produced cannot be completely reliable. The Word2Vec model,
however, is an efficient word embedding technique that is able to learn the semantics of
terms in context and form a dense, low-dimensional vector for each word[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Thus, we
apply Word2Vec model to extract more finely-grained features representing the
research interests of scholars in our framework, which substantially improves the
accuracy of the recommendations.
      </p>
      <p>
        Among the network analysis approaches to collaborator recommendation,
co-authorship prediction is an important line of work. These studies have employed and
combined several similarity indicators in co-authorship network to predict and recommend
collaborators[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. For example, kong et al. used Random walk with restart model (RWR)
to measure the academic impact of researchers on the collaborator network[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
However, authors’ features mainly depend on manual design and selection, so it is necessary
to realize automatic extraction of network topology features. In addition, the
computational complexity that comes with increasing dataset remains a complicated and
difficult task. The Node2vec model could transform the semantic information of nodes in
the original network into a low dimensional vector space and effectively preserves the
network structure of nodes, which can efficiently calculate the semantic connections
between nodes in the network[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Thus, to extract features automatically and accurately,
the Node2vec model is exploited to generate feature vectors of scholars’ network
topology.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>METHODOLOGY</title>
      <p>The recommendation framework for academic collaborators based on word
embedding and network embedding model proposed in this paper comprises four main steps:
data acquisition and preprocessing; measuring the similarity of research interests
between scholars; measuring the similarity of topological features between scholars; and
recommending suitable academic collaborators with the CombMNZ model. An
overview of the framework is provided in Figure 1.</p>
      <p>This step involves retrieving and downloading academic articles from the Web of
Science database, then using a professional desktop text mining
software—VantagePoint (https://www.thevantagepoint.com/)—to extract key features such as the author,
year of publication, title, and abstract. With a raw dataset of terms assembled, the
subsequent data preprocessing procedure cleans the terms and disambiguates the author
names in two separate steps.</p>
      <p>
        The procedure of cleaning terms is as follows. First, the title and abstract fields are
merged, and VantagePoint performs word segmentation. Noise is then removed and
synonyms are merged with a term clumping process based on a fuzzy semantic
matching algorithm developed by Zhang et al.[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Terms and phrases appearing more than six
times are then extracted for further analysis by experts who remove general and
irrelevant terms, such as development, methods, significant, etc. The final culled list forms
the vocabulary of core terms.
      </p>
      <p>Authors with the same names are disambiguated through a two-dimensional matrix
where the rows contain the names and the columns contain their affiliations. A fuzzy
matching algorithm then merges duplicate scholar names and institutions.
3.2</p>
      <sec id="sec-3-1">
        <title>Research interest similarity</title>
        <p>
          Word2Vec focuses on sequential combinations of words in a corpus and exploits the
idea of neural networks to train a language model that maps each word to a vector.
Word2Vec includes two model options for updating parameters to suit different
situations[
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. For our purposes, the training procedure in Skip-gram produces a more
accurate result.
        </p>
        <p>To obtain accurate eigenvectors of the scholars’ research interests, this step is to
produce accurate eigenvectors of scholar’s research interest. More specifically, given a
series of documents D = { 1,  2, … ,   } with a vocabulary of N words{ 1,  2, … ,   },
the Word2Vec model maps each word in the vocabulary to a fixed-length vector
{ ( 1),  ( 2), … ,  (  )} based on the co-occurrence relationship between documents
and words. The document vector v(  ) is then calculated by plusing each word vector
as follows:
 (  ) = ∑ =1  (  )</p>
        <p>(  ) = ∑ =1  (  )
Where  represents the number of words in the document.</p>
        <p>The author vector v(  ) is then computed by plusing each document vector according
to the co-occurrence relationships between documents and authors as follows:
Where  is the number of documents written by the author.</p>
        <p>With the fixed-dimension feature vectors of the research interests generated, the next
step is to calculate the similarity of interests between researchers. Of the many methods
of measuring similarity, we chose the popular and widely-used cosine similarity index,
formulated as
(1)
(2)
(3)</p>
        <p>Where the vector of author A is (a1, a2, …, am), and the vector of author B is (b1,
b2, …, bm).
3.3</p>
      </sec>
      <sec id="sec-3-2">
        <title>Topological similarity</title>
        <p>
          Node2Vec has been proven to maximize the likelihood of preserving network
neighborhoods and also can maps the nodes to a low-dimensional feature space[
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. To
efficiently and effectively acquire feature of scholars’ network topology, we take a
coauthorship network 
the  ∗  matrix 
= ( ,  ) as input, after running the node2vec model, we can get
to represent his/her network topology feature, where n regards
the number of nodes,  is the parameter that determines the dimension of the node’s
vector representation, and the final output is  
= { 1,  2, … ,   } .
        </p>
        <p>Calculating the cosine similarity between the topologic features of each scholar
follows the same basic principles as with the research interests described in Eq. (3).
3.4</p>
      </sec>
      <sec id="sec-3-3">
        <title>Recommendations with CombMNZ</title>
        <p>
          The goal in this stage is to integrate the two similarities and rank the candidate collab
orators from high to low according to their similarity. We have opted for a score-based
algorithm because it is the most widely used in the field of recommendation and, more
specifically, CombMNZ[
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>
          To fuse the similarity results with CombMNZ in a fair way, the dimensions of each
similarity measure first need to be standardized. The CombMNZ calculation is then[
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]:
  =  ( ,  ) ∗ ∑ =1   ∗   ( ,   ) (4)
where  ( ,  ) denotes the number of times scholar  appears in the score  of each
dimension,   ( ,   ) denotes the standardized score of the scholar in the  
item (  ≤2), and   denotes the weight of each dimension derived with a greedy
strategy.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>CASE STUDY</title>
      <sec id="sec-4-1">
        <title>Data collection and preprocessing</title>
        <p>
          To assemble our corpus, we retrieved papers published between 2010 and 2018 from
the Web of Science database using a search strategy drawn from Kwon et al. as
follows[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]:
TS=(((Self-driving or autonomous or driverless) near/4 (transport* or car or motorcar
or vehicle or automobile or aircraft or airplane or aeroplane))) or TS = (((drone near/2
autonomous) or (uav near/4 autonomous))) or TS = ((robot* near/1 (transport* or
mobile or car or motorcar or vehicle or automobile or aircraft or airplane or aeroplane))
AND (autonomous or self-driving or driverless)) or TS = (“autonomous driv*”) or TS
= (((robot* near/1 (transport* or mobile or car or motorcar or vehicle or automobile or
aircraft or airplane or aeroplane)) OR (drone or uav)) AND (path or planning or planner
or plan)) or TS = (((robot* near/1 (transport* or mobile or car or motorcar or vehicle or
automobile or aircraft or airplane or aeroplane)) OR (drone or uav)) AND (2D or 2-D
or 3D or 3-D or map or localization or tracking or navigat* or obstacle or avoid*)).
        </p>
        <p>The search returned 34,244 records. NLP preprocessing with VantagePoint yielded
8,637 core terms and phrases. Outstanding scientists were defined as those who had
published  or more papers – a criteria put forward by Price. Formally, the calculation
is  = 0.749(  )1/2, where   is the highest number of papers published by any
author in the dataset. Thus, we selected 813 researchers with five or more publications
for future analysis.</p>
        <p>From sections 3.2 to 3.4, we selected data from 2010 to 2013 to complete the
remaining three main steps of the recommendation framework. To further verify the quality
of recommendations, from Section 4.5, we divided the dataset into records from 2010
to 2013 as the training set and 2014 to 2018 as the testing set, and make a comparison
with the two most commonly used baselines: RWR and LDA.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Constructing the network of research interests with Word2Vec</title>
        <p>In this section, first, the parameters of the Word2Vec model were set to a window
size of 2 and a layer size of 128 based on the testing of a number of options. We then
generated the research interest vectors for all scholars according to Eq. (1) and Eq. (2).
Eq. (3) subsequently gave us an 813 × 813 symmetrical similarity matrix of research
interests. Alexey Matveev and Andrey Savkin had the most similar interests at
0.981981, Senqiang Zhu and Frederic Py had the least similar at 0.002712. The mean
value, median and standard deviation values were 0.544555, 0.580275 and 0.209891,
respectively.</p>
        <p>Fig. 2 shows the network scholars with a similarity score of more than 0.55. The size
of the node represents the number of published papers for each scholar, and the
thickness of the lines indicates the degree of similarity between the scholars’ research
interests.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Constructing the co-authorship network with Node2Vec</title>
        <p>In this section, first, the parameter settings for the Node2Vec model were also
determined from testing. The final values were: dimensions=128; walk-length=80; p=1; and
q=1. Eq. (3) yielded the 813 ×813 topology matrix of cosine similarity between
scholars, which was again symmetrical. Gaurav S Sukhatme and Ryan N Smith shared the
greatest similarity (0.957244), and Jian Liu and Yi Chao had the least (0.028196). The
mean, median and standard deviation values were 0.416515, 0.414858, and 0.125672,
respectively.</p>
        <p>Fig. 3 shows the co-authorship network based on scholars with a topological
similarity of more than 0.47. The size of nodes indicates the number of collaborators
associated with that scholar. The thickness of the lines represents the degree of similarity
between the two connected scholars.</p>
      </sec>
      <sec id="sec-4-4">
        <title>Ranking the candidate collaborators with CombMNZ</title>
        <p>As a preliminary assessment of the framework’s ability to make appropriate
recommendations, we randomly selected Roland Siegwart as the target scholar and generated
a final list of recommendations. The top-10 ranked candidates are shown in Table 1.</p>
        <p>An in-depth manual review of Siegwart’s academic background shows these
recommendations to be appropriate. For example, Stachniss and Siegwart have overlapping
interests in mobile robots, sensor design, navigation system design, positioning, motion
planning and more, and have both published many influential papers. In addition, both
scholars often attend the IEEE International Conference on Robotics &amp; Automation.
Based on this analysis, it is reasonable to conclude that the framework can recommend
realistic and fruitful collaborations.
4.5</p>
      </sec>
      <sec id="sec-4-5">
        <title>Comparative evaluations</title>
        <p>From the literature, we found the two most popular and widely-used methods for
collaborator recommendations are the LDA model and the RWR indicator coupled with
a topological model. To compare the quality of recommendations produced by these
approaches with those of our framework, we divided the dataset into records from 2010
to 2013 as the training set and 2014 to 2018 as the testing set. And then we randomly
chose 20 target authors for comparison by Precision, Recall, and F1 scores. The results
are given in Figs. 4 to 6.</p>
        <p>In Figs. 4 to 6, we find that our framework makes higher precision, recall and f1
score than the two current benchmark solutions, these results emphasize the advantages
of our approach. More specifically, we drew the following insights from this analysis:</p>
        <p>(1) Fusing similarity indicators based on research interests and topological
structures significantly increased the quality of the recommendations.</p>
        <p>(2) The Word2Vec method solved the problems of context less and scalability
associated with traditional text mining technology.</p>
        <p>(3) The Node2Vec method removed the need to manually design and define
indicators, saving on manpower and produced recommendations based on global network
features.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>CONCLUSION</title>
      <p>Overall, the main innovation of our paper is to develop a novel framework for
recommending academic collaborators with similar research interests and network
topology features through incorporating word embedding and network embedding, and
produces more accurate recommendation compared with the existing methods. Meanwhile,
the framework can not only help researchers and private-enterprise R&amp;D managers to
provide valuable reference for cooperation, but also is feasible and can now be the basis
for further improvement/inspiration.</p>
      <p>The limitations of our current research offer opportunities for future inquiry. These
are summarized as follows. (1) Word embedding and network embedding techniques
both contain some parameters; however, methods of training these parameters for
optimal benefit is a task that falls into the field of machine learning. (2) We have based our
recommendations on only two criteria: the similarity of research interests and the
coauthorship features. However, other factors can also indicate the likelihood of a good
collaboration, such as citations, or institutional ties. In future, we will consider adding
more of these factors into our framework.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Sylvan</surname>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>Ben R.</given-names>
          </string-name>
          : What is research collaboration?.
          <source>Research Policy</source>
          <volume>26</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          (
          <year>1997</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Sooho</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Barry</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Scientific collaboration || the impact of research collaboration on scientific productivity</article-title>
          .
          <source>Social Studies of Science</source>
          <volume>35</volume>
          (
          <issue>5</issue>
          ),
          <fpage>673</fpage>
          -
          <lpage>702</lpage>
          (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Guns</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Rousseau</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Recommending research collaborations using link prediction and random forest classifiers</article-title>
          .
          <source>Scientometrics</source>
          <volume>101</volume>
          ,
          <fpage>1461</fpage>
          -
          <lpage>1473</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Abramo</surname>
          </string-name>
          , G.,
          <string-name>
            <surname>D'Angelo</surname>
            ,
            <given-names>C. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Costa</surname>
            ,
            <given-names>F. D.</given-names>
          </string-name>
          :
          <article-title>Research collaboration and productivity: is there correlation?</article-title>
          <source>Higher Education</source>
          <volume>57</volume>
          (
          <issue>2</issue>
          ),
          <fpage>155</fpage>
          -
          <lpage>171</lpage>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Kong</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            <given-names>W.</given-names>
          </string-name>
          et al.:
          <article-title>Exploring dynamic research interest and academic influence for scientific collaborator recommendation</article-title>
          .
          <source>Scientometrics</source>
          <volume>113</volume>
          (
          <issue>1</issue>
          ),
          <fpage>369</fpage>
          -
          <lpage>385</lpage>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>L. T.</given-names>
          </string-name>
          :
          <article-title>Mvcwalker: Random walk-based most valuable collaborators recommendation exploiting academic factors</article-title>
          .
          <source>IEEE Transactions on Emerging Topics in Computing</source>
          <volume>2</volume>
          (
          <issue>3</issue>
          ),
          <fpage>364</fpage>
          -
          <lpage>375</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Pham</surname>
            ,
            <given-names>M. C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klamma</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Jarke</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A clustering approach for collaborative filtering recommendation using social network analysis</article-title>
          .
          <source>Journal of Universal Computer Science</source>
          <volume>17</volume>
          (
          <issue>4</issue>
          ),
          <fpage>583</fpage>
          -
          <lpage>604</lpage>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Dong</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , et al.:
          <article-title>Link Prediction and Recommendation across Heterogeneous Social Networks</article-title>
          .
          <source>2012 IEEE 12th International Conference on Data Mining. IEEE</source>
          . pp.
          <fpage>181</fpage>
          -
          <lpage>190</lpage>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ping</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>De-Gen</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <article-title>Tf-idf and rules based automatic extraction of chinese keywords</article-title>
          .
          <source>Journal of chinese computer systems</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M. I.</given-names>
          </string-name>
          <article-title>Latent Dirichlet Allocation</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>3</volume>
          (
          <issue>4-5</issue>
          ),
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <article-title>Distributed representations of sentences and documents</article-title>
          .
          <source>Proceedings of the 31st International Conference on Machine Learning (ICML-14)</source>
          ,
          <fpage>1188</fpage>
          -
          <lpage>1196</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <article-title>Measuring multiple evolution mechanisms of complex networks</article-title>
          .
          <source>Scientific Reports</source>
          ,
          <volume>5</volume>
          ,
          <issue>10350</issue>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Kong</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            <given-names>W.</given-names>
          </string-name>
          et al.
          <article-title>Exploring dynamic research interest and academic influence for scientific collaborator recommendation</article-title>
          .
          <source>Scientometrics</source>
          <volume>113</volume>
          (
          <issue>1</issue>
          ),
          <fpage>369</fpage>
          -
          <lpage>385</lpage>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jian</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Wenwu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>A survey on network embedding</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          , pp (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Porter</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Newman</surname>
          </string-name>
          , N.C.
          <article-title>: “Term clumping” for technical intelligence: a case study on dye-sensitized solar cells</article-title>
          .
          <source>Technol. Forecast. Soc. Chang</source>
          <volume>85</volume>
          ,
          <fpage>26</fpage>
          -
          <lpage>39</lpage>
          (
          <year>2014a</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G. S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>Advances in Neural Information Processing Systems</source>
          ,
          <volume>3111</volume>
          -
          <fpage>3119</fpage>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Eunice</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iris</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Humphrey</surname>
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Yiu-Kai N</surname>
          </string-name>
          .
          <article-title>Making personalized movie recommendations for children</article-title>
          .
          <source>International Conference on Information Integration &amp; Web-based Applications &amp; Services. ACM</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ounis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Voting techniques for expert search</article-title>
          .
          <source>Knowledge &amp; Information Systems</source>
          ,
          <volume>16</volume>
          (
          <issue>3</issue>
          ):
          <fpage>259</fpage>
          -
          <lpage>280</lpage>
          (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Kwon</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Porter</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Youtie</surname>
          </string-name>
          , J.:
          <article-title>Research addressing emerging ideas has greater scientific impact</article-title>
          .
          <source>Research Policy</source>
          <volume>48</volume>
          (
          <issue>9</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>