<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Method of Ontology-aided Expertise Matching for Facilitating Knowledge Exchange</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Eduard Babkin</string-name>
          <email>eababkin@hse.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolay Karpov</string-name>
          <email>nkarpov@hse.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tatiana Babkina</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Research University Higher School of Economics</institution>
          ,
          <addr-line>25/12 Bolshaja Pecherskaja Ulitsa, Nizhny Novgorod 603155</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>43</fpage>
      <lpage>54</lpage>
      <abstract>
        <p>The paper proposes a new method for facilitating knowledge exchange by seeking relevant university experts for commenting actual information events arising in the open environment of a modern economical cluster. This method is based on a new mathematical model of ontology concepts matching. We propose to use in the formal core of our method a new modification of Latent Dirichlet allocation. The method and the mathematical model of ontology matching were validated in the form of a software-based solution: the newly designed decision support system titled EXPERTIZE. The system regularly monitors different text sources in the Internet, performs document analysis and provides university employees with critical information about relevant events according a developed matching algorithm. In the proposed solution we made several contributions to the advances of knowledge processing, including: new modifications of topic modeling method suitable for application in expert finding tasks, integration of new algorithms and existing ontology services to show feasibility of the solution.</p>
      </abstract>
      <kwd-group>
        <kwd>expert finding</kwd>
        <kwd>natural language processing</kwd>
        <kwd>topic modeling</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Emerging and successful growing of new forms of inter-organizational cooperation
known as regional, innovation or university clusters [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] in national economies became
a significant phenomenon of the modern world-wide socio-economical system.
Sustainable exchange of expertise and professional knowledge between stakeholders of
innovation clusters plays an important role in knowledge-based economics [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. For this
task an university undoubtedly should be a catalyst which provides expert evaluation
and opinions. Critical problems and major strategic choices should be commented,
discussed and exposed for multiple stakeholders including industry mass-media and
society.
      </p>
      <p>Until now there is no big success of tight integration of university community
within the framework of emerging innovative clusters. Informational links are
developed by ad hoc manner, major activities are implemented inside the stable
university-based structures like incubators and business parks. Communication with
business experts and mass media shows that in modern turbulent information
environments it is the paradigm of information and knowledge exchange which should
be modernized. The modernized paradigm of information and knowledge exchange
should facilitate reactive or even proactive behavior of university community in
response to critical emerging economic or social phenomena in the open environment
of innovative cluster-based economy of knowledge.</p>
      <p>Traditional analytical methods which provide modern university community with
current information about important discussion topics and critical issues lack of
comprehensiveness and become too slow. In nowadays practice of universities the best
solutions primarily include manual analysis of mass media and internet resources and
further slow distribution of information about relevant public events through the
inefficient hierarchical organizational structure (from the schools, faculties towards
department and employees).</p>
      <p>We believe that advanced methods of automated and automatic knowledge
management belong to critical scientific foundations of modernization the paradigm of
information and knowledge exchange. A specifically designed combination of
automated text processing and ontology-based knowledge engineering may improve
quality of information analysis and reduce university’s response time.</p>
      <p>
        There are many interesting systems which approaches are close to our knowledge
exchange idea. The one of it is Media Information Logistics project (Media-ILOG)
which is concerns the domain of mass media too. The goal of the Media-ILOG [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
was to improve information flow inside a local newspaper JonkopingsPosten.
      </p>
      <p>In our research we limited the scope of the aforementioned global problem to the
key issue of real-time matching between relevant university experts and actual
information events arising in the open environment of the economical innovation
cluster. We offer a solution of that issue in the form of new automated method of
experts finding for facilitating knowledge exchange between the university and
heterogeneous community of the innovation cluster.</p>
      <p>
        In contrast to Media-ILOG which is used semantic matching approach proposed by
Billig et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] The core of our method is a modification of Latent Dirichlet allocation.
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] It is algorithmically implemented in the newly designed decision support system
titled EXPERTIZE. The system regularly monitors different text sources in the
Internet, performs document analysis and provide university employees with critical
information about relevant events according the specific relevance matching algorithm.
      </p>
      <p>The high level design structure of EXPERTIZE software system includes several
principal components. They are Crawler, Data Modeler, Data Store, GUI and Matcher.
We match an input document not only with a single expert from our dataset, but with a
scientific areas of interest, which is a category of the formal ontology. Each category is
represented as a probability distribution of latent topics, so we match distribution of
latent topics in the query document with the category using the maximum-likelihood
estimation.</p>
      <p>
        In the result of software implementation EXPERTIZE software system has been
implemented as a software service. Now it is in an operating state, and regularly
collects data from the several information resources available in Internet: library of
HSE9 and Elibrary10. Open systems interfaces allow EXPERTIZE get real-time access
to the areas of domain interest of the employees of HSE from the InfoPort service [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <sec id="sec-1-1">
        <title>9 publications.hse.ru 10 elibrary.ru 44</title>
        <p>A set of practical use cases show that EXPERTIZE properly matches the actual
information about discussion topics and information events.</p>
        <p>The article has the following structure. After the introduction Section 2 contains
related works overview in the information modeling and semantic matching to experts’
domains. In Section 3 we observe essentials and formal foundations of our method.
Main design decisions and functionality of EXPERTIZE software system are described
in Section 4. Section 5 provides the readers with case study of application of that
system in a real life information environment. Section 6 concludes the work, giving
comparison results of our method and other known approaches and defining open
research questions for further investigation.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2 Overview of Relevant Formal Methods for Expert Finding</title>
      <p>
        As soon as our task is to match ontology concepts of expertise with plain text of news
it is strongly related to the common expertise retrieval task. The past decade has
appeared tremendous interest in expertise retrieval as an emerging subdiscipline. From
2005 the Enterprise Track at the Text REtrieval Conference (TREC) provided a
common platform for researchers to empirically assess methods and techniques
devised for expert finding [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The TREC Enterprise test collections are based on
public facing web pages of large knowledge-intensive organizations, such as the World
Wide Web Consortium (W3C) and the Commonwealth Scientific and Industrial
Research Organisation (CSIRO).
      </p>
      <p>
        Balog et al. 2012 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] highlights state of the art models and algorithms relevant to
this field. They classified expert finding approaches as follows:
• profile-based model;
• document-based model;
• hybrid model.
      </p>
      <p>
        A profile-based model for expert finding using information retrieval proposed in
Balog and de Rijke [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. A candidate’s skill is represented as a score over documents
that are relevant given a knowledge area. The relevance of a document is estimated
using standard generative language model techniques.
      </p>
      <p>
        In the other approach, the method of document-based expert finding does not create
a profile for each expert. It uses documents to match candidates to queries. The idea is
to first find documents that are relevant to the topic and then locate the experts
associated with these documents. The document models are also referred to as
querydependent approaches. Later, Fang and Zhai [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] presented a general probabilistic
model for expert finding and showed how the document-based model can be adapted
in this schema.
      </p>
      <p>
        Balog et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] applied this approach to a language model–based framework for
expert finding. They also used the profilebased approach in their system and showed
that the document-based approach performs better than the profile-based model.
Serdyukov and Hiemestra [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] proposed a hybrid model for expert finding which
combines both profile- and document-based approaches.
      </p>
      <p>
        Semantic analysis of texts for expert finding with required competencies proposed
by Fomichov on the basis of Formal Concept Analysis [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The approach allows to
build and compare semantic representations of expert profile using the theory of
Krepresentation and a model of linguistic database.
      </p>
      <p>
        A topic modeling approach for expert finding proposed by Balog et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Instead
of modeling candidate profiles or documents, they built a model for each input query
and used this model to calculate the probability of candidates given queries. Their
approach is similar to the document likelihood method, which is used in language
model–based information retrieval. Based on their results, this model underperforms
the profile- and document based approaches. The main reason of its poor performance
is the sparsity of the models built from the queries. Their definition of topic, however,
is different from ours. The term topic in their work refers to query words that users use
to search for experts, whereas in the present work we use the term topic as a set of
concepts that are extracted from a collection using a topic modeling algorithm. There
are multiple known methods for topic modeling of document which are Latent
Semantic Analysis (LSA) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] Latent Dirichlet allocation (LDA) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] et al.
      </p>
      <p>The topic modeling approach is based on the assumption that words in a document
are independent of one another (bag of words) and of their order in the text. Similarly,
documents in a corpus D are independent of one another and unordered. Distribution of
words W is determined by the set of latent topics Z . Each topic has its own word
distribution (phi) and each document has distribution over topics (theta).</p>
      <p>
        Traditional topic-based information retrieval approach is exploited by Wei and
Croft, 2006 [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. The extracted topics are used for information retrieval; whereas the
to-be-retrieved documents are used in the retrieval step, i.e., the distribution of topics
over words (phi) is used for estimating P(Q / Z) , where Q – is a set of word in query.
The distribution of documents over topics (theta) is used for estimating P(Z / D) .
      </p>
      <p>
        Another topic-based model is proposed by Momtazi and Naumann [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. This model
outperforms the state-of-the-art profile- and document-based models. To-be-retrieved
documents are not used in the retrieval step. Instead, we only use these documents for
training LDA, i.e., to be-retrieved documents are used as a corpus to extract topics in
an off-line process. Then, in the retrieval step, we only use the distribution of topics
over words (phi) for estimating both P(Q / Z) and P(e / Z) where e – is an expert
label.
      </p>
      <p>
        In a paper [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] the researchers show how to use a topic-based model with scientific
ontology, where each document labeled with a category in scientific classification
taxonomy C. They represent each category c as a conditional probabilistic distribution
P(Z / c) which denotes the probability of category c being labeled with topic z. By
utilizing LDA, P(Z / c) is a Z -dimension vector of topic distribution. The main
requirement for this approach is to estimate the probability P(zk / c) , which cannot be
obtained directly from LDA. However, according to the Bayes formula authors
calculate P(zk / c) by
      </p>
      <p>P(zk / c) =</p>
      <p>P(c / zk )P(zk )
∑ P(c, zk )
k
(1)
where P(c / zk ) and P(zk ) can be obtained from LDA. As soon as ∑ P(c, zk ) is
k
constant for different c and P(zk ) is uniform distribution we have</p>
      <p>P(Z / c) ∝ P(c / Z)
(2)</p>
      <p>
        On the basis of explored papers the best way to solve our task is to match between
relevant university experts and actual information events using topic-based model,
which is proposed by Momtazi and Naumann [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Thus, we should implement the
model for papers in Russian language concluded in our enterprise dataset. With the
help of approach described in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] this topic-based approach can be applied to use with
scientific ontology. To show feasibility of the solution, we archive an integration of
new algorithms and existing ontology services.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3 The Essentials and Formal Foundations of the Method Proposed</title>
      <p>
        In our previously designed InfoPort system [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] for solving the expert finding
problem we proposed to translate a user-specified query to a corresponding SPARQL
query which is evaluated against a specific set of RDF repositories. The query result
consisted of a relevant category of scientific classification taxonomies and keywords.
The search algorithm of InfoPort system retrieved all persons who labeled with this
query.
      </p>
      <p>In the current research our new system EXPERTIZE works automatically: it gets
news event as a query and matches it to the most relevant scientist, who can provide
expert evaluation and opinions about it. In other word we arrange experts in order to
relevance to the event.</p>
      <p>On the one hand news events are represented as news in natural language format,
thus we have ability to extract semantic information from the text. On the other hand
each expert has texts in the form of written papers or records of spoken interviews and
tutorials. This material contains rich semantic information about personal interests and
abilities.</p>
      <p>
        There are some formal models suitable for implementation of context analysis such
as a Distributional Semantic Model (DSM) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ][
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and Latent Semantic Analysis
[
        <xref ref-type="bibr" rid="ref20">20</xref>
        ][
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and Latent Dirichlet Allocation [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>In our project we use an extension of Latent Dirichlet Allocation which is a
generative formal model that uses latent groups to explain results of observations –
data similarity in particular. For instance, if the observations are words in the
documents, one can posit that each document is a combination of a small number of
topics and that each word in the document is connected with one of the topics. Latent
Dirichlet Allocation (LDA) is one of topic-modeling methods and was first introduced
by its authors as a graphical model for topic detection.</p>
      <p>
        In our approach by training the LDA model, we form the statistical portrait of its
author. A person writing a text has a set of topics in their mind, and each document has
a certain distribution of these topics. The author first selects the topic to write on;
within this topic, there is a distribution of words that may occur in any document that
contains this topic. The next word in the text is generated within the distribution. Then
the same procedure is repeated. On each iteration, the author either selects a new topic
or continues to use the previous one, and generates the next word within the active
topic [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>The first step of our method for expert finding is a training the model on a collection
of texts. We get an estimate of two discrete distribution functions. The following is
distribution of probabilities of words W in topics Z :</p>
      <sec id="sec-3-1">
        <title>Distribution of probabilities of topics Z in documents D :</title>
        <p>P(wi / zk ); i ∈1, W ,k ∈1, Z</p>
        <p>P(zk / dn ); k ∈1, Z ,n ∈1, D</p>
        <p>Semantic representation of query news document d0 can be also calculated using
built LDA model. It is distribution of probabilities of topics Z in documents d0 –</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Software Design of EXPERTIZE</title>
      <p>The described method for experts finding was practically implemented during
design and implementation of the system for matching between relevant university
experts and actual information events arising in the open environment of the
economical cluster. Such system was called EXPERTIZE. The following services are
distinguished in the high-level design of that system (Fig.2):
1. Web Crawler;
2. Data Modeler;</p>
      <sec id="sec-4-1">
        <title>3. Data Store;</title>
        <p>4. Graphical User Interface (GUI);
5. Matcher.</p>
        <p>
          EXPETIZE actively uses our InfoPort Service [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. That semantic service provides in
the form of formal ontology factual information about more than three hundred
employees of Higher School of Economics (HSE NRU)11 branch at Nizhny Novgorod.
The InfoPort data is represented as RDF triples. Triples include hierarchical
information as it originally is in the source. The first level is an alphabetical ordered
list of group of scientist, second is a scientist with his personal interests and papers,
and third is papers with its features.
        </p>
        <p>The components of the EXPETIZE system can be classified as Online and Offline
services. Both are interacted with InfoPort via native REST interface. Offline ones
work within monthly period to update information regularly. Online services work on
demand, when user activates it by web interface.</p>
      </sec>
      <sec id="sec-4-2">
        <title>RInefgouPloarrt</title>
        <p>offline
services</p>
      </sec>
      <sec id="sec-4-3">
        <title>EXPETIZE system</title>
      </sec>
      <sec id="sec-4-4">
        <title>Online services</title>
        <p>Native
RESTInterface</p>
      </sec>
      <sec id="sec-4-5">
        <title>InfoPort platform</title>
      </sec>
      <sec id="sec-4-6">
        <title>Store</title>
        <p>Service
• build LDA model with a given number of topics K.</p>
        <p>
          At present time, there are several methods for building LDA models, that is,
methods of searching for parameters of all distribution functions in the model. All of
the methods are iteration-based and are similar in structure to the Expectation
Maximization (EM) algorithm. They are:
- Online Variational Bayes algorithm [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ];
- Gibbs Sampling [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ];
- Expectation Propagation [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
        </p>
        <p>
          Among these algorithms, we use the Online Variational Bayes algorithm as it is the
most precise one [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. It is well implemented in the Gensim15 toolkit.
resourceintensive algorithm.
        </p>
      </sec>
      <sec id="sec-4-7">
        <title>InfoPort Store Service</title>
        <p>RESTInterface</p>
      </sec>
      <sec id="sec-4-8">
        <title>Crawler Service</title>
        <p>REST-Interface
Web GUI</p>
        <p>Offline
processing</p>
      </sec>
      <sec id="sec-4-9">
        <title>Data</title>
        <p>Modeler</p>
      </sec>
      <sec id="sec-4-10">
        <title>Matcher</title>
        <p>RSS
Newsfeed</p>
        <p>Online
processing</p>
      </sec>
      <sec id="sec-4-11">
        <title>Temporal raw data</title>
      </sec>
      <sec id="sec-4-12">
        <title>LDA model</title>
      </sec>
      <sec id="sec-4-13">
        <title>Data Store</title>
        <p>Online processing performs on demand of user by opening Web GUI. Web interface
activates RSS Newsfeed, which gets and displays 10 last news from the RSS feed and
an empty textbox. User can choose one of its 10 news or paste the text to the textbox
manually. When user specify input query the GUI transfer it to the Matcher. In turn,
this component performs online semantic search. A semantic representation of event is
matched with semantic representations of scientific categories and experts by applying
the formula (8) and (9) and selecting top 5 of the units. So, the Mather component
returns 5 URIs to the GUI.</p>
        <p>To provide user friendly output of the finding result GUI component makes a
request to the Infoport Service. It gets features of the selected units: full name, expert’s
photo URL, expert’s department.
15 http://radimrehurek.com/gensim</p>
        <p>The crawled collection includes 4132 units but only 1492 papers are in the Russian
language. So, we decide to extend collection with the help of eLibrary16 scientific
database. This is a biggest scientific database in the Russian language. We extracted a
part of this base connected only with Information Technologies field. It includes 9127
papers not older than 2011 year.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Case study</title>
      <p>Evaluation of our proposed method and the EXPERTIZE system was performed
empirically. We choose experts from our pool. This pool includes more than three
hundreds of professors and researchers of the HSE NRU branch at Nizhny Novgorod17.
According to an experts field of the study we chose news, which one can be comment
by expert and put it to EXPERTIZE. If this expert appears in the list, proposed by the
system, we mark such attempt as a successful match.</p>
      <p>Let’s take a case study. There is an expert Sidorov Dmitry V. whose profile includes
a set of scientific domain topics, which he is interested in. There are:
• w1 - innovation projects,
• w2 - venture investments,
• w3 - innovative potential estimation
• etc.</p>
      <p>Each scientific domain topics coded as one word and we have pre-created table
which is distribution of probabilities of words W in latent topics Z :
P(wi / zk ); i ∈1, W ,k ∈1, Z . It usually has small number of elements higher than zero.</p>
      <p>z1 z2 … z58 … z200
w1 0 0.04 0.1 0</p>
      <p>We find news with title «Yandex company pays for big data»18, which he can be
able to comment as an expert. This news is about investment of Russian IT giant to an
Israeli startup company. As each other documents in the collection it can be found
probabilities distribution of latent topics z in document. This news goes as an input to
the Matcher component where it converts to the probability distribution over latent
topics (4) using the pre-built LDA model. The number of topics we set equal to 200.
Table.2. Example of probabilities distribution of topics Z in documents d0 –</p>
      <p>Next, using formula (10) the algorithm chooses each categories c from scientific
classification taxonomy and finds P(d0 / C) . Top 5 of experts who has maximum
P(d0 / C) is shown in the system. A result is presented in Fig. 3.</p>
      <sec id="sec-5-1">
        <title>Document-based</title>
        <p>Candidate-based
Topic-based</p>
        <p>We choose the Topic-base approach because in comparison with other approaches
(Document-based and Candidate-based) this one gets the best results.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6 Conclusion</title>
      <p>
        In this article we presented a new approach to support rapid exchange of knowledge
in innovation clusters based on reactive experts finding. The proposed method of
expert finding uses open Internet resources and existing ontological services like
InfoPort [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] to get access to the approved skills of potential experts.
      </p>
      <p>During our research we developed a new formal method based on Latent Dirichlet
allocation, which includes a software-based solution for matching between relevant
university experts and actual information events arising in the open environment of the
economical cluster. This solution allows performing real-time matching between
Internet news and areas of interest of university employees with further quick
notification about possible participation of relevant employees in interviews,
informational programs and discussions. In the proposed solution we made several
contributions to the advances of knowledge processing, including: new modifications
of topic modeling method suitable for application in expert finding tasks, integration of
new algorithms and existing ontology services to show feasibility of the solution.</p>
      <p>A software design of decision support system EXPERTIZE was developed for
practical application of the method proposed. The first use cases of the EXPERTIZE
system show their relevance and ability to solve the task specified.</p>
      <p>
        Using topic-based model proposed by Momtazi and Naumann [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] we have
achieved about 0.43 amount of mean average precision (MAP) on our own queries.
The same approach on TREC 2005 and 2006 queries, shows 0.248 and 0.471 amount
of MAP respectively [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. So, precision of EXPERTIZE system is not much less than
achieved on TREC 2006. The estimation of recall and f-measure in our EXPERTIZE
system less interesting because in general user doesn’t need a full set of various
experts. One or two most relative experts usually enough for facilitating knowledge
exchange.
      </p>
      <p>As soon as we perform expert matching with scientific categories we can apply
cross-language expertise retrieval by applying multi-language scientific ontology. It
would be our prospective work.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Asheim</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cooke</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Clusters and regional development: Critical reflections and explorations</article-title>
          .
          <source>Econ. Geogr</source>
          .
          <volume>84</volume>
          ,
          <fpage>109</fpage>
          -
          <lpage>112</lpage>
          (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Klimova</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Litvintseva</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Universities Innovation Clusters: Approaches for National Competitiveness Paradigm, (</article-title>
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Sandkuhl</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Öhgren</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smirnov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shilov</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kashevnik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Ontology construction in practice: Experiences and recommendations from industrial cases</article-title>
          .
          <source>9th International Conference on Enterprise Information Systems</source>
          ,
          <volume>12</volume>
          -
          <fpage>16</fpage>
          ,
          <year>June 2007</year>
          , Funchal, Madeira-Portugal (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Billig</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blomqvist</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Semantic matching based on enterprise ontologies. On the Move to Meaningful Internet Systems 2007: CoopIS, DOA</article-title>
          , ODBASE, GADA, and IS. pp.
          <fpage>1161</fpage>
          -
          <lpage>1168</lpage>
          . Springer (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.: Latent</given-names>
          </string-name>
          <string-name>
            <surname>Dirichlet</surname>
            <given-names>Allocation</given-names>
          </string-name>
          , (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Babkin</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karpov</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kozyrev</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Towards Creating an Evolvable Semantic Platform for Formation of Research Teams</article-title>
          . In: Kobyliński,
          <string-name>
            <given-names>A.</given-names>
            and
            <surname>Sobczak</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . (eds.) Perspectives in Business Informatics Research. pp.
          <fpage>200</fpage>
          -
          <lpage>213</lpage>
          . Springer Berlin Heidelberg (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Craswell</surname>
          </string-name>
          , N.,
          <string-name>
            <surname>de Vries</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Soboroff</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Overview of the TREC 2005 Enterprise Track</article-title>
          . Trec. pp.
          <fpage>199</fpage>
          -
          <lpage>205</lpage>
          (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fang</surname>
          </string-name>
          , Y.,
          <string-name>
            <surname>de Rijke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Serdyukov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Si</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , others: Expertise Retrieval, (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Rijke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Determining Expert Profiles (With an Application to Expert Finding)</article-title>
          .
          <source>IJCAI</source>
          . pp.
          <fpage>2657</fpage>
          -
          <lpage>2662</lpage>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Fang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhai</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Probabilistic models for expert finding</article-title>
          .
          <source>Advances in Information Retrieval</source>
          . pp.
          <fpage>418</fpage>
          -
          <lpage>430</lpage>
          . Springer (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Serdyukov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hiemstra</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Modeling documents as mixtures of persons for expert finding</article-title>
          .
          <source>Advances in Information Retrieval</source>
          . pp.
          <fpage>309</fpage>
          -
          <lpage>320</lpage>
          . Springer (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Fomichov</surname>
          </string-name>
          , V.: Semantics-Oriented
          <source>Natural Language Processing: Mathematical Models and Algorithms</source>
          . Springer (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bogers</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Azzopardi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Rijke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Van Den Bosch</surname>
          </string-name>
          , A.:
          <article-title>Broad expertise retrieval in sparse data environments</article-title>
          .
          <source>Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          . pp.
          <fpage>551</fpage>
          -
          <lpage>558</lpage>
          . ACM (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Dumais</surname>
          </string-name>
          , S.T.:
          <article-title>Latent semantic analysis</article-title>
          .
          <source>Annu. Rev. Inf. Sci. Technol</source>
          .
          <volume>38</volume>
          ,
          <fpage>188</fpage>
          -
          <lpage>230</lpage>
          (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Croft</surname>
          </string-name>
          , W.B.:
          <article-title>LDA-based document models for ad-hoc retrieval</article-title>
          .
          <source>Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          . pp.
          <fpage>178</fpage>
          -
          <lpage>185</lpage>
          . ACM (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Momtazi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naumann</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Topic modeling for expert finding using latent Dirichlet allocation</article-title>
          , (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiong</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tian</surname>
          </string-name>
          , J.:
          <article-title>Towards expert finding by leveraging relevant categories in authority ranking</article-title>
          .
          <source>Proceedings of the 20th ACM international conference on Information and knowledge management</source>
          . pp.
          <fpage>2221</fpage>
          -
          <lpage>2224</lpage>
          . ACM (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Baroni</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Distributional memory: A general framework for corpusbased semantics</article-title>
          , (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Turney</surname>
          </string-name>
          , P.D.:
          <article-title>Similarity of semantic relations</article-title>
          , (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Landauer</surname>
            ,
            <given-names>T.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Foltz</surname>
            ,
            <given-names>P.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Laham</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>An introduction to latent semantic analysis</article-title>
          .
          <source>Discourse Process</source>
          .
          <volume>25</volume>
          ,
          <fpage>259</fpage>
          -
          <lpage>284</lpage>
          (
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>David</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Blei</surname>
          </string-name>
          , Matthew D. Hoffman:
          <article-title>Online Learning for Latent Dirichlet Allocation</article-title>
          , http://books.nips.cc/papers/files/nips23/NIPS2010_1291.pdf, (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Thomas L. Griffiths</surname>
          </string-name>
          , Mark Steyvers:
          <article-title>Finding scientific topics</article-title>
          , (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Thomas</surname>
            <given-names>Minka</given-names>
          </string-name>
          , John Lafferty:
          <article-title>Expectation-propagation for the generative aspect model</article-title>
          .
          <source>Proceedings of the Eighteenth conference on Uncertainty in artificial intelligence</source>
          . pp.
          <fpage>352</fpage>
          -
          <lpage>359</lpage>
          . Morgan Kaufmann Publishers Inc. (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>