<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Articles Available for the Journal
Number of Requests Lower Bound Upper Bound
5</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Search for an Appropriate Journal - in Depth Evaluation of Data Fusion Techniques</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Markus Wegmann</string-name>
          <email>markus.wegmann@live.de</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas Henrich</string-name>
          <email>andreas.henrich@uni-bamberg.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Media Informatics Group, University of Bamberg</institution>
          ,
          <addr-line>Bamberg, Germany https://</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Based on available or generated metadata, collection objects can be categorized and organized in disjoint sets or classes. In our approach we search for appropriate classes of a categorized collection based on object instance queries. More concretely, our collection consists of scientific papers as object instances which are classified by the journals they are published in. Our textual query searches for the most appropriate classes respectively journals. At LWDA 2017 [1] we introduced a voting-based approach for searching these appropriate journals: Utilizing randomly removed articles from the article collection as query instances we searched for journals as classes, having potentially similar or related articles and topics. To evaluate the relevance, we determined the rank of the requested article's journal, assuming that our request, respectively article title, is a significant example for its journal. A complete automation of search and evaluation enables us to send a huge number of requests against the collection and to evaluate and fine tune the techniques. In this contribution we maintain our base approach of search and evaluation while adding search on abstracts, variations of similarity measures, new voting techniques, and evaluations considering our collection structure regarding the journal/article count distribution.</p>
      </abstract>
      <kwd-group>
        <kwd>Expert Search</kwd>
        <kwd>Expertise Retrieval</kwd>
        <kwd>IR Systems</kwd>
        <kwd>Collection Search</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Motivation and Related Work</title>
      <p>
        One of the approaches introduced in literature for expertise retrieval [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is based
on relevant documents retrieved by a search query. These documents vote for
their associated authors as candidate experts. In this work, we transfer this
approach to another domain: Our keyword search on a bibliographic collection
yields matching articles which vote for the journals where the articles have been
published, as beneficial sources or targets. Our aim is to identify a technique
which yields and ranks journals that potentially contain other publications and
resources which match the information need of the user. This information need
targets journals rather than single articles.
      </p>
      <p>In a first step, we don’t evaluate the discussed techniques using manually
created test data or test users. Instead, we use article-titles from the collection
itself to automatically send these titles as search requests. Since we have the
information in which journal a single article has been published, we can measure
the position of this respective journal in the result ranking and evaluate the
algorithms.</p>
      <p>
        This work is based on research in data fusion techniques and their application
in the field of expertise retrieval. Different approaches show, that combining
multiple retrieval results using voting models can improve retrieval effectiveness [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
In their survey, Balog et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] present different approaches used in expertise
retrieval including the document-based voting model. Rank- and score-based fusion
techniques are listed and evaluated, mostly based on the work of MacDonald et
al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Furthermore, normalization methods are applied for the underlying
candidate expert profiles to gain better results. In the mentioned works, it becomes
quite significant that the documents in the upper ranks together with their score
values have a disproportionately high impact on the quality of the fusion results;
exponential variants of fusion techniques can have better results and prove this
fact [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        In the paper at hand, we investigate how such approaches perform in our
setting. We present significant extensions to our previous paper at [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]: We maintain
our base approach of search and evaluation while adding search on abstracts,
variations of similarity measures, new voting techniques, and evaluations
considering our collection structure regarding the journal/article count distribution.
      </p>
      <p>
        It should be mentioned that existing journal recommenders from publishers
– like EndNote’s manuscript matcher, Elsevier’s journal finder, or Springer’s
journal suggester – are obviously related to our approach. However, these systems
apply complex ranking schemes using much more information than our simple
approach discussed in this paper. The aim of our paper is to investigate the
capability of rather simple voting techniques in the sketched scenario – and
similar scenarios such as company search based on their web pages [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] or the
search for scholarly collections based on the single collection items [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Hence,
a comparison with the existing journal recommenders might be an interesting
next step but is out of scope for this paper.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Results and Conclusions from our Last Contribution</title>
      <p>
        For our contribution in 2017 we took data from the dblp computer science
bibliography3. dblp offers bibliographic metadata, links to the electronic editions
of publications, and consists of nearly 3,800,000 publications. We restricted our
investigations to journal articles, extracted from the offered dump. Based on this
collection and our voting and search paradigm, CombMAX, taking only the first
ranked article of each journal into account, yielded the best results regarding the
journal ranking. For all measured searches, the utilized Votes algorithm yielded
3 dblp Homepage, http://dblp.uni-trier.de/. Last accessed 4th Jun 2018
the worst results. Analyzing the voting process based on sums of articles’ scores
we concluded that an improved CombSUM technique considering only the upper
ranking articles might deliver more accurate results than a CombSUM that takes
all results into account. Furthermore, to emphasize the first articles’ results, we
planned to include RR (Reciprocal Rank) as a voting algorithm in our test runs
which performed well in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Collection and Setup</title>
      <p>
        The experiments presented in this paper are based on a dump of 154,771,162
scientific papers offered by AMiner [
        <xref ref-type="bibr" rid="ref7 ref9">7, 9</xref>
        ]. In contrast to our preceding paper where
the utilized collection mostly covered computer science subjects, this collection
includes papers from conferences and journals across all sciences.
      </p>
      <p>
        AMiner is a project from Tsinghua University, led by Jie Tang, which deals
with search and mining of academic social networks and offers search/mining
services for the academic community. Within this research, publication data
from online databases including dblp bibliography, ACM Digital Library,
CiteSeer, and others are merged [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The complete collection is divided in three
compressed files containing the data in json format which can be downloaded
from the homepage. Not all records contain the complete metadata; fields like
issue or abstract are not filled reliable. For our experiment we parsed an extract
of 2,000,000 articles and gained 864,330 articles which had an abstract,
corresponding to 38,145 journals. These extracted articles were sent to Elasticsearch
where we carried out the experiments.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Applied Techniques</title>
      <p>Our search consists of two parts: first, we are searching over the collection titles
like we did in case of the dblp collection. In a second step, again we take the
removed titles from the collection and use them to search over the abstracts. For
similarity measures we use Elasticsearch’s TF/IDF similarity and in addition
three variations of BM25 similarity. These constellations lead to combinations
of voting techniques, field searches, and similarity measures shown in table 1.</p>
      <p>Throughout all techniques and levels we applied Elasticsearch’s built in
stopword elimination and stemming mechanisms on similarity measure level.
4.1</p>
      <p>Applied Similarity Measures for document ranking
For flat, title-based article search – i.e. the underlying document ranking – we
use Elasticsearch’s TF/IDF and BM25 algorithm in three parametrizations.
Similarity based on TF/IDF: The score(d, q) of a document d given a query
q which consists of terms t is computed as score(d, q) = Pt∈q tf (t, d) · idf (t2) ·</p>
      <p>TF/IDF,
BM25 (3 variations)
norm(d) . The term frequency tf describes the frequency of term t in document
d and is defined as tf (t, d) = √f requency. The inverse document frequency for
t across the collection is computed as idf (t) = 1 + log docnFurmedqo(tc)s+1 where
numdocs is the number of documents in the collection and docF req(t) is the
number of documents containing term t. In addition, a document length
nor1
malization is added with norm(d) = √numT erms .</p>
      <p>Similarity based on BM25: For BM25, the score(d, q) is computed as
score(d, q) = Pt∈q idf (t) · tf(t,dt)f+(kt,(d1)−·(bk++b1·)a|vDgd|l ) . |D| represents the document
length, and avgdl is computed as the average document length over all documents
in the collection. The inverse document frequency idf for term t is computed as
idf (t) = log 1 + numDdooccsF−rdeoqc(Ft)r+e0q.(5t)+0.5 with numdocs and docFreq(t) defined
as before. In our experiments, we use one standard parameterization for BM25
(k = 1.2 and b = 0.75) complemented by two variations. Variation 1: k = 3 and
b = 0.1; Variaton 2: k = 3 and b = 1.0.
4.2</p>
      <p>
        Applied Voting Techniques for journal ranking
Based on the article ranking six approaches introduced in expert search are
adopted to derive a journal ranking, respectively, a collection ranking. In general,
the voting model can be based on different inputs: the number of items in the
search result associated with a collection, the ranks of the items associated with
a collection, and the score values calculated for the items associated with a
collection [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>Let R(q) be the set of articles retrieved for the query q and score(j, q) the
computed score for journal j and query q, we apply six different voting models:
Votes: This metric takes the number of found articles for every journal as the
score: scoreVotes (j, q) = |{art ∈ R(q) ∩ art ∈ j}|
CombSUM: For every journal, CombSUM sums up the scores of the articles:
scoreCombSUM (j, q) = Part∈R(q) ∧ art∈j score(art, q)
CombSUM TOP n: For every journal, this aggregation sums up the top n (in
our work: n ∈ {5; 10}) scores of the articles for each journal:
scoreCombSUM TOP n (j, q) = Part∈R(q) ∧ art∈j ∧ rank(art,j)≤n score(art, q).
rank(art, j) represents the rank of art if only articles of j are considered.
CombMAX: This metric takes the first result stemming from j, respectively,
the article with the highest ranking, as voting candidate with its score:
scoreCombMAX (j, q) = max {score(art, q) | art ∈ R(q) ∧ art ∈ j}
RR: This algorithm takes the results from the underlying similarity measure in
their order and revalues them by applying values of a harmonic series:
scoreRR(j, q) = Part∈R(q) ∧ art∈j rank(1art,q)
5</p>
    </sec>
    <sec id="sec-5">
      <title>Experimental Results</title>
      <p>In this section we start presenting the results across all applied voting techniques,
initially disregarding the structure of the collection (journals and their associated
article count).</p>
      <p>Title Search: First experimental results are gained by searching with 10,000
randomly removed titles over the remaining 854,330 titles of the collection. Table
2 shows the results for all applied voting techniques based on TF/IDF and BM25
as similarity measure. The values in the 1st quartile column specify the rank for
which 25% of the queries have ranked the journal from which the query was
originally taken at least as good. Similarly, the median and 3rd quartile columns
state the rank under which 50%, respectively, 75% of the searches include the
searched journal. As guessed in our last paper, voting techniques emphasizing
the first ranks of the underlying article search perform better than methods
which include all article results like Votes or CombSUM. While CombMAX,
considering only the first article result for each journal, performed best in our
last paper, the newly utilized CombSUM TOP n voting techniques involving only
the first n results per journal yield even better results. RR, rescoring the articles
by their order with values of a harmonic series, produces nearly similarly good
results. For BM25 we applied the standard parameterization. Though BM25
performs slightly better than TF/IDF, this new experiment series confirms our
assumption that the underlying similarity algorithm (retrieval model) does not
fundamentally change the ranking behaviour of the collection ranking methods.
Consideration of Abstracts: In contrast to the dblp dataset, the AMiner
collection contains abstracts which we include in our search experiments. Table
3 shows the results for getting the associated journal. Please note that we search
with 10,000 titles as queries in a collection only consisting of abstracts in this
scenario. Regarding the Votes and CombSUM aggregations, rankings, and
therefore the result quality decrease significantly. An noteworthy enhancement can be
observed regarding the 3rd quartile of CombSUM Top 10 and CombSUM Top
5. Again, in case of abstracts, the choice of the underlying similarity measure
between TF/IDF and BM25 does not change the ranking considerably.</p>
      <p>Figure 1 shows the overall results. In case of abstracts, aggregation techniques
like CombSUM and Votes show significantly worse results, whereas CumbSUM
TOP 10 and CombSUM TOP 5 yield better results than RR which performs
better for the search on titles.</p>
      <p>Impact of Journal Size: In a second step the interest goes to how the
voting techniques perform dependent on the article distribution over the journals
collection. Keeping the introduced search fields, similarity measures, and
voting techniques, we divided our test queries regarding the article count of the
respective journal of origin. We examined the classifications shown in table 4.</p>
      <sec id="sec-5-1">
        <title>Abstract/BM25</title>
      </sec>
      <sec id="sec-5-2">
        <title>Aggregation Profile</title>
      </sec>
      <sec id="sec-5-3">
        <title>Abstract/TF-IDF Title/BM25</title>
      </sec>
      <sec id="sec-5-4">
        <title>Searchfield/Similarity</title>
      </sec>
      <sec id="sec-5-5">
        <title>CombMAX CombSUM TOP 10 RR</title>
      </sec>
      <sec id="sec-5-6">
        <title>CombSUM CombSUM TOP 5 Votes</title>
      </sec>
      <sec id="sec-5-7">
        <title>Title/TF-IDF</title>
        <p>Figure 2 shows the performance of requests having up to 100 articles in their
corresponding journal. Across all utilized ranking techniques, results are inferior
to the statistics regarding the entire collection (Fig. 1). Requests having 100 up
to 499 corresponding articles achieve ranking results comparable to the overall
statistics. Regarding increasing values for the number of associated articles, it
turns out that the ranking results get better for all techniques. Further,
considering Fig. 3, CombSUM using the complete sum of the scores and Votes using
the number of results for each journal, perform better than CombMAX and RR
which have a strong emphasize on the first result.</p>
        <p>System Behaviour and Average Rank Distribution: In a next step we
investigate if any of the introduced techniques fundamentally favours certain
requests dependent on their corresponding journal’s article count. In order to
compare probability of relevance with probability of retrieval, we divided the
collection into 40 bins. At an amount of 864,330 articles, each bin gains a capacity
of 21,608. The 38,145 journals from the collection are ordered by their number
of articles and passed through in ascending order. As long as the cumulated sum
of the journals’ articles does not exceed n · 21, 608, a journal gets associated with
the current nth bin. Fig. 4 shows the distribution of the 38,145 journals over the
bins. We compared the number of queries taken from each bin (corresponds to
probability of relevance) and the number of Top 1 results stemming from each bin
(corresponds to probability of retrieval) for our 10,000 title queries. Fig. 5 shows
these two distributions over the bins for the RR voting technique with underlying
800
l 600
a
n
r
u
o
J
d
e
t
rge 400
a
tf
o
k
n
a
R200</p>
        <p>0
40
30
l
a
n
r
u
o
J
d
tee 20
g
tfr
a
o
k
n
a
R10
0</p>
        <sec id="sec-5-7-1">
          <title>Abstract/BM25</title>
        </sec>
        <sec id="sec-5-7-2">
          <title>Abstract/TF-IDF Title/BM25</title>
        </sec>
        <sec id="sec-5-7-3">
          <title>Search Field/Similarity Measure</title>
        </sec>
        <sec id="sec-5-7-4">
          <title>Title/TF-IDF</title>
        </sec>
        <sec id="sec-5-7-5">
          <title>Aggregation Profile</title>
        </sec>
        <sec id="sec-5-7-6">
          <title>CombMAX</title>
        </sec>
        <sec id="sec-5-7-7">
          <title>CombSUM</title>
        </sec>
        <sec id="sec-5-7-8">
          <title>CombSUM TOP 10 RR</title>
        </sec>
        <sec id="sec-5-7-9">
          <title>CombSUM TOP 5 Votes</title>
          <p>Abstract/BM25</p>
          <p>Abstract/TF-IDF Title/BM25</p>
          <p>Search Field/Similarity Measure
9
7
2
7
1
6
6
2
5 9
40
0</p>
          <p>10 20 30</p>
          <p>Bins by Journal Size (small on the left, big on the right)
TF/IDF similarity measure. Regarding RR, CombMAX, and the CombSUM
TOP n techniques, no basic and systematic deviations of favourizations can be
observed.</p>
          <p>CombSUM and Votes show a significant tendency yielding journals with a
hight amount of articles as the Top 1 result. Fig. 6 shows an exemplary chart
for CombSUM in combination with TF/IDF. Whereas the grey graph as the
requested journals’ article count is equal to Fig. 5, CombSUM almost exclusively
yields journals having a high number of articles as top results. The same effect
occurs applying the Votes technique.</p>
          <p>Average Rank Distribution: Based on the described bin distributions we
also look at the query results and their average scoring in each bin. Fig. 7 shows
the average ranks for all applied voting techniques using TF/IDF similarity. Led
by RR, voting techniques emphasizing the upper ranks yield the best average
ranking results in lower bins (small journals). For the last ten bins, this tendency
is reversed and the aggregating voting techniques like CombSUM and Votes are
predominant.</p>
          <p>BM25 Parameter Change: In our experiment applying BM25 we also changed
the parameter values for k and b as shown in section 4.1. Notably worse results
could only be gained in case of abstracts, setting b = 0.1, dampening the
document length normalization. Figure 8 shows the overall results considering all
applied techniques.</p>
          <p>Requested Journal/Max Score Journal Requested Journal Max Score (TOP1) Journal
10 20 30</p>
          <p>Bins by Journal Size (small on the left, big on the right)</p>
          <p>Requested Journal/Max Score Journal Requested Journal Max Score (TOP1) Journal
Fig. 6. Number of queries taken from each bin (Requested Journal) vs. number of Top
1 results stemming from each bin (Max Score). Exemplary presentation for CombSUM
voting technique and TF/IDF similarity measure.
10 20 30</p>
          <p>Bins by Journal Size (small on the left, big on the right)
Result average Rank RCoRmbMAX CombSUM TOP 5 Votes</p>
          <p>CombSUM CombSUM TOP 10
40
in 400
B
n
i
k
n
a
R
e
g
a
r
ve 200
A
0</p>
          <p>0
600
l
a
n
r
u
Jod 400
e
t
e
g
tr
a
fo 200
k
n
a
R
0
5
2
srct/BM
ta
b
A</p>
          <p>)
01-3
5(.</p>
          <p>2
srct/BM
ta
b
A</p>
          <p>)
1-3
5(
2
srct/BM
ta
b
A</p>
          <p>F
srct/TF-ID
ta
b
A
5
2
Title/BM</p>
          <p>)
01-3
5(.</p>
          <p>2
Title/BM</p>
          <p>)
1-3
5(
2
Title/BM</p>
          <p>F</p>
          <p>Title/TF-ID</p>
          <p>Searchfield/Similarity
Fig. 8. Box plots for the ranks of the journals from which the respective query title
was taken considering variations of BM25 parametrization</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion and Future Work</title>
      <p>The search experiments over the AMiner collection confirm the results from our
dblp study: Summing up all found articles as voting candidates for their journal
by adding their scores or number does not perform well.</p>
      <p>CombMAX as the best performing technique in our last study is now
outperformed by RR, CumbSUM TOP 10 and TOP 5. Regardless of the applied
underlying similarity measure it turns out that the leading, top ranked articles as
voting candidates provide the best results: Whereas RR yields the best ranking
results regarding journals with few articles, the CumbSUM TOP n aggregations
perform best when requesting journals with a high amount of articles.</p>
      <p>Regarding the journals’ article count, modifying RR and CombSUM TOP
n by applying a correction factor or taking dynamically more or less voting
candidates into account could be an interesting alternative.</p>
      <p>Contrary to our expectations, the search over abstracts does not yield
significantly better results across all techniques. Voting techniques emphasizing the
first articles as voters yield slightly better rankings regarding the median and 3rd
quartile. As a modification, we plan to extract the top words out of an article’s
abstract and utilize them as metadata for our search experiments.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Henrich</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wegmann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Searching an Appropriate Journal for your Paper - an Approach Inspired by Expert Search and Data Fusion</article-title>
          . In: Lernen, Wissen, Daten,
          <string-name>
            <surname>Analysen (LWDA) Conference</surname>
            <given-names>Proceedings</given-names>
          </string-name>
          , Rostock, Germany,
          <source>September 11-13</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ounis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Searching for Expertise: Experiments with the Voting Model</article-title>
          .
          <source>In: THE COMPUTER JOURNAL</source>
          , Vol.
          <volume>52</volume>
          No.
          <issue>7</issue>
          , pp.
          <fpage>729</fpage>
          -
          <lpage>748</lpage>
          . Published by Oxford University Press on behalf of
          <source>The British Computer Society</source>
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Lee</surname>
          </string-name>
          , Joon Ho:
          <article-title>Analyses of multiple evidence combination</article-title>
          .
          <source>In: Proc. 20th annual international ACM SIGIR conf. on Research and development in information retrieval (SIGIR '97)</source>
          . ACM, New York, NY, USA, pp.
          <fpage>267</fpage>
          -
          <lpage>276</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fang</surname>
          </string-name>
          , Y.,
          <string-name>
            <surname>de Rijke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Serdyukov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Si</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Expertise Retrieval</article-title>
          .
          <source>In: Foundations and Trends in Information Retrieval</source>
          , Vol.
          <volume>6</volume>
          , pp.
          <fpage>180</fpage>
          -
          <lpage>184</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Blank</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boosz</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Henrich</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>IT company atlas upper Franconia: a practical application of expert search techniques</article-title>
          .
          <source>In: Proc. of the 31st ACM Symposium on Applied Computing (SAC '16)</source>
          . ACM, New York, NY, USA, pp.
          <fpage>1048</fpage>
          -
          <lpage>1053</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Gradl</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Henrich</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A novel approach for a reusable federation of research data within the arts and humanities</article-title>
          .
          <source>In: Digital Humanities 2014 - Book of Abstracts. EPFL - UNIL. Lausanne, Switzerland</source>
          , pp.
          <fpage>382</fpage>
          -
          <lpage>384</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Zhang, J.,
          <string-name>
            <surname>Yao</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z.</surname>
          </string-name>
          :
          <article-title>ArnetMiner: Extraction and Mining of Academic Social Networks</article-title>
          .
          <source>Proc. ACM SIGKDD International Conf. on Knowledge Discovery and Data Mining (SIGKDD</source>
          '
          <year>2008</year>
          ). pp.
          <fpage>990</fpage>
          -
          <lpage>998</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ounis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Voting for candidates: Adapting data fusion techniques for an expert search task</article-title>
          .
          <source>In: Proc. of the ACM International Conf. on Information and Knowledge Management</source>
          ,
          <source>CIKM '06</source>
          , New York, NY, USA, pp.
          <fpage>387</fpage>
          -
          <lpage>396</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. Open Academic Graph Homepage, https://aminer.org/open-academic-graph.
          <source>Last accessed 4th Jun</source>
          <year>2018</year>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>