<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An initial Analysis of Topic-based Similarity among Scienti c Documents based on their Rhetorical Discourse Parts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Carlos Badenes-Olmedo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jose Luis Redondo-Garc a</string-name>
          <email>jluisred@amazon.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oscar Corcho</string-name>
          <email>ocorchog@fi.upm.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Amazon Research</institution>
          ,
          <addr-line>Cambridge</addr-line>
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidad Politecnica de Madrid, Ontology Engineering Group</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>s, what may limit the capacity of nding the right paper for a speci c research purpose. Given the size limitations and the concise nature of abstracts, they usually omit explicit references to some contributions and impacts of the paper. Therefore for certain information retrieval tasks they cannot be considered as the most appropriate excerpt of the paper to base these operations on. In this paper we have studied other kinds of summaries, built upon textual fragments falling under certain categories of the scienti c discourse, such as outcome, background, approach, etc, in order to decide which one is more appropriate in order to substitute the original text. In particular, two novel measures are proposed: (1) internalrepresentativeness, which evaluates how well a summary describes what the full-text is about and (2) external-representativeness, which evaluates the potential of a summary to discover related texts. Results suggest that summaries explaining the method of a scienti c article express a more accurate description of the full-content than others. In addition, more relevant related articles are also discovered from summaries describing the method, together with those containing the background knowledge or the outcomes of the research paper.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        In this paper we present our rst steps on the analysis of the quality of research
article summaries. Our goal is to nd the strengths and weaknesses of approaches
leveraging exclusively on abstracts against those based on scienti c discourse
categories such as approach, challenge, background, outcomes and future work. Since
the main contributions and impacts of a research article are not always included
explicitly in the abstract, as in the case of [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] describing an architecture, where
details about the model architectures are missing, they cannot always be
considered as the most adequate scienti c summary of a research paper. In order to
judge on this accuracy, two novel measures are proposed based on the capability
of the summary to substitute the original paper: (1) internal-representativeness,
which evaluates how well the summary represents the original full-text and (2)
external-representativeness, which evaluates the summary according to how the
summary is able to produce a set of related texts that are similar to what the
original full-text has triggered.
      </p>
      <p>The paper is organized as follows: Section 2 highlights recent studies on text
mining research articles and presents the steps followed to measure the
representativeness of abstracts and research article summaries based on rhetorical
categories. It describes both the classi er used to identify those categories in
papers and the representational model and similarity metric used to compare
textual units. Experimental results comparing the di erent kind of summaries
are shown in Section 3. Finally, Section 4 presents conclusions from the
experiments.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background and Approach</title>
      <p>
        Recent studies [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ][
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] have shown that text mining of full research articles give
consistently better results than using only their corresponding abstracts. Given
the size limitations and concise nature of abstracts, they often omit descriptions
or results that are considered to be less relevant but still are important in certain
Information Retrieval (IR) tasks. Thus, when other researchers cite a particular
paper, 20% of the keywords that they mention are not present in the abstract
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        In this paper, we show our initial analysis about the representativeness of
research article summaries, considering those based exclusively on abstracts and
those based on their discursive structure (approach, challenge, background,
outcomes and future work )[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The representativeness of a summary with respect
to the original full-text is de ned as the degree of relation with the original one
(internal-representativeness ), along with the capacity of mimicking the full text
when nding related items (external-representativeness ). In order to quantify
this notions of internal-external representativeness, a probabilistic topic model
is trained over the entire set of papers to have a vectorial representation of each
text retrieved from a paper: full content-based and summary-based. The
vectorial representations of full-papers is used to measure the distance between them
and those derived from abstract or summaries (internal-representativeness ), and
also to nd similar documents (external- representativeness ) based on the
distance between their vectorial representations. An upper distance threshold is
speci ed to lter less similar pairs and compose a set of related papers for each
paper. Then, a comparison in terms of precision and recall is performed between
sets obtained by only using the vectorial representation of full-papers, against
sets produced by using other kind of summaries.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Annotation of Rhetorical Discourse Parts</title>
        <p>
          First of all, we need to identify the rhetorical parts of a research paper. Some
approaches have been proposed to summarize scienti c articles [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] taking
advantage of the citation context and the document discourse model. We have used
the scienti c discourse annotator proposed by [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] to automatically create
summaries from scienti c articles by classifying each sentence as belonging to one
of the following scienti c discourse categories: approach, challenge, background,
outcomes and future work. These categories were identi ed from the schemata
proposed by [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] with the original purpose of characterizing the content of
Computer Graphics papers. The annotator is based on a Support Vector Machine
classi er that combines both lexical and syntactic features to model each
sentence in a paper. This tool3 was integrated in the librAIry [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] Rhetoric Module4
to automatically annotate research papers with their rhetorical content.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Representational Model</title>
        <p>
          A representational model is required not only to measure distances between text
fragments but, more importantly, to help to understand the di erences in their
content. Topic models are widely used to uncover the latent semantic structure
from text corpora. In particular, Probabilistic Topic Models represent documents
as a mixture of topics, where topics are probability distributions over words.
Latent Dirichlet Allocation (LDA)[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is the simplest generative topic model that
adds Dirichlet priors for the document-speci c topic mixtures, making it possible
to characterize documents not previously used during the training task. This
is a key feature for our evaluations because, although the model used for the
experiments will be trained from the full-content of papers, it will be also used
to describe the texts summaries.
        </p>
        <p>
          Thus, we have used a LDA model to describe the inherent topic distribution of
papers in the corpus. Some hyper-parameters need to be estimated: the number
of topics (k), the concentration parameter ( ) for the prior placed on documents'
distributions over topics and the concentration parameter ( ) for the prior placed
on topics distributions over terms. Since the target of this experiment is not to
evaluate the quality of the representational model, but to compare their topic
distributions, we accepted as valid values those widely used in the literature:
= 0:1, = 0:1 , and k = 2 pn=2 = 44 where n is the size of the corpus.
Similarity Measure Feature vectors in Topic Models are topic distributions
expressed as vectors of probabilities. Hence we opt for Jensen-Shannon
divergence (JSD)[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] instead of the commonly used Kullback-Liebler divergence (KLD).
The reason for this is that KLD (1) is not de ned when a topic distribution is
zero and (2) is not symmetric, what does not t well with semantic similarity
        </p>
        <sec id="sec-2-2-1">
          <title>3 http://backingdata.org/dri/library/ 4 https://github.com/librairy/annotator-rhetoric</title>
          <p>
            measures which in general are symmetric [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ]. JSD considers the average of the
distributions as follows :
          </p>
          <p>J SD(p; q) =</p>
          <p>T
X pi log 2 pi +
i=1 pi + qi</p>
          <p>T
X qi log
i=1
2 qi
qi + pi
where T is the number of topics and p; q are the topics distributions.</p>
          <p>
            And the similarity measure used in our analysis is based on the JSD
transformed into a similarity measure as follows [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] :
          </p>
          <p>similarity(Di; Dj ) = 10 JSD(p;q)
where Di; Dj are the documents and p; q the topics distributions of each of them.
(1)
(2)
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <p>
        The corpus used in the experiments was created by combining journals in di
erent scienti c domains such as Advances in Space Research, Procedia Chemistry,
Journal of Pharmaceutical Analysis and Journal of Web Semantics. In total
1,000 papers were added, 250 from each journal. Both the abstract and the
fullcontent of these documents were directly retrieved from the Elsevier API 5 by
using the librAIry [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] Harvester module 6. The code used to perform the analysis
along with the results obtained are available in GitHub7.
      </p>
      <p>Since the annotation process to automatically discovers the rhetorical parts
of a research paper (Section 2.1) is sensitive to the structure of the phrases
that are used when writing the text, only 20% of papers in the corpus could be
fully annotated with all the fragments considered. In fact, these categories are
not present in the same proportion in the corpus: approach (90%), background
(78%), outcome (73%), challenge (57%) and future work (21%)
3.1</p>
      <sec id="sec-3-1">
        <title>Internal Representativeness</title>
        <p>The internal-representativeness of a summary measures the similarity of this
summary against the original full-text research paper. This similarity is based
on the JSD between the topic distribution of each of them.</p>
        <p>
          Since LDA considers documents as bag-of-words, the text length (e.g.
fullcontent or summaries) a ects the accuracy of the topic distributions inferred
by the topic model described in Section 2.2. The occurrences of words in short
texts are less discriminative than in long texts where the model has more word
counts to know how words are related [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. In view of the above, the approach, the
background and the outcome content of a paper generate more accurate topic
distributions than those created from other approaches such as the abstract.
        </p>
        <sec id="sec-3-1-1">
          <title>5 https://dev.elsevier.com 6 https://github.com/librairy/harvester-elsevier 7 https://github.com/librairy/study-semantic-similarity</title>
          <p>
            Also, the relative presence of each of them in a paper ( gure 2) shows an
unexpected result when compared to the IMRaD format [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ]. This style proposes
to distribute the content of an abstract, and by extension the full-paper, as
follows: Introduction(25%), Methods(25%), Results (35%) and Discussion(15%).
However, the results ( gure 2) show that Method section (approach content) is
more extensive than Results section (outcome content) in our corpus.
          </p>
          <p>All pairwise similarities between full-papers, abstracts and rhetorical-based
summaries are calculated to measure the internal-representativeness of a
summary with respect to the original text, i.e. the topic-based similarity value
(equation 2) between the probability distributions of the full-text and each of
the summaries. Results (table 1) suggest than summaries created from the
approach content are more representative than others, i.e. the distribution of topics
describing the text created from the approach content is the most similar to the
one corresponding to the full-content of the paper.</p>
          <p>Min
abstract 0.0489
approach 0.0499
background 0.0463
challenge 0.0426
futureWork 0.0000
outcome 0.0485
The external-representativeness metric tries to measure how di erent is the set of
related documents obtained from summaries with respect to those derived from
the original full-text. In terms of precision, recall and f-measure, a comparison
has been performed to analyze the behavior of the summaries when trying to
discover related content compared to use the full-text of the article.</p>
          <p>By using the same topic model previously created, similarities among all pairs
of documents were also calculated according to equation 2. Then, a minimum
score or similarity threshold is required to de ne when a pair of papers are
related. Each threshold is used to create a gold-standard which relates articles
to others based on their similarity values. In order to discover that lower bound
of similarity, a study about trends in the similarity scores ( g 3) as well as
distributions of topics in the corpus ( g 4) was performed. We can see that topics
are not equally balanced across papers. This fact generates separated groups of
strongly related papers. We think this phenomena is due to our usage of a corpus
created from journals where di erent domains are equally balanced. Then, we
considered a similarity score equals to 0:99 ( g 3) as the threshold from which
strong relations appear. However, to cover di erent interpretations of similarity,
from those based on sharing general ideas or themes to those that imply to
share a more speci c content, the following list of thresholds was considered in
the experiments: 0.5, 0.6, 0.7, 0.8, 0.9, 0.95 and 0.99.</p>
          <p>For each similarity threshold, a gold-standard was created based on
considering as related those papers with a similarity value upper than the selected
threshold. Results ( gure 5) comparing the related papers inferred from the
fullcontent with those inferred from the partial-content representation (i.e. abstract
or rhetorical parts) suggest that strongly related papers are mainly discovered
by using the summary created from the approach section. The reason for this
may be based on the average size of this type of summaries or the particular
content included in this part of a paper. While other summaries include more
general-domain words, the approach content includes more speci c words that
describe the method or the nal objective of the paper. So, for higher similarity
thresholds, i.e. for strongly related papers, the recommendations discovered by
using the approach are more precise than those discovered by using the abstract.</p>
          <p>In terms of recall ( gure 6), the upward trend followed by the approach, the
outcome and the background content remarks the assumption of summaries
containing key words allow to discover more similar papers than others. Moreover,
since recall overlooks false-negatives classi cations, it suggests that these parts
of a research paper share more words than others with strongly related papers
but they may also present commonalities with highly related papers, except in
case of approach which still exhibits higher precision.</p>
          <p>As expected, only summaries created from the approach, the outcome and the
background content maintain high accuracy values ( g 7) even for high similarity
thresholds. Along with the results showed in gure 8, where the same three
rhetorical classes present the lowest standard deviation over the f-measure, they
can be considered as the most robust summaries containing the ideas that better
characterize the paper compared to others.
We have studied the Topic-based similarities among scienti c documents based
on their abstract sections with respect to summaries corresponding to their
scienti c discourse categories. For this purpose, two novel measures have been
proposed: (1) internal-representativeness and (2) external-representativeness.</p>
          <p>Results show that summaries created from the approach, outcome or
background content of a paper describe more accurately its full-content in terms of
overall ideas and related documents than abstracts. Although those summaries
are more extensive in number of characters than other with similar precision such
as the abstract content, they have proven to be particularly helpful discovering
strongly related papers, i.e. papers with a similarity value close to 1.0.</p>
          <p>
            In order to avoid an in uence of the size of the summaries on the accuracy of
the results, in future work we plan to use probabilistic topic model algorithms
oriented to handle short-texts such as BTM [
            <xref ref-type="bibr" rid="ref3">3</xref>
            ] to describe texts.
          </p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Badenes-Olmedo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Redondo-Garcia</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corcho</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Distributing Text Mining tasks with librAIry</article-title>
          .
          <source>In: Proceedings of the 17th ACM Symposium on Document Engineering (DocEng)</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M.I.</given-names>
          </string-name>
          :
          <article-title>Latent Dirichlet Allocation</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>3</volume>
          (
          <issue>4-5</issue>
          ),
          <volume>993</volume>
          {
          <fpage>1022</fpage>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Cheng,
          <string-name>
            <given-names>X.</given-names>
            ,
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            ,
            <surname>Lan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Guo</surname>
          </string-name>
          , J.:
          <source>BTM: Topic Modeling over Short Texts. Knowledge and Data Engineering, IEEE Transactions on PP(99)</source>
          ,
          <volume>1</volume>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Cohan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goharian</surname>
          </string-name>
          , N.:
          <article-title>Scienti c Article Summarization Using Citation-Context and Article's Discourse Structure</article-title>
          .
          <source>In: Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>390</volume>
          {
          <issue>400</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Dagan</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira</surname>
            ,
            <given-names>F.C.N.</given-names>
          </string-name>
          :
          <article-title>Similarity-Based Models of Word Cooccurrence Probabilities</article-title>
          .
          <source>Machine Learning</source>
          <volume>34</volume>
          (
          <issue>1-3</issue>
          ),
          <volume>43</volume>
          {
          <fpage>69</fpage>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Divoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hearst</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          :
          <article-title>Do peers see more in a paper than its authors? Advances in Bioinformatics (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Hong</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davison</surname>
            ,
            <given-names>B.D.</given-names>
          </string-name>
          :
          <article-title>Empirical study of topic modeling in Twitter</article-title>
          .
          <source>Proceedings of the First Workshop on Social Media Analytics - SOMA '</source>
          10 pp.
          <volume>80</volume>
          {
          <issue>88</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <source>Divergence Measures Based on the Shannon Entropy. IEEE Transactions on Information Theory</source>
          <volume>37</volume>
          (
          <issue>1</issue>
          ),
          <volume>145</volume>
          {
          <fpage>151</fpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>E cient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>Proceedings of the International Conference on Learning Representations (ICLR</source>
          <year>2013</year>
          ) pp.
          <volume>1</volume>
          {
          <issue>12</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Nair</surname>
            ,
            <given-names>P.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nair</surname>
          </string-name>
          , V.D.:
          <article-title>Organization of a Research Paper: The IMRAD Format</article-title>
          . In:
          <article-title>Scienti c Writing and Communication in Agriculture and Natural Resources</article-title>
          , p.
          <volume>150</volume>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ronzano</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saggion</surname>
          </string-name>
          , H.:
          <article-title>Dr. Inventor Framework: Extracting Structured Information from Scienti c Publications</article-title>
          .
          <source>In: Discovery Science: 18th International Conference</source>
          . pp.
          <volume>209</volume>
          {
          <issue>220</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Rus</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Niraula</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banjade</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Similarity Measures Based on Latent Dirichlet Allocation</article-title>
          .
          <source>In: Computational Linguistics and Intelligent Text Processing</source>
          , pp.
          <volume>459</volume>
          {
          <issue>470</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Sciences</surname>
            ,
            <given-names>E.R.S.f.P.</given-names>
          </string-name>
          ,
          <article-title>life: Harnessing the power of content - Extracting value from scienti c literature: the power of mining full-text articles for pathway analysis Harnessing the Power of content (</article-title>
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Simone</surname>
          </string-name>
          <article-title>Teufel: The Structure of Scienti c Articles - Applications to Citation Indexing and Summarization</article-title>
          .
          <source>In: CSLI Studies in Computational Linguistics</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Teufel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siddharthan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Batchelor</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Towards discipline-independent Argumentative Zoning: Evidence from chemistry and computational linguistics</article-title>
          .
          <source>Conference on Empirical Methods in Natural Language</source>
          Processing pp.
          <volume>1493</volume>
          {
          <issue>1502</issue>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Westergaard</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , St rfeldt, H.h., T nsberg, C.,
          <string-name>
            <surname>Jensen</surname>
            ,
            <given-names>L.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brunak</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Text mining of 15 million full-text scienti c articles</article-title>
          .
          <source>bioRxiv</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>