<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shutian Ma</string-name>
          <email>mashutian0608@hotmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jin Xu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jie Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chengzhi Zhang</string-name>
          <email>zhangcz@njust.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Management, Nanjing University of Science and Technology</institution>
          ,
          <addr-line>Nanjing, China, 210094</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University)</institution>
          ,
          <addr-line>Fuzhou, China, 350108</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper introduces NJUST system which is applied in the CLSciSumm 2017 Shared Task at the BIRNDL 2017 Workshop. The training corpus contains 10 articles of training set, 10 articles of development set and 10 articles of test set from CL-SciSumm 2016. Articles were created by randomly sampling documents from the ACL Anthology corpus and selecting their citing papers. In Task 1A, we utilize different measurements to compute sentence similarities. Four classifiers are trained using different features and final results are obtained by voting system. In Task 1B, rule-based methods are mainly used according to high frequency words. As to Task 2, we generate a summary within 250 word based on the identified sentences in the reference paper from its cited text spans using maximal marginal relevance.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Scientific papers are usually measured by their citances in citing papers which reveal
the extent to which a reference paper has been used by other researchers. So far, most
investigation has been focused on citation analysis from using simple index of citation
counts [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ] to complex natural language processing of citation contents [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
However, using citances can’t provide context from the reference paper, for example, the
type of information cited or where it is in the referenced paper. To understand different
perspectives of a reference paper, it’s important to generate summary from all the cited
text spans in the reference paper from citations [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8">5, 6, 7, 8</xref>
        ]. The CL-SciSumm 20172
has been designed to do automated summarization of scientific contributions for the
computational linguistics research domain, which can help readers to gain a gist of the
state-of-the-art in research for a topic.
      </p>
      <p>CL-SciSumm 2017 has been divided into two tasks. Firstly, we should identify text
spans in reference paper which most accurately reflect citance, facets of paper are also
needed to be distinguished. Second task is to generate a summary of reference paper
from the identified cited text spans. In this paper, we describe our methods applied for
CL-SciSumm 2017. As to Task 1A, we trained four classifiers and integrate all the
results by voting system. In Task 1B, rule-based methods are mainly used on identified
text span to determine which facet it belongs to. As to Task 2, we generate a summary
using maximal marginal relevance.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        This year’s CL-SciSumm 2017 takes place at the Joint Workshop on
Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries
(BIRNDL 2017)3 and is a follow-up on the shared task of CL-SciSumm 20164 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
Originally, the CL Summarization Pilot Task was conducted as a part of the
BiomedSumm Track at the Text Analysis Conference 2014 (TAC 2014)5 [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. There have been
many investigations on task problem previously [
        <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14 ref15 ref16">11, 12, 13, 14, 15, 16</xref>
        ].
      </p>
      <p>
        When doing Task 1A, most teams identified the linkage between a paper citation in
citing paper and the corresponding cited text spans in reference paper by computing
sentences similarities. CIST system applied two kinds of features, one is from lexicons,
and another is from sentence similarities [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Aggarwal and Sharma made use of
subsequences (of words) overlap [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Bi-grams were identified between generated
bagof-words to find matching statement in their study. PolyU [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] utilized TF-IDF cosine
similarity, position of sentence chunk and some lexical rules. SVM and its modification
model were chosen as the classifier for many teams [
        <xref ref-type="bibr" rid="ref11 ref12 ref15">11, 12, 15</xref>
        ]. New models have also
been proposed by combining new algorithms. Klampfl, Rexha and Kern proposed
TextSentenceRank for extracting candidate text spans which is inspired by graph based
ranking algorithms [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Nomoto introduced a composite model consisting of TF-IDF
and Neural Network [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        As for Task 1B, since the instances for the Implication and Hypothesis facets are
very limited, some teams only trained classification model on data of the other three
facets [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Machine learning models such as, decision tree [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], random forest
classifier [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], and SVM [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] were applied to conduct classification. Lexical rules are mainly
used on section headers or citance content [
        <xref ref-type="bibr" rid="ref12 ref13 ref16">12, 13, 16</xref>
        ]. Researchers will try to build
word lists for each facet which are similar words within each list. And then, they will
examine whether the subtitles of reference sentences or cited sentences contains the
following facet words or not for identification.
      </p>
      <p>
        Few teams took part in Task 2 of generating summary. CIST system calculated
sentence scores of five features: hLDA-level distribution feature, sentence-length feature,
sentence-position feature, cited text span and RST-feature. They also use discourse
facet to extract best-N sentences from all the sentences or from each cluster [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. PolyU
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] converted Task 2 into the query-focused multi-document summarization problem.
They used improved manifold ranking by modifying the prior score distribution to
inspect the importance of citances.
      </p>
      <sec id="sec-2-1">
        <title>3 Available at: http://wing.comp.nus.edu.sg/~birndl-sigir2017/</title>
        <p>4 Available at: http://wing.comp.nus.edu.sg/cl-scisumm2016/
5 Available at: http://www.nist.gov/tac/2014</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <sec id="sec-3-1">
        <title>Task Description</title>
        <p>There are two tasks in CL-SciSumm 2017 and framework is shown in Figure 1. The
training dataset contains 30 topics of documents. A topic is consisted of a Reference
Paper (RP) and Citing Papers (CPs) that all contain citations to the RP. In each CP, the
text spans (citances) have been identified that pertain to a particular citation to the RP.
In Task 1A, for each citance, we need to identify the spans of text (cited text spans) in
the RP that most accurately reflect the citance. In Task 1B, for each cited text span, we
need to identify what facet of the paper it belongs to, from five predefined facets, which
are Aim, Method, Results, Implication and Hypothesis. In Task 2, we need to generate
a structured summary of the RP from the cited text spans of the RP.</p>
        <p>Identify cited text
span in the RP</p>
        <sec id="sec-3-1-1">
          <title>Identify facet of cited text span</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>Task 1A</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>Task 1B</title>
          <p>In this task, we are asked to identify the reference sentences referred to by a given
citance. We approach this problem from the perspective of finding the sentence in RP
which is more similar with citance and treat it as a classification task. In order to get
better performance, we applied different classifiers and combined their results by
voting system. In order to train the models, three kinds of features are obtained. Short
descriptions of features are shown in Table 1.</p>
        </sec>
        <sec id="sec-3-1-4">
          <title>Feature Type</title>
        </sec>
        <sec id="sec-3-1-5">
          <title>Similaritybased features</title>
        </sec>
        <sec id="sec-3-1-6">
          <title>Doc2Vec similarity</title>
        </sec>
        <sec id="sec-3-1-7">
          <title>Rule-based features</title>
        </sec>
        <sec id="sec-3-1-8">
          <title>Position-based features</title>
        </sec>
        <sec id="sec-3-1-9">
          <title>Bigram Sid Ssid</title>
        </sec>
        <sec id="sec-3-1-10">
          <title>Sentence Position</title>
        </sec>
        <sec id="sec-3-1-11">
          <title>Section Position</title>
        </sec>
        <sec id="sec-3-1-12">
          <title>Inner Position</title>
        </sec>
        <sec id="sec-3-1-13">
          <title>Bi-gram matching value, if there is bi-gram matched,</title>
          <p>the value is 1; otherwise, value is 0.</p>
          <p>Sentence position in the full text
Sentence position in the corresponding section
The sentence position, divided by the number of
sentences
The position of the corresponding section of the
sentence chunk, divided by the number of sections
The sentence position in the section, divided by the
number of sentences in the section</p>
          <p>Based on the annotation files, we give labels to the matched sentence pairs with 1
and unmatched sentence pairs with 0. When training classifiers, we firstly tried six
different models, including SVM (kernel=linear), SVM (kernel=rbf), SVM
(kernel=sigmoid), decision tree, logistics regression and nearest neighbor. Different features are
investigated on all datasets of CL-SciSumm 2017. According to the 10 fold cross
validation results, we remove SVM (kernel=sigmoid) and nearest neighbor, and choose
different features for the remaining classifiers. The average F1 values of all features for
Task 1A on training dataset are shown in Figure 2. In order to find good features, we
trained the classifiers for 8 runs with the different class ratios of 0 and 1 labels. From
Figure 2 (a) to Figure 2 (h), the class ratio of 0 to 1 is 1, 1.5, 2, 2.5, 3, 5, 7.5, and 10.</p>
          <p>SVM(RBF) SVM(linear) Decision Tree Logistic Regression
(e) Average F1 when class ratio of 0 to 1 is 3</p>
          <p>SVM(RBF) SVM(linear) Decision Tree Logistic Regression
(f) Average F1 when class ratio of 0 to 1 is 5
图表标题
0.4
0.3 0.35 0.25
000..21.552 0.002..325 00.1.52
0.1 0.15 0.1
0.05 0.1 0.05
0 SVM(RBF0).050 SVM(linear) Decision Tree Logistic Regression 0 SVM(RBF) SVM(linear) Decision Tree Logistic Regression
(g) Average F1 whenSVcMla(RsBsF)ratio of 0 toSV1Mi(slin7e.a5r) (h) DAecviseiornaTgreeeF1 whenLogcisltaicsRsergaretsisoionof 0 to 1 is 10
sid ssid sent_position sec_position inner_position lda_sim
jaccard_sim tf_idf_sim idf_sim bigram d2v_sim
Fig. 2. Average F1 of All Features for Task 1A with Different Proportion of 0/1 Sample Size</p>
          <p>Based on these results, we can find that similarity-based features show better
performance than the others. So we keep all similarity-based features, rule-based feature and
choose some of the position-based features as the final features. Moreover, we set
different weight to each classifier while all the results are integrated by voting system.
Parameter settings are shown in Table 3.</p>
          <p>Due to the big quantitative gap between 1 and 0 labels, we trained the classifiers for
5 runs with the different proportion of 1 and 0 labels and set penalty factor as well.
Furthermore, we also set different thresholds to the voting system. Detailed information
of 1 and 0 label proportion and voting system thresholds in 5 runs is shown in table 3.
Finally, according to the requirements of Task 1A, we did tuning on obtained results.
For each citance, if the identified text spans contain more than 5 sentences, then we will
list sentences in the order of Jaccard similarity from big to small, and pick the top 5
sentences to be the final results. If we can’t identify any text span, then we will list
sentences in the order of Jaccard similarity from big to small, and pick the top 1
sentence to be the final result.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Task 1B</title>
        <p>In this task, for each cited text span, we need to identify what facet of the paper it
belongs to. We construct three dictionaries of five facets Manual Dictionary, POS
Dictionary-I and POS Dictionary-II. The first one is made manually and another two is
made according to part-of-speech tagging results. Facet identification strategy of Task
1B is shown in Figure 3.</p>
        <p>Referring to manual dictionary, we looked through each identified text span of five
facets from all the annotation files in datasets. Then we build the dictionaries by judging
every word within the sentence context manually. Two graduate students took part in
this task.</p>
        <p>Referring to POS dictionary, we firstly made part-of-speech tagging by Stanford
POS Tagger6 on the section title and sentence content in all the labeled annotation files.
Then we keep the words which are adjectives and verbs and make all words as the
automatic dictionary of section title and sentence content separately. We then list all
words by frequency order according to five facets. After removing the words whose
frequency is less than 2, the left words are the automatic dictionary of section title and
sentence content separately. This is the POS dictionary-I. Since there are more words
that related to method citation. We built POS dictionary-II by removing the method
dictionary of section title and sentence content.</p>
        <p>Based on the five different dictionaries of five facets, if the section title or sentence
content contains any one of these words in the corresponding built dictionaries, it will
be directly classified as the corresponding facet. Since the manual dictionary will be
more accurate than POS dictionary. When using manual dictionary, identified facets
will be all kept which means one sentence can have more than one facet. When using
POS dictionary, the order of judging facet is hypothesis, aim, implication, method and
result and later identified facet will override the former one. Finally, each sentence will
have five identified facets, if five facets contain more than three of one facet, then we
classify it as this facet. Else if it contains more than three different facets, we just
classify it as the facet of Method.</p>
        <p>Manual Dictionary</p>
        <p>POS Dictionary-I</p>
        <p>POS Dictionary-II</p>
        <sec id="sec-3-2-1">
          <title>Matching</title>
          <p>Hypothesis  Aim  Implication  Method  Result</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>Citance</title>
        </sec>
        <sec id="sec-3-2-3">
          <title>Sentence Content</title>
        </sec>
        <sec id="sec-3-2-4">
          <title>Section Title</title>
        </sec>
        <sec id="sec-3-2-5">
          <title>Five identified results</title>
        </sec>
        <sec id="sec-3-2-6">
          <title>Final identified results</title>
        </sec>
        <sec id="sec-3-2-7">
          <title>6 Available at: https://nlp.stanford.edu/software/tagger.html</title>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>Task 2</title>
        <p>
          Summary generation is divided into two main steps. First is to group sentences into
different clusters by bisecting K-means [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. Second is using maximal marginal
relevance (MMR) [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] to extract sentence from each cluster and combine them into a
summary.
        </p>
        <p>
          Firstly, we use vector space model to represent documents and then non-negative
matrix factorization is conducted to reduce the document dimension into 50
dimensions. Then we apply the bisecting K-means which is based on K-means. Bisecting
Kmeans can be divided into four steps: 1.Pick a cluster to split; 2.Find 2 sub-clusters
using the basic K-means algorithm; 3. Repeat step 2, the bisecting step, for a fixed
number of times and take the split that produces the clustering with the highest overall
similarity. (For each cluster, its similarity is the average pairwise document similarity,
and we seek to minimize that sum over all clusters.); 4. Repeat steps 1, 2 and 3 until the
desired number of clusters is reached. After obtaining the clusters, we list all the clusters
in the order of cluster size from big value to small value. And then, all the sentences
within each cluster are listed in the order of MMR from big value to small value. The
basic idea of MMR is straightforward [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]: if we have a set of items  and we want to
recommend a subset   ⊂  (  ℎ
query  . MMR proposes to build 
|  | =
        </p>
        <p>≪ | | ) relevant to a given
 by selecting   ∗ given   −1 = { 1∗, ⋯ ,   ∗−1}
(  ℎ

 =   −1 ∪ {  ∗} ) according to the following criteria:
  ∗ = arg</p>
        <p>max
  ∈ \  −1
[ (
1(  ,  )) − (1 −  ) max</p>
        <p>2(  ,   )]
  ∈  −1
Where</p>
        <p>
          1(∙,∙) measures the relevance between an item and a query, 
measures the similarity between two items, and the manually tuned  ∈ [
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ] trades off
relevance and similarity. In the case of  1∗, the second term disappears.
        </p>
        <p>Finally, for each time, we choose first two sentence from each cluster to build the
summary before the length of summary exceeds 250 words.</p>
        <p>(1)
2(∙,∙)
4
4.1</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <sec id="sec-4-1">
        <title>Task 1A</title>
        <p>When doing corpora preprocessing, we remove the stop words and stem words to base
forms by Porter Stemmer algorithm7. Then, we applied D2V model in Genism8 and
python package9 of LDA model to represent documents. All the classifiers were done
via Scikit-learn10 python package. The source code of our system will be successively
open on the Github website: https://github.com/KingChristenson/NJUST-CL.</p>
        <p>For classification experiments, we split training dataset into two separate datasets:
10 articles of training set and 10 articles of development set from CL-SciSumm 2016
7 Available at: http://tartarus.org/~martin/PorterStemmer/
8 Available at: http://radimrehurek.com/gensim/index.html
9 Available at: https://pypi.python.org/pypi/lda
10 Available at: http://scikit-learn.org/stable/index.html
are chosen as train dataset. 10 articles of test set from CL-SciSumm 2016 are chosen as
test dataset. There are five runs that we submitted. Precision, Recall and F1 values
which we got from the test dataset are shown in Table 4.</p>
        <p>We also draw Figure 4 (a) and Figure 4 (b) to see the trend of different evaluation
results when increasing the class ratio of 0 to 1 and thresholds for voting system.
From Figure 4 (a), we can find that with the increasing of 0/1 sample size, although
the precision value is increasing slowly, according to F1 value, the performance of Task
1A is getting worse. The same situation happened when we increasing the threshold.
So it’s important to choose the proper parameters in such classification tasks, such as
the 0/1 sample size and threshold for voting system.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Task 1B</title>
        <p>We tried all results from Task 1A, and then got the best performance by voting
system for 5 runs. Table 5 shows our Task 1B results of the train data according to different
facets.</p>
        <p>From Table 6, we can find that identification of method citation performs best since
it’s also the most common facet shown in all citations. Citation facet of result, aim and
implication shows bad performance. The poor quality of built dictionary might lead to
this results. More features should be considered when doing this task, such as the
sentence position or section title position.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and Future Work</title>
      <p>
        This document demonstrates our participant system NJUST on CL-SciSumm 2017. Our
system has tried to add some semantic information like doc vector and topic
distributions in LDA to improve the citance linkage and summarization performance. When
choosing features, we find that TF-IDF similarity and IDF similarity do better than the
similarities based Doc2Vec and LDA. In order to improve classification performance,
several classifiers are trained with different features. The final results are obtained by
voting system. When doing Task 2, we use maximal marginal relevance to rank
sentences for summary generation. According to the evaluation [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], we did the best
performance in Task 1A and also good in Task 1B, while strategy for Task 2 didn’t work
well and more work can be done in all the tasks.
      </p>
      <p>In the future work, we need to find better ways to measure sentence similarities and
use some machine learning models to do Task 1B. As to summarization, we will try to
combine the sentence with its identified facet information for organizing the sentence
order. Furthermore, more features can be added to calculate the sentence score for
ranking, such as sentence length, sentence position, etc.
. Acknowledgements
This work is supported by Major Projects of National Social Science Fund (No.
16ZAD224), Fujian Provincial Key Laboratory of Information Processing and
Intelligent Control (Minjiang University) (No. MJUKF201704) and Qing Lan Project.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Garfield</surname>
            <given-names>E</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Merton R K.</surname>
          </string-name>
          <article-title>Citation indexing: Its theory and application in science, technology</article-title>
          , and humanities [M]. New York: Wiley,
          <year>1979</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Meho</surname>
            <given-names>L I</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>K.</given-names>
          </string-name>
          <article-title>Impact of data sources on citation counts and rankings of LIS faculty: Web of Science versus Scopus and Google Scholar [J]</article-title>
          .
          <source>Journal of the Association for Information Science and Technology</source>
          ,
          <year>2007</year>
          ,
          <volume>58</volume>
          (
          <issue>13</issue>
          ):
          <fpage>2105</fpage>
          -
          <lpage>2125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bornmann</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daniel H D.</surname>
          </string-name>
          <article-title>What do citation counts measure? A review of studies on citing behavior [J]</article-title>
          .
          <source>Journal of documentation</source>
          ,
          <year>2008</year>
          ,
          <volume>64</volume>
          (
          <issue>1</issue>
          ):
          <fpage>45</fpage>
          -
          <lpage>80</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Zhang</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ding</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milojević S</surname>
          </string-name>
          .
          <article-title>Citation content analysis (cca): A framework for syntactic and semantic analysis of citation content [J]</article-title>
          .
          <source>Journal of the Association for Information Science and Technology</source>
          ,
          <year>2013</year>
          ,
          <volume>64</volume>
          (
          <issue>7</issue>
          ):
          <fpage>1490</fpage>
          -
          <lpage>1503</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Jaidka</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khoo C S G</surname>
            ,
            <given-names>Na J C</given-names>
          </string-name>
          , et al.
          <article-title>Deconstructing Human Literature Reviews-A Framework for Multi-</article-title>
          Document
          <string-name>
            <surname>Summarization</surname>
          </string-name>
          [C]//ENLG.
          <year>2013</year>
          :
          <fpage>125</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Nenkova</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McKeown K. Automatic summarization</surname>
          </string-name>
          [J].
          <source>Foundations and Trends® in Information Retrieval</source>
          ,
          <year>2011</year>
          ,
          <volume>5</volume>
          (
          <issue>2</issue>
          -3):
          <fpage>103</fpage>
          -
          <lpage>233</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Teufel</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moens</surname>
            <given-names>M.</given-names>
          </string-name>
          <article-title>Summarizing scientific articles: experiments with relevance and rhetorical status [J]</article-title>
          .
          <source>Computational linguistics</source>
          ,
          <year>2002</year>
          ,
          <volume>28</volume>
          (
          <issue>4</issue>
          ):
          <fpage>409</fpage>
          -
          <lpage>445</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Jones K S.</surname>
          </string-name>
          <article-title>Automatic summarising: The state of the art</article-title>
          [J].
          <source>Information Processing &amp; Management</source>
          ,
          <year>2007</year>
          ,
          <volume>43</volume>
          (
          <issue>6</issue>
          ):
          <fpage>1449</fpage>
          -
          <lpage>1481</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Jaidka</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chandrasekaran</surname>
            <given-names>M K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rustagi</surname>
            <given-names>S</given-names>
          </string-name>
          , et al.
          <article-title>Overview of the CL-SciSumm 2016</article-title>
          <string-name>
            <surname>Shared Task</surname>
          </string-name>
          [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>93</fpage>
          -
          <lpage>102</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Jaidka</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chandrasekaran M K</surname>
            , Elizalde
            <given-names>B F</given-names>
          </string-name>
          , et al.
          <article-title>The computational linguistics summarization pilot task</article-title>
          [C]//Proceedings of Text Ananlysis Conference, Gaithersburg, USA.
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Li</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mao</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            <given-names>Y</given-names>
          </string-name>
          , et al.
          <article-title>CIST System for CL-SciSumm 2016</article-title>
          <string-name>
            <surname>Shared Task</surname>
          </string-name>
          [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>156</fpage>
          -
          <lpage>167</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Cao</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            <given-names>D</given-names>
          </string-name>
          . PolyU at CL-SciSumm
          <year>2016</year>
          [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>132</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Aggarwal</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sharma</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Lexical</surname>
          </string-name>
          and Syntactic cues to identify Reference Scope of Citance [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>103</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Nomoto</surname>
            <given-names>T. NEAL</given-names>
          </string-name>
          :
          <article-title>A Neurally Enhanced Approach to Linking Citation</article-title>
          and Reference [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>168</fpage>
          -
          <lpage>174</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Moraes</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baki</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Verma R M</surname>
          </string-name>
          , et al. University of Houston at CL-SciSumm
          <year>2016</year>
          :
          <article-title>SVMs with tree kernels</article-title>
          and Sentence Similarity [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>113</fpage>
          -
          <lpage>121</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Klampfl</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rexha</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kern</surname>
            <given-names>R</given-names>
          </string-name>
          .
          <article-title>Identifying Referenced Text in Scientific Publications by Summarisation</article-title>
          and Classification Techniques [C]//BIRNDL@ JCDL.
          <year>2016</year>
          :
          <fpage>122</fpage>
          -
          <lpage>131</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Steinbach</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karypis</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            <given-names>V.</given-names>
          </string-name>
          <article-title>A comparison of document clustering techniques</article-title>
          [C]//KDD workshop on text mining.
          <year>2000</year>
          ,
          <volume>400</volume>
          (
          <issue>1</issue>
          ):
          <fpage>525</fpage>
          -
          <lpage>526</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Carbonell</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldstein</surname>
            <given-names>J.</given-names>
          </string-name>
          <article-title>The use of MMR, diversity-based reranking for reordering documents and producing summaries[C]//Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval</article-title>
          .
          <source>ACM</source>
          ,
          <year>1998</year>
          :
          <fpage>335</fpage>
          -
          <lpage>336</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Guo</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanner</surname>
            <given-names>S</given-names>
          </string-name>
          .
          <article-title>Probabilistic latent maximal marginal relevance[C]//Proceedings of the 33rd international ACM SIGIR conference on Research and development in information retrieval</article-title>
          .
          <source>ACM</source>
          ,
          <year>2010</year>
          :
          <fpage>833</fpage>
          -
          <lpage>834</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Kokil</surname>
            <given-names>Jaidka</given-names>
          </string-name>
          , Muthu Kumar Chandrasekaran, Devanshu Jain, and
          <string-name>
            <surname>Min-Yen Kan</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Overview of the CL-SciSumm 2017 Shared Task</article-title>
          ,
          <source>In Proceedings of the Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries (BIRNDL</source>
          <year>2017</year>
          ), Tokyo, Japan, CEUR.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>