<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Joint Extract Method from Scholarly Papers 1</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jianfan Ge</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ting Jiang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Nanjing University Of Finance &amp; Economics</institution>
          ,
          <addr-line>NanJing</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <fpage>59</fpage>
      <lpage>68</lpage>
      <abstract>
        <p>Entities and relations concisely reflect important information related to the subject matter of the literature, which is essential for understanding and analyzing it. In scientific research, methods are indispensable tools and important research objects for solving scientific problems (methods include tasks, discipline-specific methods, models, algorithms, and metrics, etc.). Therefore, 'method' is indispensable for the understanding and analysis of academic literature. This paper aims to extract method-like entities and relations from scientific abstracts using a semantically enhanced deep learning model. We explored the impact of linguistic information on the entity and relation extraction task, and to this end, we added additional POS tag information to the word vectors obtained through the pre-trained model to highlight POS tag information, which proved to be superior to the pre-trained word vectors alone. Individually, in the entity recognition part, the token sequence length of entities is considered as the feature, and in the relationship extraction part, performing max pooling over the context between entity candidates has been proven better than full context, additionally, the distance between entity candidates is embedded by us as an additional feature. Entity type is also entered as an additional feature. The sequences of rich token representations constitute a span, over which entities and relations are learned jointly. The results on several datasets show that the embedding of rich semantic information outperforms the original span-based model.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;entity extraction</kwd>
        <kwd>relation extraction</kwd>
        <kwd>method</kwd>
        <kwd>POS tag</kwd>
        <kwd>entity distance</kwd>
        <kwd>token length</kwd>
        <kwd>entity type</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In the era when the data was dense, a large number of papers are published daily[1-3]. For most
academic researchers, considering the diversity and explosive growth of research in the field, the speed
of reading is much slower than the speed of publication, making it impossible to always access the latest
methods from the most recent literature, which means that traditional manual searches to find scientific
methods become challenging[4]. Therefore, in order to help scholars form a methodological system for
their research directions and obtain the most cutting-edge methods related to the research content while
greatly saving the labor and time of researchers, it is essential to study the method to extract methods
from large-scale academic literature. As is well known, entity recognition (ER) and relationship
extraction (RE) are essential and challenging tasks in natural language processing (NLP), which can be
beneficial to information extraction from academic literature. Entity recognition and relationship
extraction from academic texts refer to identifying academic entities, such as Task, Method, Metric, and
so on, and extracting semantic relations among these entities, e.g Evaluate-For, Used-For, and so on. ER
and RE from literature are used in a wide range of academic applications, including academic
information retrieval, knowledge graph construction, question answering, article recommendation, etc.
The joint model of ER and RE from scholarly texts aims to extract the entity-relationship-entity tuples.
For example, the following sentence S1 shown in Figure 1 contains two entities; we delineate the mention
with square brackets and the corresponding entity types with suffixes：</p>
      <p>Previous research has shown that method entities identified by complex rules are more responsive to
searchers' needs than those identified by matching terms in academic literature[5]. As a result,
researchers have proposed more complex rules to accomplish the extraction task, including cue words,
language patterns, lexicality, word position, etc[6-10]. duck et al.[11] created a named entity recognizer
used in bioinformatics called bioNerDS to extract software entities and dataset entities from papers. Noun
phrases from academic papers were extracted and scored based on different rules. In the first round,
candidate entities are checked to see if they appear in the generated dictionary. In the second round,
strong rules extracted from the article, such as version information, references and URLs, are classified
into positive and negative rules and assigned different scores. In the last round, some clues, such as
specific verbs, and indicative but ambiguous titles, are combined into weak clues and assigned scores.
Candidate entities were scored according to their compliance with the rules and judged as method entities
based on their final scores.</p>
      <p>In this paper, we propose a semantically enhanced term entity relation extraction model to jointly
extract method entities and relationships from the abstracts of scientific papers. We use a recent model
called SPERT as the baseline, which uses a pre-trained transformer. A shallow entity classifier and a
shallow relationship classifier are applied to extract entities and relations, respectively. The transformer
generates embeddings of tokens in the abstract, and merges the embeddings of a span of tokens into one.
Many natural language processing tasks benefit from the use of linguistic information, such as
part-ofspeech tagging, but they are less explored in deep neural models of NER and RE. Petasis et al.[12]
believed that named entities were proper nouns (PN), which served as the name of someone or something.
From the perspective of ontology, Alfonseca and Manandhar[13] proposed that named entities were
objects used to solve specific problems. Borrega et al.[14] defined named entities in detail from the
perspective of linguistics, stipulating that only nouns and noun phrases can be used as named entities.
Although these definitions are not uniform, it can be sure that named entities at least the vast majority
are nouns or noun phrases. The relationships between entity pairs also seem to be related to entity types,
for example, we often find "used-for" relations between "generic" entities and "method" entities. In
relation extraction, contextual information between candidate entities has been proven to be superior to
global contextual information. And in addition to the important information brought by the semantics
itself, the distance information between candidate entities should also be an important feature. For this
purpose, we counted the distance between relational entities, as shown in Table 1. Through the data, we
can find that Conjunction and Feature-of method types basically appear in close entities while other
relationship types are more average. By counting the percentage of different relationships appearing
before and after the entity type, we found that the entity type characteristics also have an important
influence on the relationship, here we take a table as an example, as shown in Table 2. Therefore, we
propose a semantically enhanced model. The main contributions of our work are as follows:
• We propose an improved joint model of NER and RE for academic texts, in which the type of
entity and the distance between candidate entities that potentially constitute the relationship are
considered in RE, and the token sequence length of an entity is considered in NER.
• We enrich the initial embeddings, the initial embeddings are augmented by semantic information
and syntactic information.
• Experiments on real data sets validate the effectiveness of our proposed method.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works</title>
      <p>With the development of technology, deep learning methods are becoming the focus of research in
the field of machine learning. Deep learning models focus more on the capability of machines and
abandon complex feature engineering, which greatly reduces the labor as well as time cost required for
extracting feature engineering compared to statistical models of machine learning. Therefore, deep
learning has become a very important research direction.</p>
      <p>The supervised deep learning method used in relation extraction can solve the main problems of
manual feature extraction and error propagation that exist in classical methods. The low-level features
are combined to form more abstract high-level features. At present, supervised relation extraction
methods mainly include pipeline approaches and joint approaches.
2.1</p>
    </sec>
    <sec id="sec-3">
      <title>Pipeline approaches</title>
      <p>Pipeline approaches refer to the extraction of relations between entities directly based on the entity
recognition already done. The early pipeline approaches mainly used two types of structures,
convolutional neural networks (CNNs) and recurrent neural networks (RNNs). Among them, CNNs with
diverse convolutional kernels are good for recognizing structural features of the target, while RNNs can
fully consider the dependency between long-range words and their memory function is good for
recognizing sequences. Zeng et al.[15] used CNN to extract word-level and sentence-level features for
the first time, and improved the accuracy of the relationship extraction model by using the hidden layer
and softmax layer for relationship classification. Socher et al.[16] first used an RNN approach for entity
relation extraction, which uses recurrent neural networks to syntactically parse the sentences in the
annotated text and then obtains a vector representation considering the syntactic structure of the sentences
after continuous iterations.</p>
      <p>As the research progresses, CNN and RNN methods are continuously improved and refined, and
many variants are generated, such as long short-term memory (LSTM), and bidirectional long short-term
memory (Bi- LSTM), which are improved to solve the gradient disappearance. Xu et al.[18] proposed
an LSTM-based relation extraction method based on the shortest path of syntactic dependency analysis
tree, incorporating features such as word vector, part of speech tags, WordNet, and syntax, using
maximum pooling layer, softmax layer, etc. In addition, with the application of graph convolutional
network (GCN) in the field of natural language processing, GCN has been increasingly used for mining
and exploiting potential information between entities, providing new ideas for solving relation overlap
and entity overlap, and thus further promoting the development of relation extraction. Schlichtkrull et
al.[19] proposed the use of relational graph convolutional neural networks (R-GCNs) on two standard
knowledge bases to accomplish link prediction and entity classification, respectively, where link
prediction extracts missing relations and entity classification completes the missing attributes of entities;
Zhang et al.[20] proposed an extended graph convolutional neural network that can effectively handle
arbitrary dependency structures in parallel and facilitate the extraction of entity relationships to
effectively utilize negative class data.</p>
      <p>Although the Pipeline approach is easy to implement, the entity model and the relation model can use
independent datasets and do not need to label both entity and relation datasets, there are several
disadvantages:</p>
      <p>a) error accumulation: errors in entity extraction will affect the performance of the next step of
relation extraction.</p>
      <p>Table 1: Relation entity distance of SciERC.</p>
      <p>Ave[&gt;11]*4
0.6%
0.3%
3.0%
1.1%
2.1%
4.5%
3.1%
All
b) entity redundancy: the redundant information brought by the candidate entity pairs without
relationships will enhance the error rate and increase the computational complexity, since the extracted
entities are firstly paired, and then extract relations between entity pairs.</p>
      <p>c) lacking interaction: the intrinsic connection and dependency between the two tasks are ignored.
2.2</p>
    </sec>
    <sec id="sec-4">
      <title>Joint approaches</title>
      <p>Joint approaches can further use the potential information between the two tasks to mitigate the
disadvantage of error propagation. The difficulty of joint approaches is how to enhance the interaction
between the entity model and the relation model. In early works, the connection method relied heavily
on fine-grained feature engineering to establish the interaction between NER and RE[21-23]. Recently,
end-to-end neural networks have proven successful in extracting relational triples[15,24-26], becoming
the mainstream for joint entity and relation extraction.</p>
      <p>Based on their differences in encoding task-specific features, most existing approaches can be divided
into two categories: sequential encoding and parallel encoding. Sequential encoding generally encodes
task features in the sequential order of NER and then RE, and this encoding approach can keep the later
encoded features from affecting the first encoded features directly away, resulting in unbalanced
intertask interactions. Zeng et al.[27] and Wei et al.[28] are typical examples of this category. They extracted
features for different tasks in a predefined order. Parallel encoding uses two independent encoders to
generate task features, which have no interaction other than shared input, leading to insufficient
intertask interaction, and in contrast to sequential encoding, models built based on this scheme do not need
to worry about the effect of encoding order. , and encoded the entity and relationship information
separately, and finally completed the extraction of task-specific features in two separate submodels,
respectively. Both encoding methods have their own drawbacks, the inter-task
interaction, and in contrast to sequential encoding, models built based on this scheme do not need to
worry about the effect of encoding order, and encoded the entity and relation information separately, and
finally completed the extraction of task-specific features in two separate submodels, respectively. Both
encoding methods have their own drawbacks, the inter-task interaction in sequential encoding is
oneway with a specific order, while the problem with parallel encoding is that they only retain the shared
features and actively ignore the features that are task-beneficial separately for each task.</p>
    </sec>
    <sec id="sec-5">
      <title>3. Model 3.1</title>
    </sec>
    <sec id="sec-6">
      <title>Model Architecture</title>
      <p>In this section, we provide a detailed explanation of our model, the framework is shown in Figure 2.
We develop a joint model including five components: 1) an embedding layer, which converts tokens into
embedding vectors, 2) a POS encoder, which converts tokens into part-of-speech, 3)a fusion module,
which fuses word embedding and part-of-speech embedding into one vector,4)a shallow entity classifier,
which classifies any possible sequence of consecutive tokens and 5)a shallow relationship classifier,
which classifies relation for any given set of entity pairs.
3.2</p>
    </sec>
    <sec id="sec-7">
      <title>Embedding Layer</title>
      <p>Given a sequence of sentences  = { ,  , … ,  }, the embedding layer transforms the sentences into
the vector matrix, in which each token in the sentence is represented by a pre-trained embedding. The
embedding of each token consists of two parts: pre-trained transformer and POS embeddings.</p>
      <p>a) Pretrained transformer: The recognition and relationship extraction of academic terminology
entities are different from the conventional entity relationship extraction, which is more specific and the
relation between terms is more abstract. Traditional methods are based on manually generated features,
while pre-training techniques are now widely used in deep learning, which have achieved good
performance in computer vision, natural language processing, and other fields. Usually, a model that has
been trained on large-scale data can achieve satisfactory performance with simple training, i.e.,
finetuning. Obtaining a high-quality initial value of parameters with the help of pre-training techniques not
only reduces the training burden but also helps to improve the model generalization ability.</p>
      <p>However, as Bert received pre-training on general texts from Wikipedia and book corpora, its
performance on domain-specific tasks proved to be suboptimal in several previous works [29-30]. These
empirical findings have driven the development of domain-specific pre-trained language models. For
example, SciBERT in the scientific domain and BioBERT and ALBERT in the biomedical domain. so
the domain-specific pre-trained model SciBERT and BioBERT was used for domain academic entities,
through which academic entity features were obtained to obtain better feature representations.</p>
      <p>
        We split each sentence  into a sequence of tokens  = {  ,  ,  , … ,  ,  }, where [CLS]
and [SEP] are special symbols. The [CLS] captures the contextual information of the text, while the [SEP]
acts as a separator to separate adjacent sentences between them. We use transformer to generate
pretrained embeddings as in (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ):
Whereb , and  1 is the embedding dimension.
      </p>
      <p>b) POS Tag: Part of speech is the important information a word carries, which reflects the components
that the word plays. Dependency syntactic parse in natural language processing is the conversion of
sentences into trees based on POS tags. Xu et al.[18] proposed an LSTM-based relation extraction
method based on the shortest path of the syntactic dependency analysis tree, incorporating features such
as word vector, part-of-speech, WordNet, and syntax, and using maximum pooling layer, softmax layer,
etc. for relation classification. The addition of POS ultimately achieves the goal of effect enhancement.
So we take POS tags into consideration. We generate part of speech tags for the input sentences and
assign the POS tag of the parent word to each child word tag it generates. We use a directed embedding
matrix to generate a embedding sequence  = {  ,  ,  , … ,  ,  }for each positional tag of
dimension d2. The BERT embedding of the token and the lexical POS embedding are then stitched
together to obtain a new vector representation of dimension. We use a directed embedding matrix to
generate embeddings for each positional tag of dimension d2.</p>
      <p>=</p>
      <p>, , ,⋅⋅⋅, ,
where the output embedding is a combination of the above two vectors. The dimension of the output
embedding is  +  .
3.3</p>
    </sec>
    <sec id="sec-8">
      <title>Span Classification</title>
      <p>To detect entities, a
representation  ,⋅⋅⋅,
entity.</p>
      <p>= max
 ,⋅⋅⋅,

where , and  1+  2 is the embedding dimension.</p>
      <p>
        In order to study the influence of the length of entity span, we counted the lengths of all entity spans,
as shown in Table 3. Overall, the percentage of entity span lengths on the interval (
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ) is more than half,
reaching 58.7%, and the data on the intervals (
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4-6</xref>
        ) (
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7-10</xref>
        ) and (&gt;10) show that the possibility of entity
span becoming an entity is inversely proportional to the length. In terms of entity type breakdown, there
is also a large difference in the proportion of different entity types in each interval, for example, Generic
type has 96.9% on the (
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ) interval, which can be almost considered as Generic type only in the (
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        )
interval, while Method and Task types are relatively evenly distributed. Thus entity span length is an
important feature in entity classification. We train a specific width embedding matrix   to obtain
an embedding for a span of length k.
      </p>
      <p>where  , and  1+  2+  3 is the embedding dimension.</p>
      <p>Finally,[CLS] , which represents the sentence context, is concatenated with ′  to obtain the vector
″  , which is passed through a softmax classifier to predict the entity type.</p>
      <p>vector is obtained by doing max-pooling of the embedding
of a sequence of tokens of length k of each successive possible constituent
′</p>
      <p>=   ∘  
″ 
= ′  ∘</p>
      <p>=  ⋅ ″</p>
      <p>
        + 
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
Where  

, 2 1+  2+  3 is the embedding dimension. And 
+  ,“+1” is due to the
‘null’ entity  that denotes the absence of an entity.
      </p>
      <p>Those spans that are classified as  by the entity classifier are filtered out. For the remaining spans,
the task is to identify the relation between every pair of them. Consider a pair of spans  ,
occurs before  2 in the input sentence. We assume relations to be asymmetric, so the relation directed
from  1 to  2 may be different from that directed from  2 to  1, and each of them must be separately
where  1
classified. We take the representations, 
,⋅⋅⋅,</p>
      <p>
        , where c is the embedding of the first token following
 1 and c is that of the last token preceding  2 in the sentence, and max-pooling:
 2 and s2 to s1 are separately encoded as in (
        <xref ref-type="bibr" rid="ref8">8</xref>
        ),(
        <xref ref-type="bibr" rid="ref9">9</xref>
        ):
      </p>
      <p>Where 
, and  1+  2 is the embedding dimension.The candidate relations from span  1 to
  ,
= max</p>
      <p>,⋅⋅⋅, 


= ′ 
∘   ,</p>
      <p>∘ ′  
= ′ 
∘   ,</p>
      <p>∘ ′</p>
      <p>The results are passed through a simple classifier with a confidence interval  (results beyond  are
considered to have this relationship) and a sigmoid activation function to predict the type of relation. We
have tried to include logical information and distance between candidate entities to predict entity type.
candidate entities.</p>
      <p>
        We train a specific width embedding matrix  
to obtain an embedding for the distance between
(
        <xref ref-type="bibr" rid="ref7">7</xref>
        )
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        )
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
(
        <xref ref-type="bibr" rid="ref12">12</xref>
        )
(
        <xref ref-type="bibr" rid="ref13">13</xref>
        )
(
        <xref ref-type="bibr" rid="ref14">14</xref>
        )




 = 
′ = 
′ = 
″ = 
″ = 
⋅
∘  
∘  
′ ∘  
′ ∘  
′+
∘   
∘
      </p>
      <p>The loss function of the joint model is the sum of the cross-entropy losses of the entity classifier and
the relational classifier. End-to-end training is performed by back-propagation, and the transformer is
fine-tuned during the training process. To train the entity classifier, we used real entity spans as positive
samples and added some non-entity spans as negative samples.</p>
    </sec>
    <sec id="sec-9">
      <title>4. Experiments And Results</title>
    </sec>
    <sec id="sec-10">
      <title>4.1 Datasets</title>
      <p>Our goal is to extract method entities and relations from the scientific literature, so we evaluated our
model on SciERC, a dataset from the scientific literature that contains both entities</p>
      <p>Length of tokenized entities
12.9%
12.7%
4.3%
9.4%
8.7%
1.1%
9.0%
[&gt;10]
8.3%
7.3%
3.0%
4.2%
2.8%
0.0%
4.5%</p>
      <p>Ave[&gt;10]*4
1.8%
1.4%
1.7%
1.9%
0.7%
0.0%
0.6%
and relations.</p>
      <p>The SciERC dataset is constructed from 500 abstracts of papers in the field of artificial intelligence,
with a total of 2687 sentences. It contains six scientific entities as well as seven relations. The six
scientific entities are Task, Method, Metric, Material, Other-Scientific-Term, and Generic, while the
seven methods are Compare, Conjunction, Evaluate-For, Used-For, Feature-of, Part-of, and
Hyponym-of. We follow the official cutoff methods: train (1861), dev (275), and test (551).</p>
    </sec>
    <sec id="sec-11">
      <title>4.2 Evaluation Metrics</title>
      <p>
        We use the standard Precision (P)(
        <xref ref-type="bibr" rid="ref15">15</xref>
        ), Recall (R)(
        <xref ref-type="bibr" rid="ref16">16</xref>
        ), and F1-score(
        <xref ref-type="bibr" rid="ref17">17</xref>
        ) to evaluate the model
performance.
      </p>
      <p>stand for true positive, false positive, and false negative, respectively.
where  ,</p>
      <p>and</p>
    </sec>
    <sec id="sec-12">
      <title>4.3 Results</title>
      <p>Performance on SciERC. We report the performance on the SciERC dataset in Table 4. We compare
the experimental results for six different models, and the F1 values for all three tasks improve compared
to the baseline SPERT. Compared to the 0.36% F1 value improvement for the NER task, the performance
improvement for the RE task is relatively significant, with Boundaries RE and Strict RE improving by
1.26% and 2.23%, respectively.</p>
    </sec>
    <sec id="sec-13">
      <title>4.4 Ablation study</title>
      <p>The ablation study in Table 5 shows the effect of removing entity span length, entity type, and relation
distance on the final classification score. In the ablation experiments, we take the average of the three
best results out of 20 experiments. Intending to compare the upper limits of the effect of feature. We
observe that removing the entity span length decreases the F1 value of NER by 0.4% and has an impact
on the subsequent RE tasks as well. Removing the entity type feature reduced the F1 scores of boundaries
RE and strict RE by 1.3% and 1.8%, respectively, and removing the relation distance reduced the F1
scores of boundaries RE and strict RE by 1.16% and 1.0%, respectively, so we can conclude that entity
 =
 =
1 =</p>
      <p>∗∗
type has a significant effect on relation extraction, especially on strict RE, while as the similar distance
feature, the improvement of relationship distance for RE is significantly higher than that of entity span
for NER.</p>
    </sec>
    <sec id="sec-14">
      <title>5. Conclusion</title>
      <p>We propose a semantically enhanced deep-learning model for extracting entities and relations from
the scientific literature. We explored the impact of linguistic information on entity and relation extraction
tasks, for which we added additional lexical information to the word vectors obtained through the
pretrained model to highlight lexical information, which proved to be superior to the pre-trained word
vectors alone. Individually, in the entity recognition part, the length of the entity's token sequence is
considered as a feature, while in the relation extraction part, entity type and relation distance are added,
which also improves the accuracy of the task. In the future, extending the in-sentence feature information
to inter-sentence contextual information is a promising challenge.</p>
    </sec>
    <sec id="sec-15">
      <title>6. Funding Statement:</title>
      <p>This research was supported by the Young Scientists Fund of the National Natural Science
Foundation of China under Grant[71904078], Natural Science Foundation of Jiangsu Province of China
under Grant[BK20190793], the Postgraduate Research and Practice Innovation Program of Jiangsu
Province of China under Grant[GJFXW21001].</p>
    </sec>
    <sec id="sec-16">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Jinha</surname>
            <given-names>A E</given-names>
          </string-name>
          .
          <article-title>Article 50 million: an estimate of the number of scholarly articles in existence</article-title>
          [J].
          <source>Learned publishing</source>
          ,
          <year>2010</year>
          ,
          <volume>23</volume>
          (
          <issue>3</issue>
          ):
          <fpage>258</fpage>
          -
          <lpage>263</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Hassan</surname>
            <given-names>S U</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Safder</surname>
            <given-names>I</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akram</surname>
            <given-names>A</given-names>
          </string-name>
          , et al.
          <article-title>A novel machine-learning approach to measuring scientific knowledge flows using citation context analysis[J]</article-title>
          .
          <source>Scientometrics</source>
          ,
          <year>2018</year>
          ,
          <volume>116</volume>
          (
          <issue>2</issue>
          ):
          <fpage>973</fpage>
          -
          <lpage>996</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Xie</surname>
            <given-names>I</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Babu</surname>
            <given-names>R</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee T H</surname>
          </string-name>
          , et al.
          <article-title>Enhancing usability of digital libraries: Designing help features to support blind and visually impaired users</article-title>
          [J].
          <source>Information Processing &amp; Management</source>
          ,
          <year>2020</year>
          ,
          <volume>57</volume>
          (
          <issue>3</issue>
          ):
          <fpage>102110</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Ding</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stirling</surname>
            <given-names>K</given-names>
          </string-name>
          .
          <article-title>Data-driven discovery: A new era of exploiting the literature and data[J]</article-title>
          .
          <source>Journal of Data and Information Science</source>
          ,
          <year>2016</year>
          ,
          <volume>1</volume>
          (
          <issue>4</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Bhatia</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitra</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles C L .</surname>
          </string-name>
          <article-title>Finding algorithms in scientific articles[C]//</article-title>
          <source>Proceedings of the 19th International Conference on World Wide Web, WWW</source>
          <year>2010</year>
          , Raleigh, North Carolina, USA, April
          <volume>26</volume>
          -
          <issue>30</issue>
          ,
          <year>2010</year>
          . ACM,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Katsurai</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joo</surname>
            <given-names>S.</given-names>
          </string-name>
          <article-title>Adoption of Data Mining Methods in the Discipline of Library and Information Science[J]</article-title>
          .
          <source>Journal of Library &amp; Information Studies</source>
          ,
          <year>2021</year>
          ,
          <volume>19</volume>
          (
          <issue>1</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Lam</surname>
            <given-names>C</given-names>
          </string-name>
          , Lai F C,
          <string-name>
            <surname>Wang C H</surname>
          </string-name>
          , et al.
          <article-title>Text mining of journal articles for sleep disorder terminologies[J]</article-title>
          .
          <source>PloS one</source>
          ,
          <year>2016</year>
          ,
          <volume>11</volume>
          (
          <issue>5</issue>
          ):
          <fpage>e0156031</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Li</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yan</surname>
            <given-names>E.</given-names>
          </string-name>
          <article-title>Co-mention network of R packages: Scientific impact and clustering structure[J]</article-title>
          .
          <source>Journal of Informetrics</source>
          ,
          <year>2018</year>
          ,
          <volume>12</volume>
          (
          <issue>1</issue>
          ):
          <fpage>87</fpage>
          -
          <lpage>100</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Wang</surname>
            <given-names>Y</given-names>
          </string-name>
          , Zhang C.
          <article-title>Finding more methodological entities from academic articles via iterative strategy: A preliminary study</article-title>
          [J].
          <source>training</source>
          ,
          <year>2019</year>
          ,
          <volume>2787</volume>
          :
          <fpage>2</fpage>
          .
          <fpage>73</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Zhu</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            <given-names>J</given-names>
          </string-name>
          .
          <article-title>Discovering Relationships between Data Structures and Algorithms</article-title>
          [J].
          <source>J. Softw.</source>
          ,
          <year>2013</year>
          ,
          <volume>8</volume>
          (
          <issue>7</issue>
          ):
          <fpage>1726</fpage>
          -
          <lpage>1735</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Duck</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nenadic</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brass</surname>
            <given-names>A</given-names>
          </string-name>
          , et al.
          <article-title>bioNerDS: exploring bioinformatics' database and software use through literature mining[J]</article-title>
          .
          <source>BMC bioinformatics</source>
          ,
          <year>2013</year>
          ,
          <volume>14</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Petasis</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cucchiarelli</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Velardi</surname>
            <given-names>P</given-names>
          </string-name>
          , et al.
          <article-title>Automatic adaptation of Proper Noun Dictionaries through cooperation of machine learning and probabilistic methods[C]//</article-title>
          <source>Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval</source>
          .
          <year>2000</year>
          :
          <fpage>128</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Alfonseca</surname>
            <given-names>E</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manandhar</surname>
            <given-names>S.</given-names>
          </string-name>
          <article-title>An unsupervised method for general named entity recognition</article-title>
          and
          <source>automated concept discovery[C]//Proceedings of the 1st international conference on general WordNet</source>
          , Mysore, India.
          <year>2002</year>
          :
          <fpage>34</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Borrega</surname>
            <given-names>O</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Taulé</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martı</surname>
            <given-names>M A</given-names>
          </string-name>
          .
          <article-title>What do we mean when we speak about Named Entities</article-title>
          [C]//Proceedings of Corpus Linguistics.
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Zeng</surname>
            <given-names>Daojian</given-names>
          </string-name>
          , Liu Kang,
          <string-name>
            <given-names>Lai</given-names>
            <surname>Siwei</surname>
          </string-name>
          , et al．
          <article-title>Relation classification via convolutional deep neural network [C]∕∕</article-title>
          <source>Proc of the 25th Int Conf on Computational Linguistics．Stroudsburg: ACL</source>
          ,
          <year>2014</year>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Socher</surname>
            <given-names>R</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huval</surname>
            <given-names>B</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            <given-names>C D</given-names>
          </string-name>
          , et al. Semantic
          <string-name>
            <surname>Compositionality through Recursive MatrixVector Spaces</surname>
          </string-name>
          [C]// Joint Conference on Empirical Methods in
          <source>Natural Language Processing &amp; Computational Natural Language Learning</source>
          .
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Luan</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ostendorf</surname>
            <given-names>M</given-names>
          </string-name>
          , et al.
          <article-title>Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction[J]</article-title>
          .
          <source>arXiv preprint arXiv:1808.09602</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Xu</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            <given-names>S</given-names>
          </string-name>
          , et al.
          <article-title>Semantic Relation Classification via Convolutional Neural Networks with Simple Negative Sampling</article-title>
          [J].
          <source>Computer Science</source>
          ,
          <year>2015</year>
          ,
          <volume>71</volume>
          (
          <issue>7</issue>
          ):
          <fpage>941</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Schlichtkrull</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kipf</surname>
            <given-names>T N</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bloem</surname>
            <given-names>P</given-names>
          </string-name>
          , et al.
          <article-title>Modeling Relational Data with Graph Convolutional Networks</article-title>
          [J].
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Zhang</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            <given-names>W . Attention</given-names>
          </string-name>
          <article-title>Guided Graph Convolutional Networks for Relation Extraction</article-title>
          [J].
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Yu</surname>
            <given-names>X</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lam</surname>
            <given-names>W .</given-names>
          </string-name>
          <article-title>Jointly identifying entities and extracting relations in encyclopedia text via a graphical model approach</article-title>
          .
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Qi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Ji H . Incremental Joint Extraction of Entity Mentions</article-title>
          and Relations[C]// Meeting of the Association for Computational Linguistics.
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Miwa</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sasaki</surname>
            <given-names>Y</given-names>
          </string-name>
          .
          <article-title>Modeling joint entity and relation extraction with table representation[C]//Proceedings of the 2014 conference on empirical methods in natural language processing</article-title>
          (EMNLP).
          <year>2014</year>
          :
          <fpage>1858</fpage>
          -
          <lpage>1869</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Gupta</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eral H B</surname>
            , Hatton
            <given-names>T A</given-names>
          </string-name>
          , et al.
          <article-title>Nanoemulsions: formation, properties and applications</article-title>
          [J].
          <source>Soft matter</source>
          ,
          <year>2016</year>
          ,
          <volume>12</volume>
          (
          <issue>11</issue>
          ):
          <fpage>2826</fpage>
          -
          <lpage>2841</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Katiyar</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cardie</surname>
            <given-names>C</given-names>
          </string-name>
          .
          <article-title>Going out on a limb: Joint extraction of entity mentions and relations without dependency trees[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</article-title>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers).</given-names>
          </string-name>
          <year>2017</year>
          :
          <fpage>917</fpage>
          -
          <lpage>928</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Shen</surname>
            <given-names>X</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McDanal</surname>
            <given-names>C</given-names>
          </string-name>
          , et al.
          <source>SARS-CoV-2 variant B. 1.1</source>
          .
          <article-title>7 is susceptible to neutralizing antibodies elicited by ancestral spike vaccines[J]</article-title>
          .
          <source>Cell host &amp; microbe</source>
          ,
          <year>2021</year>
          ,
          <volume>29</volume>
          (
          <issue>4</issue>
          ):
          <fpage>529</fpage>
          -
          <lpage>539</lpage>
          .
          <year>e3</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Wu</surname>
            <given-names>Q</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeng</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            <given-names>R</given-names>
          </string-name>
          .
          <article-title>Joint trajectory and communication design for multi-UAV enabled wireless networks[J]</article-title>
          .
          <source>IEEE Transactions on Wireless Communications</source>
          ,
          <year>2018</year>
          ,
          <volume>17</volume>
          (
          <issue>3</issue>
          ):
          <fpage>2109</fpage>
          -
          <lpage>2121</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Wei</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>WaNg</surname>
            <given-names>Y</given-names>
          </string-name>
          , et al.
          <article-title>A Novel Cascade Binary Tagging Framework for Relational Triple Extraction[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</article-title>
          .
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Beltagy</surname>
            <given-names>I</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lo</surname>
            <given-names>K</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohan</surname>
            <given-names>A</given-names>
          </string-name>
          .
          <article-title>SciBERT: A pretrained language model for scientific text</article-title>
          [J].
          <source>arXiv preprint arXiv:1903.10676</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Lee</surname>
            <given-names>J</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yoon</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            <given-names>S</given-names>
          </string-name>
          , et al.
          <article-title>BioBERT: a pre-trained biomedical language representation model for biomedical text mining</article-title>
          [J].
          <source>Bioinformatics</source>
          ,
          <year>2020</year>
          ,
          <volume>36</volume>
          (
          <issue>4</issue>
          ):
          <fpage>1234</fpage>
          -
          <lpage>1240</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Zhong</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            <given-names>D.</given-names>
          </string-name>
          <article-title>A frustratingly easy approach for entity and relation extraction[J]</article-title>
          .
          <source>arXiv preprint arXiv:2010.12812</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Luan</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wadden</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            <given-names>L</given-names>
          </string-name>
          , et al.
          <article-title>A general framework for information extraction using dynamic span graphs[J]</article-title>
          .
          <source>arXiv preprint arXiv:1904.03296</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Wadden</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wennberg</surname>
            <given-names>U</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            <given-names>Y</given-names>
          </string-name>
          , et al.
          <article-title>Entity, relation, and event extraction with contextualized span representations[J]</article-title>
          .
          <source>arXiv preprint arXiv:1909.03546</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Yan</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            <given-names>C</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fu</surname>
            <given-names>J</given-names>
          </string-name>
          , et al.
          <article-title>A partition filter network for joint entity and relation extraction[J]</article-title>
          .
          <source>arXiv preprint arXiv:2108.12202</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Eberts</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ulges</surname>
            <given-names>A</given-names>
          </string-name>
          .
          <article-title>Span-based joint entity and relation extraction with transformer pre-training[J]</article-title>
          .
          <source>arXiv preprint arXiv:1909.07755</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>