<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>IRIT at CheckThat! 2018</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Romain Agez</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Clement Bosc</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cedric Lespagnol</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Josiane Mothe</string-name>
          <email>Josiane.Mothe@irit.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Noemie Petitcol</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ESPE, IRIT, UMR5505, CNRS &amp; Universite de Toulouse</institution>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universite P.</institution>
          <addr-line>Sabatier de Toulouse, UPS</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The 2018 CLEF CheckThat! is composed of two tasks: (1) Check-Worthiness and (2) Factuality. We participated to task (1) only which purpose is to evaluate the check-worthiness of claims in political debates. Our method to achieve this goal is to represent each claim by a vector of ve computed values that correspond to scores on ve criteria. These vectors are then used with machine learning algorithms to classify claims as check-worthy or not. We submitted three runs using di erent machine learning algorithms. The best result we achieved using the o cial measure MAP ranks our run that uses non linear SVM the 12th over the 16 submitted runs. Our run that uses linear SVMis ranked 2nd with the Mean Precision@1 measure.</p>
      </abstract>
      <kwd-group>
        <kwd>Information retrieval fact-checking tional label</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The CLEF CheckThat! rst task aims at predicting which claims in political
debates should be prioritized for fact-checking. All the background and detailed
information about the task are available on the task description paper provided
by the organizers of the task [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>To achieve this goal, the task organizers released several textual transcripts
of political debates with each sentence being annotated according to whether it
is check-worthy or not.</p>
      <p>This paper describes the participation of the Universite de Toulouse team
(o cial name RNCC) at CLEF 2018 CheckThat! pilot task for check-worthiness.</p>
      <p>We preprocessed the data by representing each sentence corresponding to a
transcription of what a speaker said in the debate by a vector containing the
score of this sentence for ve di erent criteria. We then trained three classi ers
using these vectors to submit three di erent runs.</p>
      <p>The remaining of this paper is organized as follows: Section 2 gives a
description of the pilot task. Section 3 details the model we developed and the submitted
runs. Then Section 4 details the results we obtained. Finally, Section 5 concludes
this paper.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Task Description</title>
      <sec id="sec-2-1">
        <title>Objectives</title>
        <p>
          The Check-Worthiness task aims to predict which statements in a political
debate should be fact-checked. Indeed, nowadays, information objects are
spreading faster and faster on the Internet and especially on social networks. This
spreading is named the virality of the information [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
        </p>
        <p>During a political debate, any of the statements made by the participants
can be reused without checking its factuality and it even can become viral.
CheckThat! aims at providing journalists with a list of statements members of
the debate made that should be checked before they are reused by others.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Dataset</title>
        <p>There are two datasets : one to train the model and one to test it. Both sets
consist of political debates transcribed into texts.</p>
        <p>
          They are annotated so that each row indicates the sentence number, the
speaker, the transcription of the sentence that the speaker said. The training
dataset includes in addition a label that indicates whether this sentence is to be
fact-checked or not. The training set contains three political debates while the
test set contains seven debates [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
2.3
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Evaluation metric</title>
        <p>
          The task has been evaluated according to di erent measures. The o cial measure
is MAP which calculates the usual mean of the average precision. Then, other
measures were used as Mean Reciprocal Rank which allows to obtain reciprocals
of rank of the rst relevant document as well as Mean Precision at x which
performs the average of x best candidates. Details on the measures used can be
found in the task overview [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>Evaluations are carried out on primary and contrastive runs. Primary run
corresponds to the results le of the participant's main model ; the decision of the
main run was the participant's decision. Contrastive runs match the secondary
models the participant used.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Method and runs</title>
      <p>
        We computed ve of the criteria from the Information Nutritional Label for
Online Documents proposed by [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These criteria and the methods we developed
in this work to calculate their score are as follows:
{ Factuality and Opinion : Determines whether a sentence represents a
fact or a personal opinion. These two features are based on the same
algorithm. Each value is the opposite of the other, it is either 0 or 1. We use a
Multi-layer Perceptron classi er, using LBFGS gradient descent [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. This
neural network is composed of 500 neurons in the rst hidden layer and
5 neurons in the second hidden layer. The activation function used is the
recti ed linear unit function ("relu"). We used a MLP classi er because it
was the best performing classi er over Random Forest, Support Vector
Machine and Linear Regression. The datasets to train the neural network come
from various Wikipedia articles3 for factual sentences and from Opinosis4 for
opinion sentences. The features used to classify a sentence are ne-grained
part-of-speech tags extracted with spaCy5.
{ Controversy : Determines the degree of controversy in a text. We count
the number of controversial issues in the text based on the Wikipedia Article
List of controversial issues6. For each issue referenced in the wiki article, we
also take in account the anchor text labels7 to nd the synonyms and other
appellations of the issues in all of the Wikipedia database. For example :
Donald Trump is in the list of controversial issues. Other names can link to
his Wikipedia page such as "45th President of America". These names are
called anchor text labels and will be recognized as a controversial issue.
{ Emotion : Determines the intensity of emotion in a sentence. We use the
list of 2; 477 emotional words and valuation from AFINN8 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] (ex : abusive
= -3, proud = 2). We sum the absolute value of the positive and negative
valuations of the emotional words found in the sentence and we divide it by
the total number of words in the sentence :
      </p>
      <p>
        (X posW ordV alue + X jnegW ordV aluej)=totalN umberW ords
{ Technicality : Determines the degree of technicality in a text. We count
the number of domain-speci c terms in the text. For that, we use NLTK9 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
to perform part of speech tagging (adjective = JJ, name = NN, etc.). Then,
we use the RE library10 to match, from tags, a regular expression de ned
in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] which identi es the terminological noun phrases (NPs). NPs represent
domain-speci c terms in the text. We extract all the NPs from the text and
3 Each of the following URL should be preceded by https://en.wikipedia.org/wiki
/World War I, /Industrial Revolution, /October Revolution, /Fermi paradox,
/Steam engine, /Barack Obama, /Amazon (company), /Netherlands,
/Triangular trade, /Song dynasty, /Nanking Massacre, /The Holocaust
4 http://kavita-ganesan.com/opinosis/
5 spaCy is a library for Natural Language Processing in Python. It provides NER,
POS tagging, dependency parsing, word vectors and more.
      </p>
      <p>https://spacy.io/
6 https://en.wikipedia.org/wiki/Wikipedia:List of controversial issues
7 https://en.wikipedia.org/wiki/Anchor text
8 http://www2.imm.dtu.dk/pubdb/p.php?6010
9 Natural Language ToolKit, https://www.nltk.org/
10 Regular Expression, https://docs.python.org/3/library/re.html
keep those which appear more than once. We then calculate the ratio of the
number of these NPs over the number of words in the text.</p>
      <p>(X N P s)=totalN umberW ords</p>
      <p>We decided to use only these criteria as features because our goal was to test
the Information Nutritional Label on a concrete task.
3.1</p>
      <sec id="sec-3-1">
        <title>Models</title>
        <p>Each of our three runs uses its own model to compute a check-worthiness score.
For each of our models, we preprocessed the data using the criteria previously
described. We computed the ve features for each sentence that has to be
evaluated for check-worthiness. These sentences are then represented by a vector
containing ve features, one for each criterion score.</p>
        <p>For our INL SVM RBF (primary run) and INL SVM Lin ( rst contrastive)
runs, we decided to use the Support Vector Machine in sklearn 11 with the
probability setting set to "True". We used a RBF kernel for INL SVM RBF
run and a linear kernel for the INL SVM Lin run. For our INL RF (second
contrastive) run, we used the random forest classi er in sklearn.</p>
        <p>To train our models, we used the three annotated debates provided by the
clef2018-factchecking github repository12.</p>
        <p>To obtain a score of check-worthiness, we computed the probability for each
sentence to be check-worthy using the classi ers. The score of a sentence was
then normalized by the highest score obtained for this sentence divided by the
highest probability computed, so that the scores are between 0 and 1.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>Seven teams submitted runs to this task for a total of 16 runs.</p>
      <p>
        Table 1 presents the results of our three runs and the best submitted run
according to the MAP measure, which is from the Copenhagen team [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
11 http://scikit-learn.org/stable/modules/svm.html
12 https://github.com/clef2018-factchecking/clef2018-factchecking/tree/master/data/task1/English
      </p>
      <p>Overall, the INL SVM Lin run obtained better results than the INL SVM RBF
run; that was somehow unexpected since non linear kernel have been shown to
work better in other information retrieval applications. The INL SVM Lin run
has been ranked twelfth according to the main measurement (Mean Average
Precision), but obtained better rank when considering other measures: it is ranked
fth according to the Mean Reciprocal Rank and second according to the Mean
Precision@1. These ranks mean that our INL SVM Lin run would be good if the
purpose of the task was nding the most check-worthy claim instead of nding
all the check-worthy claims. However, we need to deeper analyse the results to
understand why.</p>
      <p>Post-hoc experiments showed that the least important criterion is
Technicality. This may be due to the fact that the method we use to compute this
feature was meant to work with large texts and it is not appropriate for a single
sentence. The most important criterion is Emotion. We can assume that a claim
has greater chances to be check-worthy if it is highly emotional. The speaker
thinks less about what he says and it is more likely that his claims are not fully
accurate. We will check this hypothesis in future work.</p>
      <p>Table 2 presents the weight of the 5 features for our INL SVM Lin model.
The weights of the features for our INL RF model are similar.</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and perspectives for future works</title>
      <p>
        In this paper we proposed three models to solve the CLEF2018 CheckThat!
challenge (task 1 Check Worthiness) which deals with the evaluation of the
check-worthiness of statements in political debates. We used random forest and
support vector machine to learn models that make use of the Information
Nutritional Label features [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. We show that these models perform pretty well when
considering the Mean Precision@1 measure, which ranks our run that uses a
support vector machine with a linear kernel 2nd over 16 submitted runs.
      </p>
      <p>
        We are currently working on better calculation of the ve features. We would
like to complete the representations of the texts by using content-based
components like it is done in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. While the objective is di erent (virality prediction),
some of the features may also be useful for the task tackled by CheckThat!. To
improve more our models, we would also like to investigate the use of
wordembedding since we are using successfully this approach in other tasks [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and
this approach also worked well according to Hansen et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] in the CheckThat!
context. As future work, we will also take in consideration the sentences around
the one to be classi ed and who said these sentences.
      </p>
      <p>Finally, we will test these models on other datasets such as social networks.
For example, we will consider a Twitter-based dataset where each tweet would
have a score indicating its worthiness for fact-checking taking into account
hashtags and tweet sources.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Berger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milkman</surname>
            ,
            <given-names>K.L.</given-names>
          </string-name>
          :
          <article-title>What makes online content viral</article-title>
          ?
          <source>Journal of marketing research 49(2)</source>
          ,
          <volume>192</volume>
          {
          <fpage>205</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bird</surname>
          </string-name>
          ,
          <string-name>
            <surname>Steven</surname>
            ,
            <given-names>E.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klein</surname>
          </string-name>
          , E.:
          <article-title>Natural language processing with python (</article-title>
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Fuhr</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giachanou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grefenstette</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hanselowski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jarvelin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Liu,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Mothe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Nejdl</surname>
          </string-name>
          ,
          <string-name>
            <surname>W.</surname>
          </string-name>
          , et al.:
          <article-title>An information nutritional label for online documents</article-title>
          .
          <source>In: ACM SIGIR Forum</source>
          . vol.
          <volume>51</volume>
          , pp.
          <volume>46</volume>
          {
          <fpage>66</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simonsen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lioma</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The Copenhagen Team Participation in the Check-Worthiness Task of the Competition of Automatic Identi cation and Veri cation of Claims in Political Debates of the CLEF-2018 Fact Checking Lab</article-title>
          . In: Cappellato,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.Y.</given-names>
            ,
            <surname>Soulier</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . (eds.)
          <source>CLEF 2018 Working Notes, Conference and Labs of the Evaluation Forum. CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Avignon, France (
          <year>September 2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hoang</surname>
            ,
            <given-names>T.B.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mothe</surname>
          </string-name>
          , J.:
          <article-title>Predicting information di usion on twitter{ analysis of predictive features</article-title>
          .
          <source>Journal of Computational Science</source>
          (
          <year>2017</year>
          ), https://doi.org/10.1016/j.jocs.
          <year>2017</year>
          .
          <volume>10</volume>
          .010
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Justeson</surname>
            ,
            <given-names>J.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>S.M.:</given-names>
          </string-name>
          <article-title>Technical terminology: some linguistic properties and an algorithm for identi cation in text</article-title>
          .
          <source>Natural language engineering 1</source>
          (
          <issue>1</issue>
          ),
          <volume>9</volume>
          {
          <fpage>27</fpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramiandrisoa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>IRIT at TRAC</article-title>
          .
          <source>In: Proceedings of the First Workshop on Trolling, Aggression and Cyberbulling (TRAC)</source>
          ,
          <source>27th International Conference on Computational Linguistics, COLIN 18. International Committee on Computational Linguistics</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barron-Cedeno</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suwaileh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marquez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaghouani</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Atanasova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kyuchukov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Da San Martino, G.:
          <article-title>Overview of the CLEF2018 CheckThat! Lab on Automatic Identi cation and Veri cation of Political Claims</article-title>
          . In: Bellot,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Trabelsi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Mothe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Murtagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.Y.</given-names>
            ,
            <surname>Soulier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>SanJuan</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>Cappellato</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferro</surname>
          </string-name>
          , N. (eds.)
          <article-title>Experimental IR Meets Multilinguality, Multimodality, and Interaction</article-title>
          .
          <source>Proceedings of the Nineth International Conference of the CLEF Association (CLEF</source>
          <year>2018</year>
          ).
          <source>Lecture Notes in Computer Science (LNCS) 11018</source>
          , Springer, Heidelberg, Germany (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Nielsen</surname>
            ,
            <given-names>F.A.</given-names>
          </string-name>
          :
          <article-title>A nn (mar</article-title>
          <year>2011</year>
          ), http://www2.imm.dtu.dk/pubdb/p.
          <source>php?6010</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>