<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>CLEF</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>UB ET at CheckThat! 2020: Exploring Ad hoc Retrieval Approaches in Veri ed Claims Retrieval</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Edwin Thuma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Motlogelwa Nkwebi Peace</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Leburu-Dingalo Tebo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mudongo Monkgogi</string-name>
          <email>mudongomg@ub.ac.bw</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Botswana</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>22</volume>
      <fpage>22</fpage>
      <lpage>25</lpage>
      <abstract>
        <p>In this paper, we explore three di erent ad hoc retrieval approaches to rank veri ed claims, so that those that verify the input claim are ranked on top. In particular, we deploy DPH Divergence from Randomness (DFR) term weighting model to rank the veri ed claims. In addition, we deploy the Sequential Dependence (SD) variant of the Markov Random Fields (MRF) for term dependence to re-rank documents (veried claims) that have query terms (input claim) in close proximity. Moreover, we deploy LambdaMART, which is a learning to rank algorithm that use machine learning techniques to learn an appropriate combination of features into an e ective ranking model.</p>
      </abstract>
      <kwd-group>
        <kwd>Check-Worthiness</kwd>
        <kwd>Claim Retrieval</kwd>
        <kwd>Proximity Search</kwd>
        <kwd>Learning to Rank</kwd>
        <kwd>Ad-hoc Retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Information posted on social media platforms such as Twitter is not often
factchecked by an authoritative entity before being published [
        <xref ref-type="bibr" rid="ref2">2, 11</xref>
        ]. In some
instances, these posts on social media are coming from unreliable sources whose
main objective is to disinform the general public. Such an action often yields
undesirable results. For example, disinformation is often used in political campaigns
in order to in uence the outcome of political elections. It is for this reason that
the Information Retrieval (IR) and the natural language processing community
have invested signi cant e ort in developing techniques to address
disinformation, misinformation, factuality and credibility [
        <xref ref-type="bibr" rid="ref2">2, 11</xref>
        ]. This is evidenced by the
CheckThat! lab1, which is running under the Conference and Labs of the
Evaluation Forum (CLEF)2. The CheckThat! Lab at CLEF 2020 is the third version
of the lab. The other editions are the CheckThat! 2018 and CheckThat2019. The
main purpose of these labs is to foster research in the development of techniques
that would enable identi cation and veri cation of claims. In this paper, we
present the results of our participation to the CheckThat! 2020 Task 2: Claim
Retrieval, where we explore three di erent ad hoc retrieval approaches to rank
veri ed claims, so that those that verify the input check-worthy tweet are ranked
on top.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Background</title>
      <p>In this section, we present a brief but essential background on the di erent
adhoc retrieval approaches used in our investigation. In particular, we start by
providing a description of the DPH term weighting model in Section 2.1. This
is followed by a description of the learning to rank techniques in Section 2.2.
2.1</p>
      <sec id="sec-2-1">
        <title>DPH Term Weighting Model</title>
        <p>
          For all our experimental investigation, we used the parameter-free DPH term
weighting model from the Divergence from Randomness (DFR) framework [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
The DPH term weighting model calculates the score of a document d for a given
query Q as follows:
scoreDP H(d; Q) = Pt2Q qtf norm tf log((tf avlg l ) ( tNfc )) + 0:5 log(2 tf (1 tMLE))
(1)
where qtf , tf and tf c are the frequencies of the term t in the query Q , in the
document d and in the collection C respectively. N is number of documents in
the collection C, avg l is the average length of documents in the collection C
and l is the length of the document d. tMLE = tlf and norm = (1 tMLE)2 .
tf+1
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Learning to Rank Approach</title>
        <p>
          Learning to rank techniques are algorithms that use machine learning
techniques to learn an appropriate combination of features into an e ective
ranking model [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. This e ective ranking model can be leant through the following
steps [
          <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
          ]:
1. Top K retrieval: Using a set of training queries that have relevance
assessment, retrieve a sample of k documents using an initial weighting model such
as DPH.
2. Feature extraction: For each document in the retrieved sample, extract a set
of features. These features can either be query-dependent (term weighting
models, term dependence models) or query-independent (click count, fraction
of stopwords). The feature vector for each document is labelled according to
the already existing relevance judgements.
3. Learning: Learn an e ective ranking model by deploying an e ective learning
to rank technique on the feature vectors of the top k documents.
The learned model can be deployed in a retrieval setting as follows:
4. Top K retrieval: For each unseen query, the top k documents are retrieved
using the same retrieval strategy as in step (1)
5. Feature extraction: A set of features are extracted for each document in the
sample of k documents. These features should be the same as those extracted
in step (2).
6. Re-rank the documents: Re-rank the documents for the query by applying a
learned model on every feature vector of the documents in the sample. The
nal ranking of the documents is obtained by sorting the predicted scores in
descending order.
        </p>
        <p>
          In this work, we deploy LambdaMART [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], which is a tree-based learner.
A tree-based learner builds a set of regression trees T . The nal score of a
document d is obtained by traversing the nodes of a particular tree t, according
to the decisions based on the vector of feature values of the document fd [
          <xref ref-type="bibr" rid="ref3 ref6">3, 6</xref>
          ].
The leaf node of the tree traversed represents the nal score of the document d.
This can be expressed as:
score(d; Q) =
        </p>
        <p>X t(fd)
t2T
(2)
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experimental Setting</title>
      <p>FAQ Retrieval Platform: For all our experiments, we used Terrier-4.23 [8],
an open source Information Retrieval (IR) platform. All the documents used in
this study were rst pre-processed before indexing and this involved tokenising
the text and stemming each token using the full Porter stemming algorithm [10].
We indexed the collection using blocks in order to save positional information
with each term.</p>
      <p>
        Training Learning to Rank Techniques: For our learning to rank approach,
we used the Terrier-4.2 Fat4[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] framework. Fat is a method of allowing many
features to be computed within one run of Terrier. To train and test LambdaMART,
we used the default parameter values of the algorithms.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Description of the Di erent Runs</title>
      <p>
        T2-EN-UB ET-DPH: For all our runs, we used the parameter-free DPH
Divergence From Randomness term weighting model in Terrier-4.2 IR platform to
score and rank the documents (veri ed claims)
T2-EN-UB ET-DPH LTR: We used T2-EN-UB ET-DPH as the baseline
system. As improvement, we deployed a learning to rank technique. For our
3 http://terrier.org/
4 http://terrier.org/docs/v4.0/learning.html
learning to rank technique, we used the training and development tweets with
their qrels for training and validation. We used the Terrier-4.2 Fat framework to
retrieve 1000 documents for each topic (tweet) using the DPH term weighting
model, and then calculated several additional query dependent features in
Table 1. Using these features, we used Jforests to learn a LambdaMART model. We
then applied this learned model on the test tweets to generate a nal ranking.
T2-EN-UB ET-DPH MRF: We used T2-EN-UB ET-DPH as the baseline
system. As improvement, we deployed the Sequential Dependence (SD) variant
of the markov random eld for term dependence [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to re-rank documents
(veri ed claims) that have query terms (input claim) in close proximity. Sequential
Dependence only assumes a dependence between neighbouring query terms.
5
      </p>
    </sec>
    <sec id="sec-5">
      <title>Results and Discussion</title>
      <p>T2-EN-UB ET-DPH
T2-EN-UB ET-DPH LTR
T2-EN-UB ET-DPH MRF
0.843
0.818
0.838
0.868
0.862
0.865
0.873
0.864
0.869
0.840 0.300 0.185
0.815 0.307 0.186
0.835 0.300 0.184</p>
      <p>Table 2 presents our evaluation results. The o cial evaluation measure for
Task 2: Claim Retrieval is MAP@k, where k = 5. We also present Precision@k.
The results of this study suggests that ad-hoc retrieval approaches such as term
weighting models, proximity (Dependence) models and learning to rank
techniques can be used to rank veri ed claims for a given check-worthy tweet.
Overall, our primary submission T2-EN-UB ET-DPH LTR ranked third out of
10 submissions. It is worth noting that an attempt to improve the retrieval
performance using a learning to rank technique resulted in a degradation in
performance. An examination of the data revealed that for a majority of
checkworthy tweets, there was a single veri ed claim. This lack o su cient training
data could have resulted in the degradation in retrieval performance. For
example, after performing a rst-pass retrieval with DPH and attempting to improve
the ranking with our learned ranking model, some veri ed claims that verify the
input claim ranked lower than in the previous ranking.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>In this paper, three di erent ad hoc retrieval approaches were explored to
determine their e ectiveness in raking veri ed claims so that those that verify the
input claim are ranked on top. The results of this study suggests that term
weighting models such as DPH can be used to rank veri ed claims for a given
check-worthy tweet. In our attempt to improve the retrieval e ectiveness using a
learning to rank technique, we noticed a degradation in retrieval performance. In
future, we will explore using su cient training data in our learning to rank
technique coupled with additional query dependent and query independent features.
Similarly, re-ranking the veri ed claims using markov random eld for term
dependence resulted in the degradation in performance. In our experiments, default
parameter settings were used. Further research could usefully explore using
different parameters settings such as varying the window size in order to improve
the retrieval performance.
8. I. Ounis, G. Amati, Plachouras V., B. He, C. Macdonald, and Johnson. Terrier
Information Retrieval Platform. In Proceedings of the 27th European Conference
on IR Research, volume 3408 of Lecture Notes in Computer Science, pages 517{519,
Berlin, Heidelberg, 2005. Springer-Verlag.
9. V. Plachouras and I. Ounis. Multinomial Randomness Models for Retrieval with
Document Fields. In Proceedings of the 29th European Conference on IR Research,
pages 28{39, Berlin, Heidelberg, 2007. Springer-Verlag.
10. M.F. Porter. An Algorithm for Su x Stripping. Readings in Information Retrieval,
14(3):313{316, 1997.
11. Shaden Shaar, Alex Nikolov, Nikolay Babulkov, Firoj Alam, Alberto
BarronCeden~o, Tamer Elsayed, Maram Hasanain, Reem Suwaileh, Fatima Haouari,
Giovanni Da San Martino, and Preslav Nakov. Overview of CheckThat! 2020 English:
Automatic identi cation and veri cation of claims in social media.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>G.</given-names>
            <surname>Amati</surname>
          </string-name>
          , E. Ambrosi,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bianchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gaibisso</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Gambosi. FUB</surname>
          </string-name>
          , IASICNR and University of Tor Vergata at
          <article-title>TREC 2007 Blog Track</article-title>
          .
          <source>In Proceedings of the 16th Text REtrieval Conference (TREC-2007)</source>
          , pages
          <fpage>1</fpage>
          {
          <fpage>10</fpage>
          ,
          <string-name>
            <surname>Gaithersburg</surname>
          </string-name>
          , Md., USA.,
          <year>2007</year>
          . Text REtrieval Conference (TREC).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Alberto</surname>
          </string-name>
          Barron-Ceden~o, Tamer Elsayed, Preslav Nakov, Giovanni Da San Martino, Maram Hasanain, Reem Suwaileh, Fatima Haouari, Nikolay Babulkov, Bayan Hamdan, Alex Nikolov, Shaden Shaar, and Zien Sheikh Ali. Overview of CheckThat! 2020:
          <article-title>Automatic identi cation and veri cation of claims in social media</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Yasser Ganjisa ar, Rich Caruana, and
          <string-name>
            <given-names>Cristina</given-names>
            <surname>Lopes</surname>
          </string-name>
          .
          <article-title>Bagging gradient-boosted trees for high precision, low variance ranking models</article-title>
          .
          <source>In Proceedings of the 34th international ACM SIGIR conference on Research and development in Information, SIGIR '11</source>
          , pages
          <fpage>85</fpage>
          {
          <fpage>94</fpage>
          , New York, NY, USA,
          <year>2011</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>T</surname>
          </string-name>
          .-Y. Liu.
          <article-title>Learning to Rank for Information Retrieval</article-title>
          .
          <source>Foundations and Trends in Information Retrieval</source>
          ,
          <volume>3</volume>
          (
          <issue>3</issue>
          ):
          <volume>225</volume>
          {
          <fpage>331</fpage>
          ,
          <year>June 2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>C.</given-names>
            <surname>Macdonald</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.L.</given-names>
            <surname>Santos</surname>
          </string-name>
          ,
          <string-name>
            <surname>and I. Ounis.</surname>
          </string-name>
          <article-title>The whens and hows of learning to rank for web search</article-title>
          .
          <source>Information Retrieval</source>
          ,
          <volume>16</volume>
          (
          <issue>5</issue>
          ):
          <volume>584</volume>
          {
          <fpage>628</fpage>
          ,
          <year>October 2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>C.</given-names>
            <surname>Macdonald</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.L.T.</given-names>
            <surname>Santos</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Ounis</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>He</surname>
          </string-name>
          .
          <article-title>About learning models with multiple query-dependent features</article-title>
          .
          <source>ACM Transactions on Information Systems (TOIS)</source>
          ,
          <volume>31</volume>
          (
          <issue>3</issue>
          ):
          <volume>11</volume>
          :1{
          <fpage>11</fpage>
          :
          <fpage>39</fpage>
          ,
          <year>August 2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Donald</given-names>
            <surname>Metzler</surname>
          </string-name>
          and
          <string-name>
            <given-names>W. Bruce</given-names>
            <surname>Croft</surname>
          </string-name>
          .
          <article-title>A markov random eld model for term dependencies</article-title>
          . In Ricardo A.
          <string-name>
            <surname>Baeza-Yates</surname>
          </string-name>
          , Nivio Ziviani, Gary Marchionini, Alistair Mo at, and John Tait, editors,
          <source>SIGIR 2005: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Salvador, Brazil,
          <source>August 15-19</source>
          ,
          <year>2005</year>
          , pages
          <fpage>472</fpage>
          {
          <fpage>479</fpage>
          . ACM,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>