<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Hybrid Model to Rank Sentences for Check-worthiness</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rudra Dhar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Subhabrata Dutta</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dipankar Das</string-name>
          <email>dipankar.dipnil2005g@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Jadavpur University</institution>
          ,
          <addr-line>West Bengal</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we describe a system submitted to the CLEF 2019 CheckThat Shared Task. We implement an ensemble of a logistic regression model and an LSTM-based neural network model to predict the worthiness of a sentence for fact checking. Our key idea is to train two separate models with high precision and high recall on binary classi cation task, and then use the binary class probability as a check-worthiness score. Our system achieves a reciprocal rank 0.4419 and mean average precision of 0.1162 for ranking sentences according to their check-worthiness.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>With the advent of web technology, mass media have achieved a revolutionary
new form. People all around the globe share information with each other to
construct their opinion about everything, that too at an unprecedented speed.
In this age of communication, quality of information becomes utmost important.
Rumors, fake news, malicious tampering of reality can cause substantial amount
of economic as well as social calamity now more than ever. This makes
factchecking an essential part of media and information systems. To handle such
large amount of information sharing, automation of such a system is necessary
if not mandatory.</p>
      <p>
        As Hassan et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] suggest, automatic fact checking can be divided into a
two-fold task: identi cation of check-worthy sentences, and checking their
trustworthiness based on some reliable source. In this task, we attempt to build a
system which can assign a score to an input sentence indicating its check-worthiness.
This score can vary from 0 (not check-worthy) to 1 (fully check-worthy). The
organizers provided the required data set comprised of 16421 sentences. These
sentences are binary labelled, 0 and 1, corresponding to not check-worthy and
fully check-worthy respectively. Out of these 16421 sentences, 440 are labelled
as fully check-worthy. We insist on building a system using this training data
only. We attempt to build a classi er framework which assigns a probability
score to each sentence, and hypothesize that this probability score corresponds
to the check-worthiness of the sentence. Our framework is an ensemble of two
separately trained classi ers: one based on Long Short-Term Memory (LSTM)
networks, and another based on Logistic Regression classi er. Our key strategy
is to build one model predicting check-worthy sentences with high precision (less
false positives) and another one with high recall (less false negatives), so that
their ensemble will perform more accurately.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Logistic Regression Component</title>
      <p>
        To train our logistic regression model, we extract the following features for each
sentence:
1. Syntactic N-grams [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] computed using Syntactic N-gram builder1; we
generate the dependency parse tree of the sentence using Stanford CoreNLP [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
and compute syntactic n-grams for dependecy paths of length one (arc), two
(bi-arc), three (tri-arc) and four (quad-arc).
2. Subjectivity score of the sentence, using TextBlob2.
3. Cumulative Entropy of the sentence as,
      </p>
      <p>CE =
1</p>
      <p>
        X(tf (t) (log(jT j)
jT j t2T
log(tf (t)))
where T is the set of terms in the corpus and tf (t) is the frequency of term
t 2 T in the sentence.
4. Count of negative, neutral and positive polarity words computed using
SentiWordNet [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
5. LIX score [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] representing the readability of the sentence.
6. Count of named entities present in the sentence.
      </p>
      <p>With this feature set, we train a logistic regression classi er using
Scikitlearn.</p>
      <p>LSTM Cell
Fully connected
layer
. . .</p>
      <p>Final
check-worthiness
score</p>
      <p>Average
S1</p>
      <p>S2
Selected
features</p>
      <p>
        Logistic regression
Pretrained
word
embedding
POS
embedding
We use GloVe [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] pre-trained word embeddings of size 300 to be used as word
representations. Along with that, we input parts-of-speech tags of the words
(tagged using Stanford CoreNLP) as one-hot vectors, which selects randomly
initialized POS embeddings. These two embeddings are concatenated and input
to an LSTM layer, which learns a many-to-one mapping from the sequence of
word-POS tag vectors to a single representation of the sentence. We also use
the manually extracted features as an input to a fully connected layer, learning
a dense low dimensional representation from the sparse feature vectors. The
output of the LSTM layer and the fully connected layer is then concatenated
and input to another fully connected layer, which outputs a probability score.
1 https://github.com/jmnybl/syntactic-ngram-builder
2 https://textblob.readthedocs.io/en/dev/
As already stated, the LSTM component and logistic regression components are
trained separately. For a single sentence, if the output score of the LSTM model
is greater than 0:5, then the predicted label is 1, otherwise 0. We train the LSTM
model using the Adam optimizer, with a batch size equal to 256.
      </p>
      <p>As the training data-set has high class imbalance, we train the LSTM model
with a class-weighted binary cross-entropy loss. We weigh the positive class with
weight w = log NN10 where N0; N1 corresponds to the number of negative and
positive class samples in the training data. The class imbalance automatically
puts a bias in the logistic regression model towards the negative class, making its
positive predictions highly precise; on the other hand, the weighted loss function
forces the LSTM model to bias towards the positive class, resulting in a higher
recall.
4</p>
    </sec>
    <sec id="sec-3">
      <title>Testing</title>
      <p>To produce the nal check-worthiness score, we average the scores predicted
by the two components. We test our system on the test dataset provided by
the organizers. Test dataset contains 7080 sentences. In Table 3 we present the
evaluation results of our system for various metrics.
We made a system to predict the check-worthiness of a sentence for CLEF 2019
CheckThat (Task 1). We did not use any external data. Here we used a combined
logistic regression model and an LSTM-based neural network model to get a
better probability score for check-worthiness. We achieved a reciprocal rank of
0.4419 and a mean average precision of 0.1162 .</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Atanasova</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karadzhov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohtarami</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Da San Martino, G.:
          <article-title>Overview of the clef-2019 checkthat! lab on automatic identi cation and veri cation of claims. task 1: Check-worthiness</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Baccianella</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Esuli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sebastiani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Sentiwordnet 3.0: an enhanced lexical resource for sentiment analysis and opinion mining</article-title>
          .
          <source>In: Lrec</source>
          . vol.
          <volume>10</volume>
          , pp.
          <volume>2200</volume>
          {
          <issue>2204</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Bjornsson, C.H.:
          <article-title>Readability of newspapers in 11 languages</article-title>
          . Reading Research Quarterly pp.
          <volume>480</volume>
          {
          <issue>497</issue>
          (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Elsayed</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Barron-Ceden~o,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Hasanain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Suwaileh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            , Da San Martino, G.,
            <surname>Atanasova</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>Overview of the CLEF-2019 CheckThat!: Automatic Identi cation and Veri cation of Claims. In: Experimental IR Meets Multilinguality, Multimodality, and</article-title>
          <string-name>
            <surname>Interaction. LNCS</surname>
          </string-name>
          , Lugano, Switzerland (
          <year>September 2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hassan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adair</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hamilton</surname>
            ,
            <given-names>J.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tremayne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The quest to automate fact-checking</article-title>
          .
          <source>In: Proceedings of the 2015 Computation+ Journalism Symposium</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Surdeanu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bauer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Finkel</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McClosky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The stanford corenlp natural language processing toolkit</article-title>
          . In:
          <article-title>Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations</article-title>
          . pp.
          <volume>55</volume>
          {
          <issue>60</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Glove:
          <article-title>Global vectors for word representation</article-title>
          .
          <source>In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</source>
          . pp.
          <volume>1532</volume>
          {
          <issue>1543</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Sidorov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Velasquez</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamatatos</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanona-Hernandez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Syntactic n-grams as machine learning features for natural language processing</article-title>
          .
          <source>Expert Systems with Applications</source>
          <volume>41</volume>
          (
          <issue>3</issue>
          ),
          <volume>853</volume>
          {
          <fpage>860</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>