=Paper= {{Paper |id=Vol-2266/T1-4 |storemode=property |title=Fully Automatic Approach to Identify Factual or Fact-checkable Tweets |pdfUrl=https://ceur-ws.org/Vol-2266/T1-4.pdf |volume=Vol-2266 |authors=Sarthak Anand,Rajat Gupta,Rajiv Ratn Shah,Ponnurangam Kumaraguru |dblpUrl=https://dblp.org/rec/conf/fire/AnandGSK18 }} ==Fully Automatic Approach to Identify Factual or Fact-checkable Tweets== https://ceur-ws.org/Vol-2266/T1-4.pdf
Fully Automatic Approach to Identify Factual or
                         Fact-checkable Tweets


      Sarthak Anand1 , Rajat Gupta2 , Rajiv Ratn Shah3 , and Ponnurangam
                                Kumaraguru3?
          1
            Netaji Subhas Institute of Technology, New Delhi 110078, INDIA
          2
            Maharaja Agrasen Institute of Technology, Delhi 110086, INDIA
       3
         Indraprastha Institute of Information Technology, Delhi 110020, INDIA
                               sarthaka.ic@nsit.net.in



        Abstract. This paper presents the solution of the team MIDAS of IIIT
        Delhi for the IRMiDis track in FIRE 2018. We present our solution for
        the identication of factual or fact-checkable tweets from a dataset that
        consists of about 50,000 tweets posted during the 2015 Nepal earthquake.
        We provide a rule based approach for this task and compare it with a
        semi-supervised approach. After preprocessing steps including tokeniza-
        tion and cleaning, we calculate a factuality score on the basis of number
        of proper-nouns and quantitative values within a tweet and nally rank
        them according to the score. Experimental results show that this simple
        rule based approach provides comparable results in comparison to that
        of semi-supervised approach.

        Keywords: Social media analysis, Unsupervised learning, Information
        retrieval, Microblogs, Disaster


1     Introduction
Social media usage has considerably increased over the last decade. People often
use the social media for various purposes and create a huge amount of user-
generated content. In addition to the reporting of news or events social media
platforms are increasingly being used for aiding relief operations during various
mass emergencies, e.g., during Kerala oods 2018.
    However, messages posted on these sites often contain rumors and false in-
formation. In such situations, identication of factual or fact-checkable tweets,
i.e., tweets that report some relevant and veriable fact is extremely important
for eective coordination of post-disaster relief operations. Additionally, cross
verication of such critical information is a practical necessity to ensure the
trustworthiness. Considering the scale of these platforms it is not feasible to
manually check and verify dierent user-generated content on time. Since it is
very important to reach to a person who is stuck in such emergencies on time,
?
    This work was done when Ponnurangam Kumaraguru was on sabbatical at Interna-
    tional Institute of Information Technology, Hyderabad.
2      Sarthak et al.

automated IR techniques are needed to identify, process and verify the credibility
of information from multiple sources.
    With this paper we provide one such approach which has shown the best
performance in the FIRE challenge 2018 [1] on identifying factual tweets.

2   Related Work
Identifying factual and non factual tweets can be treated as a supervised clas-
sication problem. A lot of work have already been done related to supervised
based classication [7] [8] [4]. All these works require large amounts of manually
labeled dataset.
    Despite most works focus on supervised techniques, some works also em-
ployed unsupervised techniques as well. For instace, Bjorn Schuller et al. [6]
worked on knowledge based approach which does not demand labeled training
data. Moreover, Shailesh S. Deshpand et al. [2] proposed a rule based approach
for the classication of sentences. They tested it for identifying specic and non
specic sentences. They computed several features for each sentence for com-
puting a specicity score for each sentence. Similar to their approach we extract
features from sentences such as the number of proper nouns(PROPN) and the
number of quantitative values(NUM) and compute a factuality score(higher score
indicates more factual information ). In our approach, we use the factual score
for ranking the tweets in order of factual information and use the top k sentences
as fact-checkable tweets.

3   Problem and Data Description
Information retrieval from micro-blogs during disasters challenge had 2 sub-
tasks. Sub-task 1 was about, identifying factual or fact-checkable tweets related
to Nepal disaster and ranking them on the basis of their factuality scores. Sub-
task 2 was about, mapping the fact-checkable tweets with appropriate news
articles. The submission was categorized into 3 types based on the amount of
manual intervention i.e. Fully automatic, Semi automatic, and Manual.
    Data Description Dataset for sub-task 1 consists of about 50,000 tweets
posted during the 2015 Nepal earthquake. Dataset for sub-task 2 included around
6,000 news articles related to the 2015 Nepal earthquake. Refer [1] for more
details.

4   Automatic Methodology
The problem at hand is to use tweets and rank them based on the information
they contain. The following sections describe in detail the various steps that have
been performed to achieve the results and intuition behind our approach.
 1. Pre-processing of tweets, POS tagging and nding proper-nouns and quan-
    titative values, are described in Section 4.2.
      Fully Automatic Approach to Identify Factual or Fact-checkable Tweets       3

 2. Finally computing a factuality score based on proper-nouns and quantitative
    values, is described in Section 4.3.

4.1    Intuition
Similar to the ndings of Shailesh S. Deshpande et al. [2], in our study we nd
that tweets that contain some factual information consists of some name entities
like an organization like UN or NDRF, or proper noun such as PM Modi and
quantitative information such as date, time or numbers(e.g., 5 dead or 5 tonnes ).
Based on this study we try to score a tweet on the basis of number of proper
nouns and quantitative values which we call as factuality score.

4.2    Data Preprocessing and POS tagging
Since the data given to us is raw, noisy and also prone to more errors, it cannot be
directly used for analysis. It is necessary to perform some preprocessing to make
the data more suitable so that we can perform POS tagging on the sentences.
The following preprocessing steps were performed:
 1. Tokenization: Tokenization refers to the breaking down of the given text
    into individual words. We use the Spacy's word tokenizer to perform tok-
    enization of the tweets.
 2. Normalization: We perform the following steps, very specic to tweets to
    normalize our corpus:
       Stop-words and punctuation removal: Usually tweets consists of
        mentions, hash-tags, URLs, punctuation marks and emoji's. They are
        not useful in determining the amount of information within a tweet and
        hence are removed from our corpus.
POS tagging In our approach, we have used two major features for computing
factuality score, i.e., the number of proper nouns and quantitative values within
a tweet. We use spacy's POS tagger for this purpose.

4.3    Computing Factuality Score
Submitted Approach 1 In this approach we compute the number of proper
nouns and number of quantitative values within the tweet. For mapping the
score to 0 and 1 we divide the number of PROPN and NUM by maximum
values achieved in their respective eld. Finally, we take average of both these
values. The Table 1 shows examples for calculating the factuality score. The
underlined words refer to proper-nouns and italicized words refer to numbers.
For these examples, note that the maximum values of PROPN and NUM were
17 and 13, respectively. (Shortcomings and suggestions for this approach are
described in Section 7)
1
    Github code available at: https://github.com/isarth/Fire_task_1
4       Sarthak et al.

                         Table 1. Computing factuality score
                  Tweet                   PROPN score NUM score Factuality score
   Currently working rescue Army
 CHINA, INDIA, FRANCE, ISREAL,
TURKEY, GERMANY, USA, UK, UAE, 13/17= 0.7647                   0.0          0.3823


    Missing tourists in earthquake
  include 15 French, 12 Russians,10   5/17 = 0.2941 5/13=0.3846             0.3393
Canadians, 9 Americans and 8 Spanish.


       Quake wake-up call for govt,
      need better building tech: Eert            0                0           0



5     Semi Automatic Methodology
For comparing our automatic approach with supervised approach. We manually
labeled around 1,500 tweets as factual and non-factual and treat the sub-task
1 (refer Section 3) as binary classication problem. The condence score of the
classier is treated as the factuality score, which is nally used for ranking the
tweets. The following section describes in detail various steps that have been
performed for the semi-automatic approach.
1. Manually labeling a small set of tweets from the dataset.
2. Pre-processing steps, already described in Section 4.2
3. Training a binary classier and nally ranking tweets according the con-
   dence score (see Section 5.1 for details).

5.1   Binary Classier
For classifying tweets as factual and Non-factual, we train both Fasttext [3] cbow
and bi-gram models. We split our labeled dataset into two parts training and
validation. Table 2 shows the performance of both the classiers. Finally for
ranking tweets in order of factuality, we treat the condence score of bi-gram
classier as our factuality score.

                    Table 2. Performance of FastText classiers
                   Fast Text classier Validation Accuracy
                          CBOW                       0.756
                          Bi-gram                    0.796
    Fully Automatic Approach to Identify Factual or Fact-checkable Tweets    5

6   Result and Analysis
Finally Table 3 compares the results of automatic and semi-automatic approach
in the FIRE'18 challenge. Table 4 summarizes the nal results of other teams
that participated in the FIRE'18 task for automatic submission. We were ranked
rst in the competition with an NDCG score of 0.6835. The lowest NDCG score
achieved in the competition was 0.1271. Table 5 summarizes the nal results of
other teams that participated in the FIRE'18 task for semi-automatic submis-
sion. We were ranked second in that task. For detailed results refer [1].

Table 3. Results show that automatic approach is comparable with semi automatic
approach.

                Method       P@100 R@100 MAP NDCG@100 NDCG
             Semi Automatic 0.960    0.1148 0.1345     0.6007   0.6899
               Automatic    0.880    0.1292 0.1329     0.5649   0.6835




        Table 4. Top ve automatic submission (our submission: MIDAS)
                Teams        P@100 R@100 MAP NDCG@100 NDCG
            MIDAS et al. 0.8800 0.1292 0.1329          0.5649   0.6835
            FASTNU et al. 0.7000     0.0885   0.0801   0.5723   0.6676
              UEM et al. 0.6800      0.1427   0.1178   0.5332   0.6396
              UEM et al. 0.6400      0.1069   0.0767   0.5237   0.5276
            IIT BHU et al. 0.9300    0.1938   0.1568   0.8645   0.4532




Table 5. Final ranking of all teams for semi-automatic (our submission: MIDAS)
submission
                Teams        P@100 R@100 MAP NDCG@100 NDCG
             DAIICT et al.   0.400   0.2002 0.1471     0.4021   0.7492
            MIDAS et al. 0.9600 0.1148 0.1345          0.6007   0.6899
            IIT BHU et al. 0.390     0.0447 0.0401     0.3272    0.620




7   Conclusion and Future Work
We have presented our automatic approach for calculating the factuality score
on the basis of number of proper-nouns and quantitative values within a tweet
which provided comparable results with semi automatic approach in FIRE'18
6       Sarthak et al.

Information Retrieval from Micro-blogs during Disasters (IRMiDis) task. The
best automatic submission achieved the NDCG score of 0.6835, that made our
team stand at rst position globally in terms of NDCG score.
    On further exploring we nd two minor issues in the automatic approach
described in Section 4.3 are:
 1. Because we are dividing by the maximum value in each eld to obtain
    PROPN and NUM score. The individual scores will not contribute equally
    for the factuality score.
 2. Also, tweets with large number of just quantitative values or proper-nouns
    will achieve high factuality score.
To overcome the above mentioned issues, we suggest having an upper-bound to
the PROPN and NUM values as λ. Hence for computing the individual score
we take min(propn/num, λ) and nally to map score between 0 and 1 we divide
by λ and take the average of both the scores. Futher exploration can be done of
nding value of λ. These shortcomings remain, as to be solved as future work.
    We also aim to extend the model by making it more ecient by using dierent
techniques we did not explore such as using other features like TFIDF [5] score
of words, combined with the ones we already tried. Further knowledge based
classication [6] can also be explored .

References
1. Basu, M., Ghosh, S., Ghosh, K.: Overview of the FIRE 2018 track: Information
   Retrieval from Microblogs during Disasters (IRMiDis). In: Proceedings of FIRE
   2018 - Forum for Information Retrieval Evaluation (December 2018)
2. Deshpande, S.S., Palshikar, G.K., Athiappan, G.: Unsupervised approach to sen-
   tence classication (2010)
3. Joulin, A., Grave, E., Bojanowski, P., Mikolov, T.: Bag of tricks for ecient text
   classication. CoRR abs/1607.01759 (2016)
4. Lei Shen, J.Z.: Empirical evaluation of rnn architectures on sentence classication
   task
5. Ramos, J.: Using tf-idf to determine word relevance in document queries
6. Schuller, B., Knaup, T.: Learning and knowledge-based sentiment analysis in movie
   review key excerpts
7. Wang, S., Manning, C.D.: Baselines and bigrams: Simple, good sentiment and topic
   classication (2012)
8. Zhang, X., Zhao, J.J., LeCun, Y.: Character-level convolutional networks for text
   classication. CoRR abs/1509.01626 (2015)