=Paper= {{Paper |id=Vol-1737/T3-1 |storemode=property |title=Overview of the Mixed Script Information Retrieval (MSIR) at FIRE-2016 |pdfUrl=https://ceur-ws.org/Vol-1737/T3-1.pdf |volume=Vol-1737 |authors=Somnath Banerjee,Kunal Chakma,Sudip Kumar Naskar,Amitava Das,Paolo Rosso,Sivaji Bandyopadhyay,Monojit Choudhury |dblpUrl=https://dblp.org/rec/conf/fire/BanerjeeCNDRBC16 }} ==Overview of the Mixed Script Information Retrieval (MSIR) at FIRE-2016== https://ceur-ws.org/Vol-1737/T3-1.pdf
 Overview of the Mixed Script Information Retrieval (MSIR)
                      at FIRE-2016

                                  Somnath Banerjee                     Kunal Chakma
                                Jadavpur University, India            NIT Agartala, India
                                sb.cse.ju@gmail.com kchax4377@gmail.com
           Sudip Kumar Naskar                        Amitava Das                            Paolo Rosso
           Jadavpur University, India                IIIT Sri City, India      Technical University of Valencia, Spain
      sudip.naskar@cse.jdvu.ac.in               amitava.das@iiits.in                   prosso@dsic.upv.es
                               Sivaji Bandyopadhyay                         Monojit Choudhury
                               Jadavpur University, India                   Microsoft Research india
                       sbandyopadhyay@cse.jdvu.ac.in                    monojitc@microsoft.com

ABSTRACT                                                           search that commercial search engines have to tackle. Five
The shared task on Mixed Script Information Retrieval (MSIR)       teams participated in the shared task.
was organized for the fourth year in FIRE-2016. The track             In FIRE-2014, the scope of subtask-1 was extended to
had two subtasks. Subtask-1 was on question classifica-            cover three more South Indian languages - Tamil, Kannada
tion where questions were in code mixed Bengali-English            and Malayalam. In subtask-2, (a) queries in Devanagari
and Bengali was written in transliterated Roman script.            script, and (b) more natural queries with splitting and join-
Subtask-2 was on ad-hoc retrieval of Hindi film song lyrics,       ing of words, were introduced. More than 15 teams partici-
movie reviews and astrology documents, where both the              pated in the 2 subtasks [8].
queries and documents were in Hindi either written in De-             Last year (FIRE-2015), the shared task was renamed from
vanagari script or in Roman transliterated form. A total of        “Transliterated Search” to “Mixed Script Information Re-
33 runs were submitted by 9 participating teams, of which 20       trieval (MSIR)” for aligning it to the framework proposed
runs were for subtask-1 by 7 teams and 13 runs for subtask-2       by [9]. In FIRE-2015, three subtasks were conducted [16].
by 7 teams. The overview presents a comprehensive report           Subtask-1 was extended further by including more Indic lan-
of the subtasks, datasets and performances of the submitted        guages, and transliterated text from all the languages were
runs.                                                              mixed. Subtask-2 was on searching movie dialogues and
                                                                   reviews along with song lyrics. Mixed script question an-
                                                                   swering (MSQA) was introduced as subtask-3. A total of
1.   INTRODUCTION                                                  10 teams made 24 submissions for subtask-1 and subtask-2.
   A large number of languages, including Arabic, Russian,         In spite of a significant number of registrations, no run was
and most of the South and South East Asian languages like          received in subtask-3.
Bengali, Hindi etc., have their own indigenous scripts. How-          This year, we hosted two subtasks in the MSIR shared
ever, the websites and the user generated content (such as         task. Subtask-1 was on classifying code-mixed cross-script
tweets and blogs) in these languages are written using Ro-         question; this task was the continuation of last year’s subtask-
man script due to various socio-cultural and technological         3. Here Bengali words were written in Roman transliter-
reasons[1]. This process of phonetically representing the          ated Bengali. Here Bengali words were written in Roman
words of a language in a non-native script is called translit-     transliterated Bengali. The subtask-2 was on information
eration. English being the most popular language of the            retrieval of Hindi-English code-mixed tweets. The objective
web, transliteration, especially into the Roman script, is         of subtask-2 was to retrieve the top k tweets from a corpus
used abundantly on the Web not only for documents, but             [7] for a given query consisting of Hind-English terms where
also for user queries that intend to search for these docu-        the Hindi terms are written in Roman transliterated form.
ments. This situation, where both documents and queries               This paper provides the overview of the MSIR track in
can be in more than one scripts, and the user expectation          the Eighth Forum for Information Retrieval Conference 2016
could be to retrieve documents across scripts is referred to       (FIRE-2016). The track was coordinated jointly by Mi-
as Mixed Script Information Retrieval.                             crosoft Research India, Jadavpur University, Technical Uni-
   The MSIR shared task was introduced in 2013 as “Translit-       versity of Valencia, IIIT Sriharikota and NIT Agartala. De-
erated Search” at FIRE-2013 [13]. Two pilot subtasks on            tails of these tasks can also be found on the website https:
transliterated search were introduced as a part of the FIRE-       //msir2016.github.io/.
2013 shared task on MSIR. Subtask-1 was on language iden-             The rest of the paper is organized as follows. Section 2
tification of the query words and subsequent back translit-        and 3 describe the datasets, present and analyze the run
eration of the Indian language words. The subtask was              submissions for the Subtask-1 and Subtask-2 respectively.
conducted for three Indian languages - Hindi, Bengali and          We conclude with a summary in Section 4.
Gujarati. Subtask-2 was on ad hoc retrieval of Bollywood
song lyrics - one of the most common forms of transliterated
2.    SUBTASK-1: CODE-MIXED CROSS-
                                                                               Table 2: Subtask-1: Question class statistics
      SCRIPT QUESTION ANSWERING                                                   Class               Training Testing
   Being a classic application of natural language process-                       Person (PER)           55        27
ing, question answering (QA) has practical applications in                        Location (LOC)         26        23
various domains such as education, health care, personal                          Organization (ORG)     67        24
assistance, etc. QA is a retrieval task which is more chal-                       Temporal(TEMP)         61        25
lenging than the task of common search engines because the                        Numerical(NUM)         45        26
purpose of QA is to find accurate and concise answer to                           Distance(DIST)         24        21
a question rather than just retrieving relevant documents                         Money(MNY)             26        16
containing the answer [10]. Recently, the code-mixed cross-                       Object(OBJ)            21        10
script QA research problem was formally introduced in [3].                        Miscellaneous(MISC)     5         8
The first step of understanding a question is to perform ques-
tion analysis. Question classification is an important task in
question analysis which detects the answer type of the ques-
tion. Question classification helps not only filter out a wide
range of candidate answers but also determine answer selec-
tion strategies [10]. Furthermore, it has been observed that
the performance of question classification has significant in-
fluence on the overall performance of a QA system.
   Let, Q = {q1 , q2 , . . . , qn } be a set of factoid questions as-
sociated with domain D. Each question q : hw1 w2 w3 . . . wp i,
is a set of words where p denotes the total number of words
in a question. The words, w1 , w2 , w3 , . . . , wp , could be En-
glish words or transliterated from Bengali in the code mixed
scenario. Let C = {c1 , c2 , . . . , cm } be the set of question
classes. Here n and m refer to the total number of questions                   Figure 1: Classwise distribution of dataset
and question classes respectively.
   The objective of this subtask is to classify each given ques-
tion q i ∈ Q into one of the predefined coarse-grained classes             AMRITA CEN [2] team submitted 2 runs. They used
cj ∈ C. For example, the question “last volvo bus kokhon                bag-of-words (BoW) model for the Run-1. The Run-2 was
chare?” (English gloss: “When does the last volvo bus de-               based on Recurrent Neural Network (RNN). The initial em-
part?”) should be classified to the class ‘TEMPORAL’.                   bedding vector was given to RNN and the output of RNN
                                                                        was fed to logistic regression for training. Overall, the BoW
2.1    Datasets                                                         model outperformed the RNN model by almost 7% ons F1-
   We prepared the datasets for subtask-1 from the dataset              measure.
described in [3] which is the only dataset available for code-             AMRITA-CEN-NLP [4] team submitted 3 runs. They
mixed cross-script question answering research. The dataset             approached the problem using Vector Space Model (VSM).
described in [3] contains questions, messages and answers               Weighted term based on the context was applied to overcome
from the sports and tourism domains in code-mixed cross-                the shortcomings of VSM. The proposed approach achieved
script English–Bengali. The dataset contains a total of 20              upto 80% accuracy in terms of F1-measure.
documents from two domains, namely sports and tourism.                     ANUJ [15] also submitted 3 runs. The author used term
There are 10 documents in the sports domain which consist               frequency âĂŞ inverse document frequency (TF-IDF) vec-
of 116 informal posts and 192 questions, while the 10 doc-              tor as feature. A number of machine learning algorithms,
uments in the tourism domain consist of 183 informal posts              namely Support Vector Machines (SVM), Logistic Regres-
and 314 questions. We initially provided 330 labeled factoid            sion (LR), Random Forest (RF) and Gradient Boosting were
questions as the development set to the participants after ac-          applied using Grid Search to come up with the best param-
cepting the data usage agreement. The testset contains 180              eters and model. The RF model performed the best among
unlabeled factoid questions. Table 1 and Table 2 provide                the 3 runs.
statistics of the dataset. Question class specific distribution            BITS PILANI [5] submitted 3 runs. Instead of apply-
of the datasets is given in Figure 1.                                   ing the classifiers on the code-mixed cross-script data, they
                                                                        convert the data into English. The translation was per-
                                                                        formed using Google translation API 1 . Then they applied
        Table 1: MSIR16 Subtask-1 Datasets
 Dataset Questions(Q) Total Words Avg. Words/Q                          three machine learning classifiers for each run, namely Gaus-
 Trainset       330       1776         5.321                            sian Naı̈ve Bayes, LR and RF Classifier. However, Gaussian
 Testset        180       1138         6.322                            Naive Bayes classifier outperformed the other two classifiers.
                                                                           IINTU [6] was the best performing team. The team sub-
                                                                        mitted 3 runs which were based on machine learning ap-
2.2    Submissions                                                      proaches. They trained three separate classifiers namely RF,
                                                                        One-vs-Rest and k-NN, followed by building an ensemble
  A total of 15 research teams registered for subtask-1. How-           classifier using these 3 classifiers for the classification task.
ever, only 7 teams submitted runs and a total of 20 runs                The ensemble classifier took the output label by each of the
were received. All the teams submitted 3 runs except AM-
                                                                        1
RITA CEN who submitted 2 runs.                                              https://translate.google.com/
individual classifiers and selected the majority label as out-    accuracy. It can be observed from Table 3 that the highest
put. In case of tie any one label was chosen at random as         accuracy (83.333%) was achieved by the IINTU team. The
output.                                                           classification performance on the temporal (TEMP) class
   NLP-NITMZ [12] submitted 3 runs of which 2 runs were           was very high for almost all the teams. However, Table 4 and
rule based - a first set of direct rules were applied for the     Figure 2 suggest that the miscellaneous (MISC) question
Run-1 while a second set of dependent rules were used for the     class was very difficult to identify. Most of the teams could
Run-3. A total of 39 rules were identified for the rule based     not identify the MISC class. The reason could be very low
runs. Naı̈ve Bayes classifier was used in Run-2 whereas           presence(2%) of MISC class in the training dataset.
Naı̈ve Bayes updateable classifier was used in Run-3.
   IIT(ISM)D used three different machine learning based          3.    SUBTASK-2:INFORMATION RETRIEV-
classification models - Sequential Minimal Optimization, Naı̈ve
Bayes Multimodel and Decision Tree FT to annotate the                   AL ON CODE-MIXED HINDI-ENGLISH
question text. This team submitted the runs after the dead-             TWEETS
line.                                                                This subtask is based on the concepts discussed in [9].
2.3    Results                                                    In this subtask, the objective was to retrieve Code-Mixed
                                                                  Hindi-English tweets from a corpus for code-mixed queries.
  In this section, we define the evaluation metrics used to       The Hindi components in both the tweets and the queries
evaluate the runs submitted to the subtask-1. Typically, the      are written in Roman transliterated form. This subtask
performance of a question classifier is measured by calculat-     did not consider cases where both Roman and Devanagari
ing the accuracy of that classifier on a particular test set      scripts are present. Therefore, the documents in this case are
[10]. We also used this metric for evaluating the code-mixed      tweets consisting of code-mixed Hindi-English texts where
cross-script question classification performance.                 the Hindi terms are in Roman transliterated form. Given a
                   number of correctly classified samples         query consisting of Hindi and English terms written in Ro-
      accuracy =                                                  man script, the system has to retrieve the top-k documents
                     total number of testset samples
                                                                  (i.e., tweets) from a corpus that contains Code-Mixed Hindi-
  In addition, we also computed the standard precision, re-       English tweets. The expected output is a ranked list of the
call and F1-measure to evaluate the class specific perfor-        top twenty (k=20 here) tweets retrieved from the given cor-
mances of the participating systems. The precision, recall        pus.
and F1-measure of a classifier on a particular class c are
defined as follows:                                               3.1   Datasets
                 number of samples correctly classified as c        Initially we released 6,133 code-mixed Hindi-English tweets
precision(P ) =
                     number of samples classified as c            with 23 queries as the training dataset. Later we released
                                                                  a document collection containing 2,796 code-mixed tweets
                number of samples correctly classified as c       along with with 12 code-mixed queries as the testset. Query
  recall(R) =                                                     terms are mostly named entities with Roman transliterated
                   total number of samples in class c
                                                                  Hindi words. The average length of the queries in the train-
                                                                  ing set is 3.43 words and in the testset it is 3.25 words. The
                                     2.P.R                        tweets in the training set cover 10 topics whereas the testset
                   F 1 − measure =
                                     P +R                         cover 3 topics.
  Table 3 presents the performance of the submitted runs in
terms of accuracy. Class specific performances are reported       3.2   Submissions
in Table 4. A baseline system was also developed for the             This year total 7 teams have submitted 13 runs. The
sake of comparison using the BoW which obtained 79.444%           submitted runs for the retrieval task of Code-Mixed tweets
                                                                  mostly adopted preprocessing of the data and then applying
                                                                  different techniques for retrieving the desired tweets. Team
       Table 3: Subtask-1: Teams Performance                      Amrita CEN [4] removed some Hindi/English stop words to
  Team                Run ID   Correct   Incorrect   Accuracy     declutter useless words. After that, they have tokenized all
  Baseline              -        143        37        79.440      the tweets. The cosine distance was used to score the rele-
  AmritaCEN             1        145        35        80.556
  AmritaCEN             2        133        47        73.889      vance of tweets to the query. After that, the top 20 tweets
  AMRITA-CEN-NLP        1        143        37        79.444      based on the scores were retrieved. Team CEN@Amrita[14]
  AMRITA-CEN-NLP        2        132        48        73.333
  AMRITA-CEN-NLP        3        132        48        73.333
                                                                  used a Vector Space Model based approach. Team UB [11]
  Anuj                  1        139        41        77.222      has adopted three different techniques for the retrieval task.
  Anuj                  2        146        34        81.111      First, they have used Named Entity boosts where the pur-
  Anuj                  3        141        39        78.333
  BITS PILANI           1        146        34        81.111      pose was to boost the documents based on their NE matches
  BITS PILANI           2        144        36        80.000      from the query, i.e., the query was parsed to extract NEs and
  BITS PILANI           3        131        49        72.778
  IINTU                 1        147        33        81.667
                                                                  each document (tweet) that matched the given NE was pro-
  IINTU                 2       150         30        83.333      vided a small numeric boost. At the second level of boost-
  IINTU                 3        146        34        81.111      ing, phrase matching was done , i.e., documents that more
  NLP-NITMZ             1        134        46        74.444
  NLP-NITMZ             2        134        46        74.444      closely matched the input query phrase were ranked higher
  NLP-NITMZ             3        142        38        78.889      than those that did not. The UB team used Synonym Expan-
  *IIT(ISM)D            1        144        36        80.000
  *IIT(ISM)D            2        142        38        78.889      sion and Narrative based weighting as the second and third
  *IIT(ISM)D            3        144        36        80.000      techniques. Team NITA NITMZ performed stop words re-
                   * denotes late submission                      moval followed by query segmentation and finally merging.
                    Figure 2: Subtask-1: F-Measure of different teams for classes (Best run)


          Table 4: Subtask-1: Class specific performances (NA denotes no identification of a class)
        Team                 Run ID    PER      LOC      ORG      NUM     TEMP      MONEY      DIST     OBJ      MISC
        AmritaCEN              1      0.8214   0.8182   0.5667   0.9286   1.0000     0.7742   0.9756   0.5714     NA
        AmritaCEN              2      0.7541   0.8095   0.6667   0.8125   1.0000     0.4615   0.8649     NA       NA
        AMRITA-CEN-NLP         1      0.8000   0.8936   0.6032   0.8525   0.9796     0.7200   0.9500   0.5882     NA
        AMRITA-CEN-NLP         2      0.7500   0.7273   0.5507   0.8387   0.9434     0.5833   0.9756   0.1818     NA
        AMRITA-CEN-NLP         3      0.6939   0.8936   0.5455   0.8125   0.9804     0.6154   0.8333   0.3077     NA
        IINTU                  1      0.7843   0.8571   0.6333   0.9286   1.0000     0.8125   0.9756   0.4615     NA
        IINTU                  2      0.8077   0.8980   0.6552   0.9455   1.0000     0.8125   0.9756   0.5333     NA
        IINTU                  3      0.7600   0.8571   0.5938   0.9455   1.0000     0.8571   0.9767   0.4615     NA
        NLP-NITMZ              1      0.7347   0.8444   0.5667   0.8387   0.9796     0.6154   0.9268   0.2857   0.1429
        NLP-NITMZ              2      0.6190   0.8444   0.5667   0.9630   0.8000     0.7333   0.9756   0.4286   0.1429
        NLP-NITMZ              3      0.8571   0.8163   0.7000   0.8966   0.9583     0.7407   0.9268   0.3333   0.2000
        Anuj                   1      0.7600   0.8936   0.6032   0.8125   0.9804     0.7200   0.8649   0.5333     NA
        Anuj                   2      0.8163   0.8163   0.5538   0.9811   0.9796     0.9677   0.9500   0.5000     NA
        Anuj                   3      0.8163   0.8936   0.5846   0.8254   1.0000     0.7200   0.8947   0.5333     NA
        BITS PILANI            1      0.7297   0.7442   0.7442   0.9600   0.9200     0.9412   0.9500   0.5000   0.2000
        BITS PILANI            2      0.6753   0.7805   0.7273   0.9455   0.9600    1.0000    0.8947   0.4286     NA
        BITS PILANI            3      0.6190   0.7805   0.7179   0.8125   0.8936     0.9333   0.6452   0.5333     NA
        *IIT(ISM)D             1      0.7755   0.8936   0.6129   0.8966   0.9412     0.7692   0.9524   0.5882     NA
        *IIT(ISM)D             2      0.8400   0.8750   0.6780   0.8525   0.9091     0.6667   0.9500   0.1667     NA
        *IIT(ISM)D             3      0.8000   0.8936   0.6207   0.8667   1.0000     0.6923   0.9500   0.5333     NA
        Avg                           0.7607   0.8415   0.6245   0.8858   0.9613     0.7568   0.9204   0.4458     NA



Team IIT(ISM) D considered every tweet as a document              subtask-2. MAP is also referred to as “average precision at
and indexed using uniword indexing on Terrier implemen-           seen relevant documents”. The idea is that, first, average
tation. Query terms were expanded using soundex coding            precision is computed for each query and subsequently the
scheme. Terms with identical soundex code were selected           average precisions are averaged over the queries. MAP is
as candidate query and included in final queries to retrieve      represented as
the relevent tweets (documents). Further, they have used                                             Qj
                                                                                               N
three different retrieval models BM25, DFR and TF-IDF to                                    1 X 1 X
measure the similarity. However, this team submitted the                           M AP =                P (doci )
                                                                                            N j=1 Qj i=1
runs after the deadline.
                                                                  where Qj refers to the number of relevant documents for
3.3   Results                                                     query j, N indicates the number of queries and P (doci ) rep-
  The retrieval task requires that the retrieved documents        resents precision at the ith relevant document.
at higher ranks be more important than the retrieved doc-            The evaluation results of the submitted runs are reported
uments at lower ranks for a given query and we want our           in Table 5. The highest MAP (0.0377) was achieved by team
measures to account for that. Therefore, set based evalu-         Amrita@CEN which is still very low. The significantly low
ation metrics such as Precision, Recall and F-measure are         MAP values in Table 5 suggest that the task of retrieving
not suitable for this task. Therefore, we used Mean Average       Code-Mixed tweets against query terms comprising code-
Precision (MAP) as the performance evaluation metric for          mixed Hindi and English words is a difficult task and the
techniques proposed by the teams do not produce satisfac-       6.   REFERENCES
tory results. Therefore, the problem of retrieving relevant      [1] U. Z. Ahmed, K. Bali, M. Choudhury, and S. VB.
code-mixed tweets requires better techniques and method-
                                                                     Challenges in designing input method editors for
ologies to be developed for improving system performance.
                                                                     indian languages: The role of word-origin and context.
                                                                     Advances in Text Input Methods (WTIM 2011), pages
Table 5: Results for Subtask-2 showing Mean Aver-                    1–9, 2011.
age Precision                                                    [2] M. Anand Kumar and K. P. Soman.
           Team           Run ID MAP                                 Amrita-CEN@MSIR-FIRE2016: Code-Mixed
           UB             1       0.0217                             Question Classification using BoWs and RNN
           UB             2       0.016                              Embeddings. In Working notes of FIRE 2016 - Forum
           UB             3       0.0152                             for Information Retrieval Evaluation, Kolkata, India,
           Anuj           1       0.0209                             December 7-10, 2016, CEUR Workshop Proceedings.
           Anuj           2       0.0199                             CEUR-WS.org, 2016.
           Amrita CEN 1           0.0377                         [3] S. Banerjee, S. K. Naskar, P. Rosso, and
           NLP-NITMZ      1       0.0203                             S. Bandyopadhyay. The First Cross-Script
           NITA NITMZ 1           0.0047                             Code-Mixed Question Answering Corpus. Proceedings
           CEN@Amrita     1       0.0315                             of the workshop on Modeling, Learning and Mining for
           CEN@Amrita     2       0.0168                             Cross/Multilinguality (MultiLingMine 2016),
           IIT(ISM)D      1       0.0021                             co-located with The 38th European Conference on
           IIT(ISM)D      2       0.0083                             Information Retrieval (ECIR), 2016.
           IIT(ISM)D      3       0.0021                         [4] H. B. Barathi Ganesh, M. Anand Kumar, and K. P.
                                                                     Soman. Distributional Semantic Representation for
                                                                     Text Classification and Information Retrieval. In
                                                                     Working notes of FIRE 2016 - Forum for Information
4.   SUMMARY                                                         Retrieval Evaluation, Kolkata, India, December 7-10,
   In this overview, we elaborated on the two subtasks of            2016, CEUR Workshop Proceedings. CEUR-WS.org,
the MSIR-2016 at FIRE-2016. The overview is divided into             2016.
two major parts one for each subtask, where the dataset,         [5] R. Bhargava, S. Khandelwal, A. Bhatia, and
evaluations metric and results are discussed in detail. A            Y. Sharmai. Modeling Classifier for Code Mixed Cross
total of 33 runs were submitted from 9 unique teams.                 Script Questions. In Working notes of FIRE 2016 -
   In subtask-1, 20 runs were received from 7 teams. The             Forum for Information Retrieval Evaluation, Kolkata,
best performing team achieved 83.333% accuracy. The av-              India, December 7-10, 2016, CEUR Workshop
erage question classification performance obtained in terms          Proceedings. CEUR-WS.org, 2016.
of accuracy was 78.19% which was quite satisfactory con-         [6] D. Bhattacharjee and P. Bhattacharya. Ensemble
sidering this new research problem. The subtask-1 deals              Classifier based approach for Code-Mixed Cross-Script
with code-mixed Bengali-English language. In the coming              Question Classification. In Working notes of FIRE
years, we would like to include more Indian languages. The           2016 - Forum for Information Retrieval Evaluation,
participation was encouraging and we plan to continue the            Kolkata, India, December 7-10, 2016, CEUR
subtask-1 in subsequent FIRE conferences.                            Workshop Proceedings. CEUR-WS.org, 2016.
   Subtask-2 received a total of 13 run submissions from 7       [7] K. Chakma and A. Das. CMIR: A Corpus for
teams out of which one team submitted after the deadline.            Evaluation of Code Mixed Information Retrieval of
The best MAP value achieved was 0.0377 which is consid-              Hindi-English Tweets. In In the 17th International
erably low. From the results of the run submissions it can           Conference on Intelligent Text Processing and
be inferred that information retrieval of code-mixed infor-          Computational Linguistics (CICLING), April 2016.
mal micro blog texts such as tweets is a very challenging        [8] M. Choudhury, G. C. Gupta, P. Gupta, and A. Das.
task. Therefore, the stated problem opens and calls for              Overview of FIRE 2014 Track on Transliterated
new avenues of research for developing better techniques and         Search, 2014.
methodologies.                                                   [9] P. Gupta, K. Bali, R. E. Banchs, M. Choudhury, and
                                                                     P. Rosso. Query expansion for mixed-script
5.   ACKNOWLEDGMENTS                                                 information retrieval. In Proceedings of the 37th
   Somnath Banerjee, Sudip Kumar Naskar and Sivaji Bandy-            international ACM SIGIR conference on Research &
opadhyay acknowledge the support of the Ministry of Elec-            development in information retrieval, pages 677–686.
tronics and Information Technology (MeitY), Government               ACM, 2014.
of India, through the project “CLIA System Phase II”.           [10] X. Li and D. Roth. Learning question classifiers. In
   The work of Paolo Rosso has been partially funded by              Proceedings of the 19th international conference on
SomEMBED MINECO TIN2015-71147-C2-1-P research project                Computational linguistics-Volume 1, pages 1–7.
and by the Generalitat Valenciana under the grant ALMA-              Association for Computational Linguistics, 2002.
MATER (PrometeoII/2014/030).                                    [11] N. Londhe and R. K. Srihari. Exploiting Named Entity
   We would also like to thank everybody who helped spread           Mentions Towards Code Mixed IR : Working Notes for
awareness about this track in their respective capacities and        the UB system submission for MSIR@FIRE’16.
the entire FIRE team for giving us the opportunity and plat-    [12] G. Majumder and P. Pakray. NLP-NITMZ @ MSIR
form for conducting this new track smoothly.                         2016 System for Code-Mixed Cross-Script Question
     Classification. In Working notes of FIRE 2016 -
     Forum for Information Retrieval Evaluation, Kolkata,
     India, December 7-10, 2016, CEUR Workshop
     Proceedings. CEUR-WS.org, 2016.
[13] R. S. Roy, M. Choudhury, P. Majumder, and
     K. Agarwal. Overview and datasets of FIRE 2013
     track on transliterated search. In Fifth Forum for
     Information Retrieval Evaluation, 2013.
[14] S. Singh and Anand Kumar, M and Soman, KP.
     CEN@Amrita: Information Retrieval on CodeMixed
     Hindi-English Tweets Using Vector Space Models. In
     Working notes of FIRE 2016 - Forum for Information
     Retrieval Evaluation, Kolkata, India, December 7-10,
     2016, CEUR Workshop Proceedings, December 2016.
[15] A. Saini. Code Mixed Cross Script Question
     Classification. In Working notes of FIRE 2016 -
     Forum for Information Retrieval Evaluation, Kolkata,
     India, December 7-10, 2016, CEUR Workshop
     Proceedings. CEUR-WS.org, 2016.
[16] R. Sequiera, M. Choudhury, P. Gupta, P. Rosso,
     S. Kumar, S. Banerjee, S. K. Naskar,
     S. Bandyopadhyay, G. Chittaranjan, A. Das, and
     K. Chakma. Overview of FIRE-2015 Shared Task on
     Mixed Script Information Retrieval.