=Paper= {{Paper |id=Vol-1180/CLEF2014wn-QA-LiuEt2014 |storemode=property |title=The Fudan-UIUC Participation in the BioASQ Challenge Task 2a: The Antinomyra system |pdfUrl=https://ceur-ws.org/Vol-1180/CLEF2014wn-QA-LiuEt2014.pdf |volume=Vol-1180 |dblpUrl=https://dblp.org/rec/conf/clef/LiuWPZZ14 }} ==The Fudan-UIUC Participation in the BioASQ Challenge Task 2a: The Antinomyra system== https://ceur-ws.org/Vol-1180/CLEF2014wn-QA-LiuEt2014.pdf
    The Fudan-UIUC participation in the BioASQ
                     Challenge
          Task 2a: The Antinomyra system

 Ke Liu1,2 , Junqiu Wu3 , Shengwen Peng1,2 , Chengxiang Zhai4 , and Shanfeng
                                  Zhu1,2 ?
    1
    School of Computer Science, Fudan University, Shanghai 200433, P. R. China,
        2
     Shanghai Key Lab of Intelligent Information Processing, Fudan University,
                          Shanghai 200433, P. R. China
                        {antinomyra,pswgoo}@gmail.com
                               zhusf@fudan.edu.cn
3
  School of Information Science & Engineering, Central South University, 410083 P.
                                    R. China
                                oxalca@gmail.com
4
  Department of Computer Science, University of Illinois at Urbana-Champaign, IL
                                  611801, USA
                               czhai@illinois.edu



            Abstract. This paper describes the Antinomyra System that partici-
            pated in the BioASQ Task 2a Challenge for the large-scale biomedical
            semantic indexing. The system can automatically annotate MeSH terms
            for MEDLINE citations using only title and abstract information. With
            respect to the official test set (batch 3, week 5), based on 1867 annotat-
            ed citations out of all 4533 citations (June 6, 2014), our best submission
            achieved 0.6199 in flat Micro F-measure. This is 9.8% higher than the
            performance of official NLM solution Medical Text Indexer (MTI), which
            achieved 0.5647 in flat F-measure.

            Keywords: MeSH Indexing; Logistic Regression; Learning to Rank;
            Multi-Label Classification.


1       Introduction
1.1         The Problem
Medical Subject Headings (MeSH) are used by National Library of Medicine
(NLM) to index articles in MEDLINE [1]. MeSH is organized in hierarchical
structure, and slightly updated every year. In 2014, there are altogether 27149
MeSH headings5 . Usually each article is annotated by 5 to 20 MeSH headings.
Many studies have been carried out to utilize MeSH for efficiently retrieving
and mining biomedical documents for knowledge discovery [2, 3, 4, 5, 6, 7]. The
?
    Corresponding author
5
    http://www.nlm.nih.gov/mesh/introduction.html




                                            1311
accurate prediction of MeSH headings for each citation will greatly reduce the fi-
nancial and time cost of annotating biomedical documents. The BioASQ Task2a
is a large scale biomedical semantic indexing competition for automatic MeSH
annotation. Each week, thousands of new MEDLINE citations are provided to
the competition participants, who are required to submit predicted MeSH head-
ings of each citation in 21 hours. Since each MeSH heading can be deemed as a
class label, the MeSH annotation problem is a multi-label classification problem.
For each citation, our Antinomyra system tries to assign it a certain number of
MeSH headings out of all 27149 MeSH headings.

1.2   Challenges
Simple mapping does not work well Many MeSH headings do not appear
directly in the title or abstract of a target citation, which means that a simple
mapping does not work very well. This is why we resort to advanced machine
learning methods to predict MeSH headings.

The amount of information is insufficient During the competition, we
have very limited information of target citations, more specifically, only titles
and abstracts are available in this task. By contrast, MeSH indexers have the full
text for annotating MeSH headings. Since the main text contains some important
clues, predicting MeSH headings with very limited information is a big challenge
for the competition participants. Indeed, it will be very interesting if we can know
the performance of professional MeSH annotators with the same information as
the BioASQ Task2a participants. This may bring more insights on how to make
use of the information in the title and abstract.



Table 1. The frequencies of some typical MeSH headings in our Local MEDLINE
database of 12,504,999 citations.

        ID     MeSH                                 Count Frequency Rank
        6801 Humans                             8152852                  1
        8297 Male                               4777692                  2
        18570 Risk Assessment                    129816                100
        2540 Cerebral Cortex                      74513                200
        12987 Soil                                23178               1000
        12045 Regulatory Sequences, Nucleic Acid 12503                2000
        23001 Transplantation Tolerance            1532              10000
        6991 Hypnosis, Anesthetic                   199              20000



There are large variations in the occurrence of different MeSH head-
ings Some MeSH headings, such as check tags Humans, Male and Female, ap-
pear very frequently, while some others are very rare. We have counted the oc-
currence of each MeSH heading in the whole MEDLINE. Only about 150 MeSH




                                     1312
headings have occurred in more than 1% of the whole MEDLINE. That is to say,
a vast majority of MeSH headings do not appear very often. It is not surprising
that, for many MeSH headings, we lack of enough positive examples (citations)
to train accurate models. As illustrated in Table 1, MeSH heading“Hypnosis,
Anestheti” only occurs in 199 out of more than 12 million citations.



2     Related Work

In the past few years, especially for the last BioASQ challenge (task 1a) [8],
many studies have been carried out to improve the prediction accuracy of MeSH
heading suggestions [9, 10, 11]. In addition to NLM’s official solution MTI [12],
there are two other methods most closely related to our method [9, 10, 11]. The
first one is MetaLabeler, which was proposed in [13] and utilized by Tsoumakas
et al. in BioASQ Task1a for MeSH prediction [9]. In this method, firstly, for
each MeSH heading, a binary classification model was trained using linear SVM.
Secondly, a regression model was trained to predict the number of MeSH head-
ings for each citation. Finally, given a target citation, different MeSH headings
were ranked according to the SVM prediction score of each classifier, and the
top K MeSH headings were returned as the suggested MeSH headings, where K
is the number of predicted MeSH headings by the model. The second one is the
learning to rank (LTR) method, which was widely used in information retrieval
[14] and utilized by Lu et al. for automatic MeSH annotation [10, 11]. In this
method, each citation was deemed as a query and each MeSH headings as a doc-
ument. LTR method was utilized to rank candidate MeSH headings with respect
to target citation. The candidate MeSH headings came from similar citations (n-
earest neighbors). In our system, we also made use of the LTR framework for
predicting MeSH headings. However, in addition to the information from similar
citations, we also use the prediction scores from individual MeSH classifier to
improve the prediction accuracy.



3     Method

3.1   Data processing


We downloaded the whole PubMed database in an XML format. The cita-
tions without abstract or MeSH annotations were then filtered. Finally we ob-
tained 12,504,999 citations. Some citations have subsections, such as background-
s, methods and results, but we didn’t treat these subsections separately. Based
on target journals in BioASQ Task2a, we kept the latest 20,000 citations for
validation and testing, and 1,000,000 additional latest citations for training.




                                    1313
3.2   Tokenization

BioTokenizer6 [15] was used in our system for the tokenization task. We also
tried CoreNLP7 and doc2mat8 , but the performance of BioTokenizer was the
best. After the processing of BioTokenizer, the text of a citation was tokenized
and the stemming task was carried out simultaneously. Finally, based on the
tokenization result, we compiled a dictionary for converting the text of each
citation into a vector.


3.3   Primary Classifiers

Similar to MetaLabeler[9], we trained a binary classifier for each label (MeSH
heading). We call these binary classifiers as Primary Classifiers in our framework.
For efficiency, we used logistic regression instead of SVM to train these classi-
fiers. For a target citation, each Primary Classifier can predict the annotation
probability of the corresponding label.


3.4   Nearest Neighbors

Given a target citation, we used NCBI efetch9 to find its similar (neighbor) c-
itations. The MeSH headings from these neighbors were deemed as promising
candidates to annotate the target citation. It is generally believed that candi-
date MeSH headings from most similar citations are more important than those
from less similar citations. To measure the importance of each candidate MeSH
heading, we added up the similarity scores between the target citation and its
neighbor citations that contain this candidate MeSH heading. These similarity
scores could be also retrieved by NCBI efetch.


3.5   Learning to Rank Framework

Features

 We used a learning to rank framework to integrate multiple types of information,
such as the information from Primary Classifiers and nearest neighbors. For a
target citation, firstly we used the Primary Classifiers to calculate the annotation
probability (score) of every MeSH heading. Then we retrieved similar citations
for the neighbor scores. Finally, these two scores were considered as features in
the LTR framework. In the BioASQ task2a challenge, the default results of NLM
official solution MTI were also considered as a feature in the LTR framework.


Candidates
6
  See http://sifaka.cs.uiuc.edu/jiang4/software/BioTokenizer.pl
7
  See http://nlp.stanford.edu/software/corenlp.shtml
8
  See http://glaros.dtc.umn.edu/gkhome/files/fs/sw/cluto/doc2mat.html
9
  See http://www.ncbi.nlm.nih.gov/books/NBK25499/




                                     1314
  For a target citation, a MeSH label could be a candidate MeSH of this citation
if and only if it satisfied any of the two following requirements.


 1. The label appeared in the similar citations of the target citation;
 2. The labels Primary Classifier score was in the top 100 of all MeSH labels.

LTR method

  Each citation was treated as a query, and the candidate MeSH headings as
documents. Then LambdaMART[16] was used as the ranking method in the
learning to rank framework. The LTR training data contained about 30,000
citations from BioASQ task1a test set. 1000 decision trees were used in the
LambdaMART model without overfitting. After getting the LTR score of each
candidate MeSH, we used the best threshold which was tuned from the validation
set to determine how many labels we should return.


4     Experimental Results
4.1   Implementation
The whole project was coded in C++. We used some third part libraries in our
solution: Liblinear10 [17] for Logistic Regression, RankLib11 for LambdaMART
algorithm and JsonCpp12 for Input/Output json files. We also used OpenMP13
to make our task parallel.

4.2   Computational performance
The server we used for the challenge has 4 * Intel XEON E5-4650 2.7GHzs
CPU and 128GB RAM. The most computational expensive part is the training
of Primary Classifiers, which took 5 days. All other training tasks took about 1
day. However, the time cost for prediction is low. For annotating 10,000 citations,
it only took 2 hours.

4.3   Label based & Example based Performance
As shown in Table 2, we compared the performance of five different methods,
MTIFL, MTIDEF, directly mapping, MetaLabeler, and LTR. These methods
were evaluated on a test set of 9040 citations, which were published in BioASQ
task2a target journal between 2012 and 2013. By integrating multiple types
of information, LTR achieved the highest MiF of 0.61, followed by MTIDEF
10
   See http://www.csie.ntu.edu.tw/ cjlin/liblinear/
11
   See http://sourceforge.net/p/lemur/wiki/RankLib/
12
   See http://jsoncpp.sourceforge.net/
13
   See http://openmp.org/wp/




                                    1315
(0.572), MTIFL (0.564), Metalabeler (0.56) and directly mapping (0.27). Based
on this framework, in the last week of Batch 3 of BioASQ Task2a (annotated
articles: 1867/4533 on 6 June)14 , our system achieved the highest performance
in terms of both flat F-Measure (0.6199) and hierarchical F-Measure (0.5145)


Table 2. The performance comparsions of 5 methods. These methods are evaluated
on an offline testing set that has 9040 citations published between 2012 and 2013.

 Method                             MiP MiR MiF EBP EBR EBF MaP MaR MaF
 MTI FirstLine Inedx (MTIFL) 0.614 0.522 0.564 0.619 0.539 0.555 0.516 0.492 0.471
 Default MTI (MTIDEF)        0.574 0.571 0.572 0.578 0.591 0.564 0.513 0.537 0.494
 Directly Mapping            0.236 0.314 0.270 0.250 0.329 0.268 0.374 0.415 0.349
 MetaLabeler                 0.558 0.561 0.560 0.556 0.577 0.550 0.460 0.462 0.436
 LTR Without MTI Features 0.612 0.593 0.603 0.606 0.613 0.594 0.507 0.485 0.470
 LTR Ensemble                0.621 0.601 0.610 0.616 0.623 0.603 0.523 0.507 0.490




5      Discussions and Conclusions

Although our system performed very well in the competition, the system could
be further improved in several aspects. Firstly, for the local offline evaluation,
we used only flat measures (MiF) to tune our method, which leads to the un-
derperformance of our system in hierarchical measure (LCA-F). Considering the
significant difference between MiF and LCA-F, we could further improve the
performance of our system by using LCA-F to tune the model. Secondly, we
can consider some special MeSH headings separately. It is noticed that, for some
most frequent MeSH headings, such as check tags, direct prediction may improve
the annotation accuracy of these MeSH headings[18]. Finally, we have not used
any indexing rules in our system. Incorporating this kind of human knowledge
into the system would be a very promising direction for significantly increasing
the accuracy of MeSH heading recommendations.
    As a general framework, the LTR we used for MeSH heading recommenda-
tions can also be applied to some other types of tasks. Moreover, the performance
could be further improved if more information is integrated. As such, it raises
an interesting question as to what the upper bound of our system will be in the
presence of more information integrated in the LTR framework.


Acknowledgments. This work has been partially supported by National Nat-
ural Science Foundation of China (61170097), and Scientific Research Starting
Foundation for Returned Overseas Chinese Scholars, Ministry of Education, Chi-
na. Shanfeng Zhu would like to thank the China Scholarship Council for the
14
     See http://bioasq.lip6.fr/results/2a/




                                         1316
financial support on his visit at University of Illinois at Urbana-Champaign. We
would like to thank Hongning Wang and Mingjie Qian in UIUC for their helpful
suggestions and insightful discussion, thank Jieyao Deng in Fudan University
and Tianyi Peng in Tsinghua University for their help in coding works during
the competition.


References
 [1] Nelson, S. J., Schopen, M., Savage, A. G., Schulman, J. L., & Arluk, N.: The MeSH
     translation maintenance system: structure, interface design, and implementation.
     Medinfo, 11(Pt 1), 67–69. (2004)
 [2] Gu, J., Feng, W., Zeng, J., Mamitsuka, H., Zhu, S.: Efficient Semi-supervised
     MEDLINE document clustering with MeSH semantic and global content con-
     straints. IEEE Transactions on Cybernetics, 43 (4), 1265–1276, (2013)
 [3] Zhu, S., Zeng, J., Mamitsuka, H.: Enhancing MEDLINE document clustering by
     incorporating MeSH semantic similarity. Bioinformatics 25(15): 1944–1951 (2009)
 [4] Zhu, S., Takigawa, I., Zeng, J., Mamitsuka, H.: Field independent probabilistic
     model for clustering multi-field documents. Information Processing & Manage-
     ment. 45(5): 555–570 (2009)
 [5] Huang, X., Zheng, X., Yuan, W., Wang, F., Zhu, S.: Enhanced clustering of
     biomedical documents using ensemble non-negative matrix factorization. Infor-
     mation Science 181(11): 2293–2302 (2011)
 [6] Rajpal, D. K., Qu, X. A., Freudenberg, J. M., Kumar, V. D.: Mining emerging
     biomedical literature for understanding disease associations in drug discovery.
     Methods Mol Biol. 1159:171–206 (2014)
 [7] Theodosiou, T., Vizirianakis, I. S., Angelis, L., Tsaftaris, A., Darzentas, N.:
     MeSHy: Mining unanticipated PubMed information using frequencies of occur-
     rences and concurrences of MeSH terms. J Biomed Inform. 44(6):919–26. (2011)
 [8] Partalas, I., Gaussier, E., & Ngomo, A. C. N.: Results of the First BioASQ Work-
     shop. In: BioASQ@ CLEF (pp. 1-8). (2013)
 [9] Tsoumakas, G., Laliotis, M., Markantonatos, N., & Vlahavas, I. P.: Large-Scale
     Semantic Indexing of Biomedical Publications. In BioASQ@ CLEF. (2013)
[10] Huang, M., Neveol, A., Lu, Z.: Recommending MeSH terms for annotating
     biomedical articles. Journal of the American Medical Informatics Association,
     18(5): 660–667. (2011)
[11] Mao, Y., Lu, Z.: NCBI at the 2013 BioASQ challenge task: Learning to rank for
     automatic MeSH indexing[R]. Technical report. (2013)
[12] Mork, J. G., Jimeno-Yepes, A., & Aronson, A. R.: The NLM Medical Text Indexer
     System for Indexing Biomedical Literature. In BioASQ@ CLEF. (2013)
[13] Tang, L., Rajan, S., Narayanan, V.K.: Large scale multi-label classification via
     metalabeler. Proceedings of the 18th international conference on World Wide
     Web, 211–220 ACM (2009)
[14] Cao, Z., Qin, T., Liu, T. Y., Tsai, M. F., Li, H.: Learning to rank: from pairwise
     approach to listwise approach. Proceedings of the 24th international conference
     on Machine learning.129–136. ACM. (2007)
[15] Jiang, J., Zhai, C.: An empirical study of tokenization strategies for biomedical
     information retrieval. Information Retrieval, 10(4-5): 341–363. (2007)
[16] Burges, C. J.: From RankNet to LambdaRank to LambdaMART: An overview.
     Technical Report MSR-TR-2010-82, Microsoft Research. (2010)




                                      1317
[17] Fan, R. E., Chang, K. W., Hsieh, C. J., Wang, X. R., Lin, C. J.: LIBLINEAR: A
     library for large linear classification. The Journal of Machine Learning Research,
     9: 1871–1874. (2008)
[18] Yepes, A. J., Mork, J. G., Demner-Fushman, D., Aronson, A. R.: Comparison and
     combination of several MeSH indexing approaches. AMIA Annual Symposium
     Proceedings. 2013: 709–718. American Medical Informatics Association. (2013)




                                      1318