<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Applicability of Automatically Generated Thesauri to Text Classification in Specific Domains</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ksenia Lagutina</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Shchitov</string-name>
          <email>ivan.shchitov@e-werest.org</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nadezhda Lagutina</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ilya Paramonov</string-name>
          <email>ilya.paramonov@fruct.org</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>P.G. Demidov Yaroslavl State University</institution>
          ,
          <addr-line>Sovetskaya Str. 14, 150003, Yaroslavl</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The paper is devoted to comparison of the quality of text classification with the use of manually and automatically generated thesauri. For this purpose, the authors applied the BM25 algorithm with word features based on thesaurus's relations between terms. The experiments, conducted with text corpora from three specific domains (medicine, economics, and sport), showed that using an automatically generated thesaurus provides nearly the same classification quality as the manually created one. These results make the authors' approach promising for text classification in many specific domains where no thesaurus is available, as it allows to avoid consumption of high amount of resources for manual thesaurus creation.</p>
      </abstract>
      <kwd-group>
        <kwd>text classification</kwd>
        <kwd>specialized thesaurus</kwd>
        <kwd>BM25</kwd>
        <kwd>automatic thesaurus generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Automated topical text classification aims to order unstructured corpora and
assigns raw texts to pre-defined main topics. The most popular classification
algorithms are developed and studied primarily for processing news, reviews,
and spam [
        <xref ref-type="bibr" rid="ref1 ref5">1,5</xref>
        ]. These texts mostly contain common words, so their classification
does not depend on a domain.
      </p>
      <p>
        Conversely, domain-specific texts usually have non-trivial structure, many
terms, and other peculiar properties that complicate text mining. One of the way
to simplify processing and extract semantic information is the thesaurus use. A
specialized thesaurus contains almost all domain’s terms and reflects semantic
relations between them [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Studies show that embedding a thesaurus into text
processing algorithms increases classification quality significantly [
        <xref ref-type="bibr" rid="ref11 ref4">4,11</xref>
        ].
      </p>
      <p>However, construction of high-quality thesauri that can be used for text
mining, requires much expert’s time, therefore the number of thesauri in open access
is limited. An alternative for human-made thesauri is automatically created ones.
They are less popular in text classification than both manually created
specialized thesauri and general purpose thesauri. Also, the difference between results
achieved using thesauri of different types is rarely investigated in text mining
papers.</p>
      <p>The goal of the authors’ research was to compare efficiency of manually and
automatically generated specialized thesauri in application to the topical text
classification. Each thesaurus was used in conjunction with BM25 algorithm to
take into account relations between classes and words from texts. The
experiments revealed potential of both types of thesauri in use. Moreover, this
comparison allowed to figure out whether the automatically created thesauri can be
used in domains that do not have a thesaurus yet.</p>
      <p>The paper is structured as follows. Section 2 overviews the state-of-the-art in
text classification with thesauri. Section 3 explains the main steps of the
classification method and details of the thesaurus embedding. Section 4 describes text
corpora and the evaluation procedure used in experiments. Section 5 provides
numerical results of experiments and explains them. The conclusion summarizes
the paper, shortly describes key results and discusses future research.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related work</title>
      <p>Almost all existing algorithms that classify texts by topics have the same main
steps:</p>
      <sec id="sec-2-1">
        <title>1. Extraction of keyword candidates from texts.</title>
        <p>2. Computation of statistical and linguistic features of the candidates.
3. Ranking candidates using different functions and machine learning
algorithms.
4. Choice of candidates with the best scores as text classes.</p>
        <p>
          A thesaurus can be embedded into such algorithms in several ways. The
first one is to extend the candidate list by thesaurus terms related to keywords
from the initial set. This approach is implemented in the GENIE system [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
It classifies news texts using a geographical names dictionary and Eurowordnet
thesaurus. Experiments showed that such an approach allows to increase
precision and recall by 20 % comparing with the same algorithm without a thesaurus.
Another example is the method that combines the thesaurus with character-level
ConvNets [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Testing on news, reviews, DBPedia, and Yahoo Answers corpora
showed that using the WordNet thesaurus decreases the number of errors by
8–15 %.
        </p>
        <p>
          An important feature of such methods is the usage of a general purpose
thesaurus. It fits well for news, reviews, and other texts with common words,
but may work worse with documents from a concrete domain containing many
specific terms [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
        </p>
        <p>
          The second way to use the thesaurus is to compute additional features for
candidates. Nagaraj et al. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] proposed a method that finds semantic relations
between words in WordNet or Wikipedia and counts their number for each word.
The best results was shown on 20NewsGroup and Classic3 subsets, about 85–
95 %; F-measure for the business Reuters-21578 corpus was lower, from 60 to
90 % depending on the training set size.
        </p>
        <p>
          The system of review classification [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] divides reviews into two groups:
positive and negative. The solution of this task requires sentiment analysis, so
Bollegala et al. automatically created a specialized thesaurus with terms that express
human emotions and opinions. The classification algorithm computes the
number of thesaurus relations for each keyword in the text and adds this score to
the feature vector. As a result, accuracy raised from 60–70 % to 70–80 %. Such
high scores were reached because the thesaurus was built for a specific domain
taking into account its features and peculiarities.
        </p>
        <p>Summarily, the usage of a thesaurus allows to improve quality of text
classification, and results are better if the thesaurus contains information about
semantic relations between the words from the texts’ domain. For text
classification both manually and automatically generated thesauri can be used, but
comparison of the classification quality reachable with the use of these two classes
of thesauri is not properly presented in modern research.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Method of text classification using automatically generated thesaurus</title>
      <p>3.1</p>
      <sec id="sec-3-1">
        <title>Thesaurus creation</title>
        <p>
          The main goal of the authors’ research was to examine how well text classification
can be performed with the use of automatically created specialized thesauri. Such
thesauri were generated from several domain-specific corpora with the use of the
algorithm from the authors’ previous work [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. It includes the following steps of
text processing:
        </p>
        <sec id="sec-3-1-1">
          <title>1. Term extraction using TextRank algorithm.</title>
          <p>2. Construction of different semantic relations (associations, hyponym—hypernyms,
and synonyms) with the use of the combination of statistical and semantic
methods.
3. Isolated term removal.</p>
          <p>This approach provides a thesaurus with the large number of semantic
relations between terms. Comparison with the existing thesaurus showed that
automatically constructed one has quite good recall and very high precision for
synonyms.</p>
          <p>The main advantage of the algorithm is that it processes data fully
automatically and does not require any expert’s work, so the thesaurus can be quickly
constructed for any domain. Moreover, according to the authors’ previous
results, such a thesaurus usually has enough terms and relations of quite good
quality. Therefore, it can provide additional information for methods that solve
text mining problems.
3.2</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Text classification</title>
        <p>After construction the specialized thesauri were applied to the text classification
task.</p>
        <p>
          In the authors’ experiment text classification is performed by the existing
existing unsupervised classification algorithm BM25 [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] modified to use the
thesaurus. This algorithm takes as an input a corpora, where texts are not assigned
with topics, a list of classes, and an automatically constructed thesaurus from
the previous section. It contains the following steps:
        </p>
        <sec id="sec-3-2-1">
          <title>1. Extract all words from texts and compute their frequency.</title>
          <p>2. For each class create the list of related thesaurus terms.
3. Rank the text-class pairs using the terms frequency and the BM25 algorithm.</p>
          <p>Firstly, the algorithm extracts individual words from texts, builds the
inverted index for each text document, and calculate word frequency.</p>
          <p>Secondly, the thesaurus is browsed for the closest related terms of the
following classes: synonyms, hyponyms, and hyperonyms of the first order. Associations
are not included because they often link terms that can indicate different topic,
for example “blood clot” and “heart”. On the contrary, synonyms, hyponyms, and
hyperonyms obviously reflect semantic relations between words from a common
topic, so they can help to juxtapose texts with their classes better.</p>
          <p>
            Finally, the BM25 algorithm is applied to the text-class pairs for ranking each
document by the query terms. This algorithm is a popular approach for many
text mining and information retrieval tasks, including text classification [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ].
Also it does not require any additional parameters and training with treated
samples, so it can be easily extended by a thesaurus. These features make the
algorithm suitable for the authors’ research.
4
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation procedure and used text corpora</title>
      <p>In the classification experiments the authors used three text corpora from
different domains:
– PubMed text corpus (https://www.nlm.nih.gov/databases/download/pubmed_
medline.html). It has 63 classes and 1000 medical articles with 154 850
words.
– Reuters text corpus (http://www.daviddlewis.com/resources/testcollections/
reuters21578/). It has 15 classes and 1534 articles from economics domain
with 294 813 words.
– BBCSport text corpus (http://mlg.ucd.ie/datasets/bbc.html). It has 5
classes and 737 texts about sport with 253 667 words.</p>
      <p>Thesauri were generated automatically based on the chosen corpora. To
compare the classification quality when using automatically and manually
generated thesaurus the authors used the well-known MeSH thesaurus (https:
//www.nlm.nih.gov/mesh/meshhome.html) and STW thesaurus (http://zbw.
eu/stw/version/latest/about.en.html) for medical and economics domains
correspondingly.</p>
      <p>
        The tool for extracting keywords from a text corpus is based on the Topical
PageRank keyword extraction algorithm [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. It was implemented by the authors
in Python programming language. The tool takes a file with a text corpus as
an input parameter, reads all of texts, extracts keywords, and writes them to
separate files.
      </p>
      <p>The text classification algorithm is also implemented as a tool in Python. It
contains three modules: parser, query processor, ranking function. The parser
module reads and parses files with a list of classes and a set of texts to create
the data structures for further calculations. The query processor takes each class
from the list and scores the documents based on the class terms using the ranking
function. The ranking function is an implementation of the BM25 algorithm.</p>
      <p>Algorithm’s outcomes were compared with results of manual classification
that are provided with each corpus. For evaluation the authors chose most
popular quality measures: micro-average precision, recall, F-measure, and accuracy.
The precision is the fraction of documents actually belonging to given classes
among all documents that the algorithm assigned to classes. The recall is the
fraction of documents found by the algorithm that are belong to given classes
among all documents of classes. The F-measure is the harmonic mean of the
precision and recall. The accuracy is the fraction of the retrieved documents for
which the classifier made a correct decision.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <p>The results of experiments achieved with the use of the text corpora described
in the previous section are contained in Table 1 and Table 2.</p>
      <p>Table 1 displays results in absolute values. The first column of the table
contains the name of text corpora, the second one—the type of used thesauri.
“Auto” means the automatically generated thesaurus based on the text corpus.
The third and fourth columns display the number of texts in the corpus and the
number of predefined classes. For example, texts from the PubMed corpus are
classified by the following classes: toxicity, pharmacology, immunology, surgery,
et al. The fifth and sixth columns display the number of text-class pairs found
using the algorithm described in Section 3.2 and correct correspondence between
texts and classes given in advance correspondingly.</p>
      <p>These results show that the automatically generated thesaurus allows to find
more text-class pairs compared to the manually constructed one. Comparing the
original number of text-class pairs (the last column of Table 1) with the others,
we can see the following trend. The algorithm with manual thesaurus leaves
some texts unclassified or juxtaposes texts with smaller number of classes than
it should be. On the contrary, the generated thesaurus provides too many pairs
for PubMed and Reuters corpora, but for the BBCSport corpus this number is
close to the original.</p>
      <p>Table 2 displays results in relative values. The results show that the
experiments with the automatically generated thesauri have greater recall and smaller
precision than the experiments with the thesauri constructed manually. It
happens because the algorithm with the generated thesaurus finds too many texts
for classes. From one hand, it allows to find more right answers. From the other
hand, it provides a singnificant number of redundant text-class pairs. This
tendency is particularly takes part for the Reuters corpus. However, the experiment
with the BBCSport corpus shows that the generated thesaurus can provide quite
high precision for some texts (about 40 % against 33 and 12.5 % for the other
domains).</p>
      <p>The F-measure and accuracy differ slightly for algorithms with both
thesaurus types, so the difference between them is not essential.</p>
      <p>Summarily, the results allow to conclude that the automatically created
thesaurus can be successfully applied to the classification task. Meanwhile, the
approach with the automatically generated thesaurus has the following advantages.</p>
      <p>Firstly, it is suitable for different domains. Experiments for medicine and
economics domains were conducted with the same parameters of thesaurus
construction and embedding and the results showed the same trend for all measures.
Therefore, we can expect similar classification quality for other domains.</p>
      <p>Secondly, the thesaurus construction goes fully automatically and does not
require the expert’s labour and parameters adjustment. So the thesaurus can be
created and applied easily for any corpus.</p>
      <p>Nevertheless, generated thesauri do not provide enough high precision and
accuracy of results in some cases, therefore the research of their use should be
continued.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion and future research</title>
      <p>In this paper the authors compared quality of text classification when using
manually and automatically created thesauri. The experiments showed that
using manual thesauri provides higher precision and accuracy, but automatical
ones allow to find more right classes and generally are not significantly behind.</p>
      <p>The algorithm was used for two different domains with both thesaurus kinds
and the result’s trend was the same. So classification of corpora from other
domains would probably have the similar quality.</p>
      <p>The results of the research allows to assert that the automatically generated
thesaurus is unconditionally applicable for text classification. The authors
experimented with the sport news text corpus, where there is no specialized thesaurus,
and got quite high measures about 40–70 %.</p>
      <p>The future research concerns varying of thesaurus properties and
classification algorithm’s parameters.</p>
      <p>The important thesaurus features are semantic relations. When thesauri are
applied to information retrieval and text mining, the least used relations are
associations. They are dissimilar and often have different semantics. So, the
associations can be classified into several subtypes that can be differently used
in text classification. This can be important for the case when the algorithm
with an automatically generated thesaurus provides low precision, because it
processes a large number of thesaurus’ relations and extracts too many false
positive text-class pairs. Probably, discrimination of the relations would allow
to eliminate redundant text classes.</p>
      <p>The algorithm’s parameters are weights of words in the scoring function.
During the text analysis step the algorithm calculates features for classes and
keywords taking into account thesaurus’s relations between them. The
significance of various relations can be differentiated depending on the domain. If the
scoring function would assign different weights for different relations, the
algorithm would classify texts using the particular domain’s features that can lead
it to make smaller number of mistakes.</p>
      <p>Summarily, the further investigation should find out types of thesaurus
relations that reflect links between classes and texts better for a particular domain.
It would allows to vary parameters of the classification algorithm and classify
texts more precisely.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Aggarwal</surname>
            ,
            <given-names>C.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhai</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>A survey of text classification algorithms</article-title>
          .
          <source>In: Mining text data</source>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>222</lpage>
          . Springer (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Aitchison</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gilchrist</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bawden</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Thesaurus construction and use: a practical manual</article-title>
          . Psychology Press (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bollegala</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weir</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carroll</surname>
          </string-name>
          , J.:
          <article-title>Cross-domain sentiment classification using a sentiment sensitive thesaurus</article-title>
          .
          <source>IEEE transactions on knowledge and data engineering 25(8)</source>
          ,
          <fpage>1719</fpage>
          -
          <lpage>1731</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Garrido</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buey</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Escudero</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peiro</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ilarri</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mena</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>The GENIE system: Classifying documents by combining mixed-techniques</article-title>
          .
          <source>In: International Conference on Web Information Systems and Technologies</source>
          . pp.
          <fpage>231</fpage>
          -
          <lpage>246</lpage>
          . Springer (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Jindal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malhotra</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Techniques for text classification: Literature review and current trends</article-title>
          .
          <source>Webology</source>
          <volume>12</volume>
          (
          <issue>2</issue>
          ),
          <volume>1</volume>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>K.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walker</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robertson</surname>
            ,
            <given-names>S.E.</given-names>
          </string-name>
          :
          <article-title>A probabilistic model of information retrieval: development and comparative experiments</article-title>
          .
          <source>In: Information Processing and Management</source>
          . vol.
          <volume>36</volume>
          , pp.
          <fpage>809</fpage>
          -
          <lpage>840</lpage>
          . London, UK (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Lagutina</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mamedov</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lagutina</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paramonov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shchitov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Analysis of relation extraction methods for automatic generation of specialized thesauri: Prospect of hybrid methods</article-title>
          .
          <source>In: Proceedings of the 19th Conference of Open Innovations Association FRUCT. Jyvaskyla, Finland</source>
          ,
          <fpage>7</fpage>
          -
          <lpage>11</lpage>
          November
          <year>2016</year>
          . pp.
          <fpage>138</fpage>
          -
          <lpage>144</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Automatic keyphrase extraction via topic decomposition</article-title>
          .
          <source>In: Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <fpage>366</fpage>
          -
          <lpage>376</lpage>
          . Association for Computational Linguistics (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Nagaraj</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thiagarasu</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vijayakumar</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A novel semantic level text classification by combining NLP and Thesaurus concepts</article-title>
          .
          <source>IOSR Journal of Computer Engineering</source>
          <volume>16</volume>
          (
          <issue>4</issue>
          ),
          <fpage>14</fpage>
          -
          <lpage>26</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Nam</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mencía</surname>
            ,
            <given-names>E.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fürnkranz</surname>
          </string-name>
          , J.:
          <article-title>Large-scale multi-label text classification-revisiting neural networks</article-title>
          .
          <source>In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases</source>
          . pp.
          <fpage>437</fpage>
          -
          <lpage>452</lpage>
          . Springer (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Sanchez-Pi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martí</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>A.C.B.</given-names>
          </string-name>
          :
          <article-title>Improving ontology-based text classification: An occupational health and security application</article-title>
          .
          <source>Journal of Applied Logic</source>
          <volume>17</volume>
          ,
          <fpage>48</fpage>
          -
          <lpage>58</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Trieschnigg</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pezik</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , De Jong, F.,
          <string-name>
            <surname>Kraaij</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rebholz-Schuhmann</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>MeSH Up: effective MeSH text classification for improved document retrieval</article-title>
          .
          <source>Bioinformatics</source>
          <volume>25</volume>
          (
          <issue>11</issue>
          ),
          <fpage>1412</fpage>
          -
          <lpage>1418</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , LeCun, Y.:
          <article-title>Character-level convolutional networks for text classification</article-title>
          .
          <source>In: Advances in neural information processing systems</source>
          . pp.
          <fpage>649</fpage>
          -
          <lpage>657</lpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>