<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Combination of BERT and Data Oversampling for Answer Type Prediction</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Thang Ta Hoang</string-name>
          <email>tahoangthang@gmail.com</email>
          <email>thangth@dlu.edu.vn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olumide Ebenezer Ojo</string-name>
          <email>olumideoea@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olaronke Oluwayemisi A</string-name>
          <email>olaronke.oluwayemisi@gmail.com</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dalat University</institution>
          ,
          <addr-line>Lam Dong</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Instituto Politecnico Nacional (IPN), Centro de Investigacion en Computacion (CIC)</institution>
          ,
          <addr-line>Mexico City</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we address the Task 1 (of the SMART Task 2021) of predicting the answer categories and types based on target ontologies, which could be useful in knowledge-based Question Answering (QA) systems. We introduced our method by combining the power of BERT architectures with data oversampling via replacements of linked terms to Wikidata and dependent noun phrases to attain the state-ofthe-art performance. The accuracy on the DBpedia dataset is 98.5%, whereas NDCG@5 and NDCG@10 are 72.7% and 66.4% respectively. Our model has the best performance compared to other teams, with the accuracy score of 98% and Mean Reciprocal Rank (MRR) of 70% on the Wikidata dataset.</p>
      </abstract>
      <kwd-group>
        <kwd>Question Answering</kwd>
        <kwd>Answer Type Prediction</kwd>
        <kwd>Semantic Web Challenge</kwd>
        <kwd>ISWC</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        In Natural Language Processing (NLP), Knowledge Base Question Answering
(KBQA) is a task that involves searching for correct answers to a given natural
language question using knowledge bases (KB). While this task appears easy
to humans, it becomes a challenge for machines to detect the semantics in the
questions and match them to the KBs before choosing which answer is the best.
Language diversity, as well as semantic meanings, are barriers that create
lexical gaps between questions and answers in reality. At the moment, there are
two main approaches: semantic parsing-based (SP-based) methods and
information retrieval-based (IR-based) methods, with deep neural networks making
substantial contributions to improving their performance. [
        <xref ref-type="bibr" rid="ref12 ref27 ref6">6, 12, 27</xref>
        ].
      </p>
      <p>
        In SMART 2021 Semantic Web Challenge3, we engage Task 1 { category
and type prediction over two datasets, Wikidata and DBpedia [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Given a
      </p>
    </sec>
    <sec id="sec-2">
      <title>3 https://smart-task.github.io/2021/</title>
      <p>
        natural language question, our task is to build models that can predict answer
category and answer type by relying on a set of candidates from a target ontology.
There are 3 answer categories, "resource", "literal" and "boolean". If the
answer type is "boolean", the answer category will also be "boolean". If the
answer category is "literal", the answer type can be either "number", "date"
or "string". If the answer category is "resource", the answer type will be an
ontology class (DBpedia or Wikidata). Category prediction will be evaluated by
accuracy scores across all datasets. A metric lenient NDCG@k [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] (NDCG@5 and
NDCG@10) will be applied to the DBpedia dataset, and the Wikidata dataset
will be evaluated with a mean reciprocal rank (MRR).
      </p>
      <p>
        In this paper, we apply BERT classi ers [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and use an oversampling method
based on replacements of linked terms [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and dependent noun phrases analyzed
from the questions on the dataset. The paper is organized as follows: related
works will be presented in Section 2, followed by data analysis in Section 3.
In Section 4, we go over our methodology about augmenting and dividing the
dataset for the training process. Experimental results and error reports with
discussions are presented in Sections 5 and 6 respectively. Finally, in Section 7,
we focus on certain conclusions and describe further work.
2
      </p>
      <sec id="sec-2-1">
        <title>Related Works</title>
        <p>
          As the task description of this challenge, each question should belong to a unique
category class. As a result, category prediction is a multiclass classi cation
problem. On the other hand, type prediction is a multilabel classi cation in which a
given question can have multiple answer types. With answer categories and types
predicted, a KBQA system can reduce the searching time for possible answers
in the data [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ].
        </p>
        <p>
          If we consider category and type prediction as a problem of text classi
cation, there is a wide range of methods from logistic regression (LG), Naive
Bayes or Bayes networks [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], support vector machine (SVM) [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], random
forest (RF) [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], k-nearest neighbors (k-NN) [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], to deep networks (CNN, RNN,
GCN) [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] and many more hybrid approaches [
          <xref ref-type="bibr" rid="ref12 ref27 ref6">6, 12, 27</xref>
          ]. Due to the large
number of types in the corpus, type prediction is sometimes considered a
translation task in sequence-to-sequence models [
          <xref ref-type="bibr" rid="ref20 ref26">20, 26</xref>
          ], where a question is directly
inferred to answer types/relations directly or during knowledge validation.
        </p>
        <p>
          Referring to some of the papers from the previous years' challenge [
          <xref ref-type="bibr" rid="ref14 ref18 ref21">14, 18,
21</xref>
          ], BERT outperformed other approaches in predicting answer categories and
types. As a result, we opt to use BERT to tackle the problem.
        </p>
        <p>
          KBQA corpora are usually imbalanced with numerous natural language
questions, including rare questions created from language diversity and human
creativity. Therefore, oversampling and undersampling techniques are helpful to
reduce popular data and increase rare data until the dataset distribution is
more balanced or less biased. There are many methods known for oversampling
techniques such as SMOTE [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], ADASYN [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], and data augmentation (EDA [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ],
GenAug [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], contextual augmentation [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]) for text. In this paper, we employ a
simple oversampling technique to limit the number of rare types with fewer than
5 questions. Our technique will replace dependent noun phrases with the roots
and terms linked [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] to Wikidata by their aliases. Besides, cross-lingual data
augmentation should be applied to improve the questions in di erent languages
as well as overall outcome performance [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
3
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Dataset Analysis</title>
        <p>To determine the proper approach, we will rst do a preliminary analysis of the
datasets. Both Wikidata and DBpedia datasets are unbalanced and are classi ed
into three categories: resource, literal and boolean. The resource category
accounts for the majority of datasets, followed by literal boolean categories.
We conduct some investigations on older datasets, which are not signi cantly
di erent from newer datasets with noise removed.</p>
        <p>
          In the DBpedia dataset, we found a lot of questions with "empty" or None
answer types, as shown in Figure 1. One of the solutions for this problem is
that we can use a natural language inference (NLI) approach [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] to infer similar
questions, we then can grant new types to these questions. However, we prefer
to remove these empty types out of the training models and consider them as
noises.
        </p>
        <p>Type prediction is not di cult with the boolean and literal categories
because their answer types are simple. There are only a single type (boolean)
for the boolean category and three types: (date, number, and string) for the
literal category. In contrast, it is problematic with the resource category
because it contains 376 types of DBpedia and 3308 types of Wikidata. To reduce
the number of types in Wikidata, we only take the rst type of answer that will
be used for the experiment. We can also use ontology taxonomy of Wikidata to
lter the redundant types (parent-child cases) to get the lesser number of types,
but we will not do so in this work.</p>
        <p>The imbalanced features in the datasets appear in the answer types as shown
in Figure 2. When taking the rst type, which is prepared for the training process,
we notice that there are still a lot of types with only a few questions. There are
2032 types and 128 types corresponding to Wikidata and DBpedia datasets,
respectively, if counting response types with less than 5 questions. Clearly, it is
a challenge for text classi cation methods to detect enough di erences between
rare types. We debated whether to remove these rare types or keep them because
they do not contribute so much to the classi cation performance. Finally, we
decided to keep these types and apply an oversampling technique to increase the
number of questions of each type as many as possible, with the expectation that
the datasets are less imbalanced than before.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Our Methodology</title>
        <sec id="sec-2-3-1">
          <title>Preprocessing and Oversampling Methods</title>
          <p>For each question, spaCy4 v.2.3.2 was used to analyze the question structure to
get its semantic units, such as question type, subject, main verb (also ROOT),
terms (noun phrases, dependent noun phrases) to build the sentence template,
and apply entity linking (EL) methods to extract terms connected to
Wikidata. For spaCy components, we take en core web lg, STOP WORDS, lemmatizer
(Lemmatizer, ADJ, NOUN, VERB), and sentencizer pipeline.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4 https://spacy.io/</title>
      <p>The sentence template was built by a greedy algorithm which absorbs all
longest terms. First, we rank the terms according to their length. Then, for each
term, we search it in the sentence. If found, we put a pair of opening and closing
curly brackets f...g around it to form the sentence template. This algorithm
will ignore terms that are contained within larger ones.</p>
      <p>We realized that the question structure is also useful for Task 2 of this
challenge, which involves predicting the relations of questions. To easier understand
what we analyzed, take a look at the following example.
{
"question": "What periodical literature does Delta Air Lines use
,! as a moutpiece?",
"category": "resource",
"type": ["publication", "recurring", "intellectual work", "text",
,! "communication medium", "serial"],
"question_template": "What {periodical literature} does {Delta Air
,! Lines} {use} as {a mouthpiece}?",
"key_terms": ["periodical literature", "Delta Air Lines", "use",
,! "a mouthpiece"],
"subject": ["Delta Air Lines"],
"main verb": ["use"],
"entities": ["literature", "a mouthpiece", "mouthpiece"],
"question_type": "what",
"dependent_nouns": ["Delta Air Lines", "periodical
,! literature",...],
"el_terms": [ "Delta Air Lines": { "wikidata_id": "Q188920",
"label": "Delta Air Lines", "aliases": ["DAL", "Delta Air
Lines, Inc."},...],
,!
,!
...
}
There are some cases where terms are typos, so grammatical models should
be used to correct them. For each term, we depend on the API searching5 of
Wikipedia to x the typos for terms longer than eight characters. We assume
that a short term can lead to a wrong search. In some cases, the result may be
a term that is more popular than the term we desire. As a result, we may have
a wrong x that distorts the meaning of the original sentence. In the example,
we apply this API to x the typo moutpiece to mouthpiece.</p>
      <p>For the EL methods, we already built APIs for performing ELs using various
methods, such as Babelfy, OpenTapioca, Wiki er, and AIDA but we decided
to use TagMe, with WAT as an alternative. We gather linked terms to
Wikidata which have the link probability higher than 0.9. After that, we apply an
oversampling technique over linked terms to increase the number and
discrepancy of questions over rare types. We simply replace aliases of linked terms to</p>
    </sec>
    <sec id="sec-4">
      <title>5 https://www.mediawiki.org/wiki/API:Search</title>
      <p>their labels to create new questions. In the example, we can produce some new
questions like:
{
"question": "What periodical literature does Delta Air Lines use
,! as a moutpiece?",
"new questions": ["What periodical literature does DAL use as a
mouthpiece?", "What periodical literature does Delta Air
Lines, Inc. use as a mouthpiece?","What literature does Delta
Air Lines use as a moutpiece?",...],</p>
      <p>We generated more questions by replacing key terms to their roots, such
as periodical literature to literature. We assume that by using these
replacements, the models will be able to deal with the new data more e ectively.</p>
      <p>By using our strategy, we were able to adjust the number of questions and
rare types between the original and extended datasets, as shown in Table 1.
For each question, we only count single types rather than attening type lists
into strings. In comparison to the original dataset, the new datasets have nearly
three times more and less in roughly half of rare types than the original dataset.
The number of questions and answer types changes slightly when we apply some
minor preprocessing measures before the training process.
4.2</p>
      <sec id="sec-4-1">
        <title>Data Training and Models</title>
        <p>According to our initial experiments, we decided to split our datasets into two
groups (D1 and D2) each of which will be trained by two BERT models, one
for category prediction (also type prediction for the literal and boolean
categories) and the other for type prediction of the resource category. In total, we
have four BERT models for two datasets (two for DBpedia and two for
Wikidata).</p>
        <p>
          { D1: These datasets are used for category prediction and type prediction for
the boolean and literal categories. They are attened into 5 categories:
boolean, literal-date, literal-number, literal-string, and resource
as seen in the work of Setty and co-authors [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. We only use the boolean
and literal categories in the extended data, not the resource category,
because it already takes the majority of the original datasets.
{ D2: These datasets contains the questions belonging to the resource
category as well as their new questions generated from our oversampling method.
We will not present the graph distribution of answer types for questions, in
which the number of questions drops sharply from the popular to rare types.
        </p>
        <p>We adapt BERT pre-trained models to a speci c task, in this case text
classi cation. We set dropout with p = 0.1 and use linear regression to t
hidden size to n classes. In training process, we also use cross-entropy loss
and Adam optimization with learning rate = 2e-5. Other hyperparameters
are MAX LEN = 192, RANDOM SEED = 42, and BATCH SIZE = 8. We use a small
value of BATCH SIZE due to the limited resources on our server.
across both DBpedia and Wikidata datasets if the same metric is used.
Meanwhile, the numbers of types of Wikidata datasets are now slightly reduced via
the oversampling step.
5
5.1</p>
        <sec id="sec-4-1-1">
          <title>Experiments</title>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>Category and Type Prediction of Boolean and Literal</title>
      </sec>
      <sec id="sec-4-3">
        <title>Categories</title>
        <p>We apply bert-based-cased6 to D1 datasets to predict the categories and types
(only for the literal and boolean categories) over questions. As previously
stated, the questions are divided into ve categories: boolean, literal-date,
literal-number, literal-string, and resource. By this way, the model can
predict not only the answer category but also the answer type for questions which
belong to the literal category and the boolean category. The only thing left to
accomplish is to predict the answer types of questions in the resource category.
This task will be completed in the following section.</p>
        <p>We split D1 datasets into three subsets: train test, test set, and validation
set, in the ratio 8:1:1. After 10 epochs, we will save the best model based on the
highest validation accuracy. Since the organizers do not initially provide us with
the golden label set, we decide to use validation sets extracted from the original
datasets.
The previous section discusses type prediction of answers over the literal and
boolean categories. In this section, we do the same thing, but for the type
prediction of the resource category. We also apply bert-based-cased over
D2 datasets for answer types. As with D1, we divided D2 into three subsets:
training, testing and validation set with the ratio 8:1:1. Even though we apply
an oversampling method to the original dataset, as described in Section 3.2,
there are many rare types with fewer than ve questions in the datasets.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6 https://huggingface.co/bert-base-cased</title>
      <p>Table 5 shows the evaluation metrics on D2 datasets after at least 15 epochs.
Our suggestion is to train until train acc is at the highest value as possible to
maximize the model's ability to learn the rare types.</p>
      <sec id="sec-5-1">
        <title>Compare with the organizers' results</title>
        <p>Even though we achieve optimistic results from training data steps, we must
work on the test sets o ered by the organizers to have the nal evaluations.
We consider these sets as holdout sets, providing samples that do not appear
mostly in the training data. The category prediction will use accuracy scores as
the metric to evaluate our models. The metric lenient NDCG@k (NDCG@5 and
NDCG@10) will be applied to evaluate the performance of the type prediction
on the DBpedia dataset, whereas the MRR score will be used for the Wikidata
dataset.</p>
        <p>In Table 6, we obtained the accuracy score of 98.5%, NDCG@5 of 0.727,
and NDCG@10 of 0.664 for the DBpedia dataset. When compared to the best
performance, our results are among the best and quite competitive. Meanwhile,
our method performs best on Wikidata, with an accuracy of 98% and MRR of
0.7 as shown in Table 7.
6</p>
        <sec id="sec-5-1-1">
          <title>Error Reports and Discussions</title>
          <p>When working with the old datasets, we detect a tiny number of errors relating
to null answer types or the improper agreement between questions and answer
types. Some questions are too short and can not be analyzed precisely to get
semantic units, while others are too long. We consider them all as noises and
lter them out before training the data.</p>
          <p>Our analysis of question structure is not always correct. In some cases, we are
unable to obtain the necessary information, such as the question type, subject,
main verb, etc. The parsing method occasionally detects wrong dependent nouns
as shown in Table 8. In the rst example, the term 's husband leads to generate
a new wrong question "Who is the child of Ranavalona Ihusband?" when
replacing 's husband by husband (considered as the noun root). As a result,
we must avoid all replacements on possessive nouns containing 's. The second
example has no e ect on the replacements, but the parsing cannot split the long
term right ascension of malin 1 into two smaller terms, right ascension
and malin 1 to produce new questions.</p>
          <p>Example 1
question Who is the child of Ranavalona I's husband?
key terms "the child", "Ranavalona I", " 's husband "
question template Who is fthe childg of fRanavalona Igf 's husband g?
Example 2
question Is the right ascension of malin 1 less than 15.1398?
key terms " right ascension of malin 1 ", "15.1398"
question template Is the f right ascension of malin 1 g less than f15.1398g ?
Replacing dependent noun phrases in the creation of new questions does
improve the accuracy of category prediction. However, our intuition tells us that
we can distort the meaning of questions even in the least aspect. As a result, this
may a ect type prediction. We thus avoid applying this task to any dependent
noun phrases that are fewer than 8 characters long. We assume that the longer
phrases can keep the original meaning better. Since we only take linked terms
to Wikidata with those that have a link probability score over 0.9, we may leave
a lot of other helpful but equally correct linked terms with lower probabilities.</p>
          <p>Due to the occurrence of so many rare types, there are likely not enough
rare types in the training set after splitting the data into training, testing, and
validation sets. On one side, it helps the model's detection of similar types when
the rare types do not appear. On the other hand, this splitting may have an
impact on the performance of predicting rare types. Hence, we should place all
data on the training set or retrain the model until the accuracy score achieves
the best convergence across all sets.</p>
          <p>In the DBpedia dataset, the performances are relatively low in comparison to
other teams. This happens when we atten answer type lists to strings, instead
of having to use di erent models to detect single answer types based on their
ranking.
7</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>Conclusion</title>
          <p>In this paper, we participated in Task 1 of the SMART 2021 Semantic Web
Challenge, category and type prediction of answers using a set of hint ontologies.
We apply spaCy and TagMe to extract sentence components and linked terms
from questions. By using a simple oversampling method based on replacements
of linked terms and dependent nouns, we are able to expand the size of datasets
about three times, targeting to have as many questions as possible, especially
on rare answer types. Our project code can be accessed at GitHub 7.</p>
          <p>In the experiments, a pretrained BERT model, bert-base-cased, was used
to train D1 and D2 datasets to predict answer categories and types. For DBpedia,
NDCG@5 and NDCG@10 are 0.727 and 0.664 respectively, with an accuracy of
98.5%. The best results in the Wikidata dataset have an MRR score of 0.7 and
an accuracy of 98%. We discover that BERT models perform well in multiclass
and multilabel classi cation problems.</p>
          <p>In the future, we will improve the analysis parsing of question structure and
EL methods to add ontology information on top of the training data. We plan
to test various neural networks or hybrid approaches to search for a superior
method, as well as try to augment the dataset using entity linking methods
and multilingual translation. Finally, the semantic relationships between answer
types should be studied by linking to questions in order to minimize the number
of types and infer the answer types e ectively.</p>
        </sec>
        <sec id="sec-5-1-3">
          <title>Acknowledgement</title>
          <p>The work was done with partial support from the Mexican Government through
the grant A1-S-47854 of the CONACYT, Mexico, and by the Secretar a de
Investigacion y Posgrado of the Instituto Politecnico Nacional, Mexico, under
Grants 20211884, 20220859, and 20220553, EDI; and COFAA-IPN. The authors
thank the CONACYT for the computing resources brought to them through the</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>7 https://github.com/thangth1102/SMART 2021 Task1</title>
      <p>Plataforma de Aprendizaje Profundo para Tecnolog as del Lenguaje of the
Laboratorio de Supercomputo of the INAOE, Mexico and acknowledge the support
of Microsoft through the Microsoft Latin America PhD Award.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neumayer</surname>
          </string-name>
          , R.:
          <article-title>Hierarchical target type identi cation for entity-oriented queries</article-title>
          .
          <source>In: Proceedings of the 21st ACM international conference on Information and knowledge management</source>
          . pp.
          <volume>2391</volume>
          {
          <issue>2394</issue>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chawla</surname>
            ,
            <given-names>N.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bowyer</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hall</surname>
            ,
            <given-names>L.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kegelmeyer</surname>
            ,
            <given-names>W.P.</given-names>
          </string-name>
          :
          <article-title>Smote: synthetic minority over-sampling technique</article-title>
          .
          <source>Journal of arti cial intelligence research 16</source>
          ,
          <volume>321</volume>
          {
          <fpage>357</fpage>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Demszky</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Transforming question answering datasets into natural language inference datasets</article-title>
          . arXiv preprint arXiv:
          <year>1809</year>
          .
          <volume>02922</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>S.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gangal</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitamura</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hovy</surname>
          </string-name>
          , E.:
          <article-title>Genaug: Data augmentation for netuning text generators</article-title>
          . arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>01794</volume>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ferragina</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scaiella</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Tagme: on-the- y annotation of short text fragments (by wikipedia entities)</article-title>
          .
          <source>In: Proceedings of the 19th ACM international conference on Information and knowledge management</source>
          . pp.
          <volume>1625</volume>
          {
          <issue>1628</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A survey on complex question answering over knowledge base: Recent advances and challenges</article-title>
          . arXiv preprint arXiv:
          <year>2007</year>
          .
          <volume>13069</volume>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>E.H.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karypis</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Text categorization using weight adjusted k-nearest neighbor classi cation</article-title>
          . In:
          <article-title>Paci c-asia conference on knowledge discovery and data mining</article-title>
          . pp.
          <volume>53</volume>
          {
          <fpage>65</fpage>
          . Springer (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>He</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bai</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>E.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Adasyn:
          <article-title>Adaptive synthetic sampling approach for imbalanced learning</article-title>
          .
          <source>In: 2008 IEEE international joint conference on neural networks (IEEE world congress on computational intelligence)</source>
          . pp.
          <volume>1322</volume>
          {
          <fpage>1328</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. Ka e,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Kanan</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          :
          <article-title>Answer-type prediction for visual question answering</article-title>
          .
          <source>In: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          . pp.
          <volume>4976</volume>
          {
          <issue>4984</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Kobayashi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Contextual augmentation: Data augmentation by words with paradigmatic relations</article-title>
          . arXiv preprint arXiv:
          <year>1805</year>
          .
          <volume>06201</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Kowsari</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jafari Meimandi</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heidarysafa</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mendu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barnes</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brown</surname>
          </string-name>
          , D.:
          <article-title>Text classi cation algorithms: A survey</article-title>
          .
          <source>Information</source>
          <volume>10</volume>
          (
          <issue>4</issue>
          ),
          <volume>150</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Lan</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>W.X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wen</surname>
            ,
            <given-names>J.R.:</given-names>
          </string-name>
          <article-title>A survey on complex knowledge base question answering: Methods, challenges and solutions</article-title>
          .
          <source>arXiv preprint arXiv:2105.11644</source>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gliozzo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ngonga</surname>
            <given-names>Ngomo</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>A.C.</given-names>
            ,
            <surname>Usbeck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Rossiello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>U.</surname>
          </string-name>
          :
          <article-title>Semantic answer type and relation prediction task</article-title>
          (smart
          <year>2021</year>
          ). arXiv (
          <year>2022</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Nikas</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fafalios</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tzitzikas</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Two-stage semantic answer type prediction for question answering using bert and class-speci city rewarding</article-title>
          .
          <source>In: SMART@ ISWC</source>
          . pp.
          <volume>19</volume>
          {
          <issue>28</issue>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Ojo</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ta</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adebanji</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>Automatic hate speech detection using deep neural networks and word embedding</article-title>
          .
          <source>Computacion y Sistemas</source>
          <volume>26</volume>
          (
          <year>2022</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Ojo</surname>
            ,
            <given-names>O.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adebanji</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Performance study of n-grams in the analysis of sentiments</article-title>
          .
          <source>Journal of the Nigerian Society of Physical</source>
          Sciences pp.
          <volume>140</volume>
          {
          <issue>143</issue>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Ojo</surname>
            ,
            <given-names>O.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gelbukh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorov</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adebanji</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Sentiment analysis in texts on economic domain</article-title>
          .
          <source>In: Proceedings of the 19th Mexican International Conference on Arti cial Intelligence - MICAI2020</source>
          . Mexico City,
          <source>Mexico (October</source>
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Perevalov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Both</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Augmentation-based answer type classi cation of the smart dataset</article-title>
          .
          <source>In: SMART@ ISWC</source>
          . pp.
          <volume>1</volume>
          {
          <issue>9</issue>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Pranckevicius</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcinkevicius</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Comparison of naive bayes, random forest, decision tree, support vector machines, and logistic regression classi ers for text reviews classi cation</article-title>
          .
          <source>Baltic Journal of Modern Computing</source>
          <volume>5</volume>
          (
          <issue>2</issue>
          ),
          <volume>221</volume>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Rossiello</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdelaziz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bornea</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gliozzo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naseem</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kapanipathi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Generative relation linking for question answering over knowledge bases</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>321</volume>
          {
          <fpage>337</fpage>
          . Springer (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Setty</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Semantic answer type prediction using bert: Iai at the iswc smart task 2020</article-title>
          . arXiv preprint arXiv:
          <volume>2109</volume>
          .06714 (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McCann</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keskar</surname>
            ,
            <given-names>N.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiong</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
          </string-name>
          , R.: Xlda:
          <article-title>Cross-lingual data augmentation for natural language inference and question answering</article-title>
          . arXiv preprint arXiv:
          <year>1905</year>
          .
          <volume>11471</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>How to ne-tune bert for text classi cation?</article-title>
          <source>In: China National Conference on Chinese Computational Linguistics</source>
          . pp.
          <volume>194</volume>
          {
          <fpage>206</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Wasim</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Asim</surname>
            ,
            <given-names>M.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khan</surname>
          </string-name>
          , M.U.G.,
          <string-name>
            <surname>Mahmood</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>Multi-label biomedical question classi cation for lexical answer type prediction</article-title>
          .
          <source>Journal of biomedical informatics 93</source>
          ,
          <issue>103143</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Wei</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zou</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          : Eda:
          <article-title>Easy data augmentation techniques for boosting performance on text classi cation tasks</article-title>
          . arXiv preprint arXiv:
          <year>1901</year>
          .
          <volume>11196</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>A seq2seq-based approach to question answering over knowledge bases</article-title>
          .
          <source>In: Joint International Semantic Technology Conference</source>
          . pp.
          <volume>170</volume>
          {
          <fpage>181</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>A survey of question answering over knowledge base</article-title>
          .
          <source>In: China Conference on Knowledge Graph and Semantic Computing</source>
          . pp.
          <volume>86</volume>
          {
          <fpage>97</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>