<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Detailed Descriptions for Text Classification Applications</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gorka Artola</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>German Rigau</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of the Basque Country (UPV/EHU), Faculty of Informatics</institution>
          ,
          <addr-line>Manuel Lardizabal pasealekua, 1, 20018 Donostia-San Sebastián</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The development of effective domain specific text classification systems generally requires the availability of large amounts of high quality labeled domain data. In domains such as BioNLP, eHealth, NLP for Legal Purposes, NLP for Social Media and Journalism, etc., obtaining the needed volume of data manually-labeled by domain experts is not usually feasible or affordable. In this work we propose a new method for text classification based on the use of detailed class descriptions instead of using a large number of labeled instances for training the classifiers. Our method, experimentally tested on the classification of titles of scientific papers on the domain of the Sustainable Development Goals of the United Nations, consistently outperforms mainstream NLP classification approaches, radically faster and at a fraction of their cost due to it does not need a previous process of hand-labelling thousands of samples.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;text classification</kwd>
        <kwd>class descriptions</kwd>
        <kwd>sustainable development goals</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>1.1. Description of the Task</title>
      </sec>
      <sec id="sec-1-2">
        <title>1.2. Summary of Contributions</title>
        <sec id="sec-1-2-1">
          <title>The main contributions of this work are:</title>
          <p>• We propose the use of already existing or
handcrafted detailed descriptions of the classes for
multi-label sentence classification with PLMs as
a better performing and more resource-eficient
way than investing in manual-labelling of
samples.
• We propose guidelines to decide between working
in the generation of detailed descriptions or
investing in hand-labeling samples, considering the
availability or not of either detailed descriptiosn
or labeled samples, and depending on this
decision, to select the most appropriate multi-class
classification technique with PLMs.
• We establish a new SOTA for classification of
titles of scientific papers by SDGs.
• We publicly disclose the most relevant datasets
and code used in our experimentation.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        PLMs, such as  [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and   [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], have achieved
state-of-art performance on many NLP tasks [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and
among them on multi-class text classification [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The
research community has developed several lines of work
to improve text classification in diferent data availability
scenarios:
• When we have abundant unlabeled data related to
the specific application domain but lack of labeled
data Weakly-supervised techniques [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] show
promising results. The most recent of them
leverage the capacities of transformer-based PLMs,
like LOTClass [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which uses label names as
initial keywords and augments the keywords
with  ’s MLM module to train
classification models on unlabeled data, or FastClass by
      </p>
      <sec id="sec-2-1">
        <title>Xia et al. [10], that proposes the use of dense text representation techniques in semantic spaces. • In the case we have large amounts of unlabeled data, but non related to the domain or the task,</title>
        <p>
          Unsupervised text classification techniques
[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] show the capacity to improve text
classification.
• When lacking of any data, PLMs allow the
generation of improved semantically meaningful text
representation models like Sentence-BERT [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ],
and the enunciation of the text classification task
as a natural language inference (NLI) problem are
the SOTA techniques [
          <xref ref-type="bibr" rid="ref2 ref25">2</xref>
          ]. Recently, Schopf et al.
[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] proposed the combination of the
embeddingbased method Lbl2Vec and transformer-based
PLMs to further improve their performance on
unsupervised text classification.
        </p>
        <p>
          Focusing specifically on the use of descriptions of
classes, there is also a body of research studying question
answering task embodiment for text classification like
the one proposed by Chai et al. [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. These techniques
in combination with strategies for the development of
better class descriptions [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], label noise reduction
methods [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], and the recent emergence of generative large
language models (LLMs) [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] set the ground for future
research in the use of descriptions for specific domain
NLP applications.
        </p>
        <p>
          Regarding the classification of scientific papers by
SDGs, related literature describes several approaches
grouped in two diferent working principles:
• Boolean query based approaches for information
retrieval from databases like the ones developed
by Elservier [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], Digital Science [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ],
the University of Bergen [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ], the University of
Auckland [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] and the AURORA European
University Alliance*. The most relevant among them
is the AURORA SDG Queries v5 method [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] [25].
• NLP based methods like the AURORA-ML*
method [26] [27]. This approach comprises 169
   [28] based models, one
for each SDG target, fine-tuned on abstracts of
papers obtained with the AURORA SDG Queries
v5 method.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental Setup</title>
      <p>The following Datasets contain all the data used in our
experimentation:
• The "SDG-Descriptions Dataset" comprising 447
sentences of diferent semantic natures
(SDGHeadlines, SDG-Titles, SDG-Targets and
SDGIndicators) developed by the UN and published
in a dedicated website* describing the 17 SDGs.</p>
      <p>Altogether, we name the samples of this dataset
SDG-Descriptions. Considering we have 447
descriptive sentences of SDGs, we have built
training dataset with 430 entailment samples and 7,152
contradiction samples.
• The "Paper Titles Gold Dataset" with 9,382
scientific paper titles labeled by experts. This dataset
includes two families of samples that are disjoint,
i.e., no paper title appears in both families:
– "Positive samples" of titles labeled to one</p>
      <p>
        or more specific SDGs they are related to.
– "Negative samples" of titles labeled to one
or more specific SDGS they are not related
to.
*https://aurora-universities.eu/
*https://github.com/Aurora-Network-Global/TMD
*https://metadata.un.org/sdg/
This Gold Dataset is a subset of the AURORA
dataset [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], elaborated surveying expert
scientist, and that shows a human agreement level of
70.10% in this task. We have developed several
splits of this dataset for training, development
and evaluation purposes. The train-split contains
more than 8,000 positive samples, and the
testsplit contains 2,086 labeled paper titles unevenly
distributed by SDG but with the same amount
of positive and negative samples for each one of
them.
      </p>
      <p>The Classification Approaches and Models we have
experimented with are:
• Fine-tuning classifiers from general PLMs. After
experimenting with diferent general PLMs we
have selected  [29] for its better
results. We have developed diferent classifiers
ifne-tuning  on diferent amounts
of samples of the train-split of the Paper Titles
Gold Dataset, on diferent amounts of samples of
the SDG-Descriptions Dataset, and on the
combinations of both of them.
• Zero-shot classification with NLI-PLMs.
After experimenting with diferent NLI-PLMs and
querying/prompting setups, we have obtained
the best results querying    
[29] with either SDG-Headlines or SDG-Titles
and prompting the queries with the expression
"The subject is ".
• Few-shot classification. Building upon the
previous approaches, we have developed a new
method for multi-class text classification
finetuning     on pairs of
SDGDescription sentence/SDG-Headline, and
applying the resulting model for NLI based zero-shot
classification of paper titles. For the initial
finetuning we have built a training dataset with
samples composed by pairs of sentences, being the
ifrst each one of the SDG description sentences
and the second each one of the SDG-Headlines
prompted with the text "The Sustainable
Development Goal is". This way we have generated
17 samples from each SDG description sentence,
out of which the one pairing the sentence with
its correspondent SDG-Headline is labeled as
"entailment" and all the rest (16) as "contradiction".</p>
      <p>We generate a zero-shot classifier fine tuning the
    model with this dataset.</p>
      <p>The classification of each test sample is finally
performed querying the model with the
SDGHeadlines and prompting the queries with the
expression "This is".</p>
      <p>In our study we simulate this scenario fine-tuning
the base models with diferent numbers of labeled
samples (75%, 50%, 25%, 10%, 5% and 1%) of the
train split of the "Paper Titles Gold Dataset".
• A "Class descriptions available" scenario, in
which diferent amounts and types of
description sentences of the classes are available. We
simulate this in our study by splitting the
SDG-Descriptions Dataset in sub-sets of
SDGHeadlines, SDG-Titles, SDG-Targets and
SDGIndicators and fine-tuning the base models in
accumulative combinations of them.</p>
      <p>Considering that a paper title may be related to several Considering these simulations of scenarios, we have
SDGs, our Metrics on the experiments consider true studied how the baseline and the diferent
SDGpositives (TP) the right predictions on positive samples, Description based models evolve with increasing
numfalse positives (FP) the wrong predictions on negative bers of available samples and descriptive sentences for
samples and true negatives (TN) the right predictions training. In Figure 2 we can observe that the general
baseon negative samples. The Prediction Criterion used line (blue line) requires almost 3,000 labeled samples to
in this analysis of the results is Topk-3, i.e., the top 3 overcome our most simple model exclusively trained with
scores given by the models for each tested sample are 447 description sentences (yellow line). Furthermore, if
considered predictions for all considerations. we continue training our description based models with</p>
      <p>The current SOTA for the studied task and domain is increasing numbers of hand-labeled samples, we can
obthe top macro averaged F1-score of 55% ofered by the serve that the obtained fine-tuned classifier (light-green
AURORA-ML method referenced in section 2. In the ex- line), beats the top F1-score of the general baseline (+8000
perimentation we have observed that the F1-score regis- labeled samples) with only around 500 labeled samples
tered in a vanilla fine-tuning of  on the full additional to the SDG-Descriptions. Furthermore, our
train split of the "Paper Titles Gold Dataset" goes above few-shot classification model defines a new estate of art
60%. Therefore, we have considered this vanilla approach for classification of scientific papers by SDGs when using
our Baseline or the analysis of the impact of the use of all SDG-Descriptions and hand-labeled samples. At this
SDG-Descriptions. In the zero-shot approach the consid- point, the peak measured F1-score is 71.01%, slightly over
ered baseline is the direct use of     the human agreement level of 70.10% observed in the
with a collection of keywords, namely SDG-Subjects, also AURORA dataset.
enunciated by the UN and related to the SDGs that we Up to now we have studied the results as a whole, but
have not considered part of the SDG-Descriptions be- the task includes 17 diferent classes that may behave
difcause they are not shaped as the descriptive sentences ferently. Table 2 shows the detailed global and per SDG
we intend to study. results of the test performed with our Few-Shot approach</p>
      <p>These choices are the result of an extensive experi- on the Titles-Test split of the Paper Titles-Gold dataset.
mentation process comprising diferent PLMs and meta- SDG 6 "Clean water and sanitation" and SDG 17
"Partnerparameters looking for the best performing ones. ship for the goals" show the worst results. The model has
been trained with 21 sentences describing SDG 6 and 46
4. Results sentences of SDG 17, similar or higher than the number
of sentences used to train other much better
performTable 1 shows a comparison between the best macro- ing SDGs like SDG 7 "Afordable and clean energy" (13
averaged F1-scores obtained with our description-based sentences, F1-score 83.93%) or SDG 3 "Good health and
models and the baselines. Our few-shot classifica- well-being" (42 sentences, F1-score 67.7%). This suggests
tion method using 447 publicly available SDG- that there is no clear correlation or proportion between
Descriptions overcomes the general baseline trained the number of sentences included in the description and
with over 8,000 hand-labeled samples. On the other hand, the performance of the model, and that the reasons for a
our zero-shot classification using SDG-Descriptions lags better classification may relay on other features
probafar behind the zero-shot baseline bly related to the semantics of the description sentences</p>
      <p>For the analysis of these results we will consider the and the sentences to be classified. The study of the
feafollowing two scenarios of data availability: tures that make a description good for this classification
approach are lines for further research.
• A "Labeled samples available" scenario, in which</p>
      <p>diferent amounts of labeled samples are available.</p>
    </sec>
    <sec id="sec-4">
      <title>5. Error Analysis</title>
      <p>Overall, the behaviour of the model seems to follow
what could be expected by common sense on the
sceWe focus the error analysis on the results of the few-shot nario we are working on, considering (i) that each tested
model, the best performing model among those that use paper title is most likely related with several SDGs but
exclusively SDG-Descriptions for training and classifi- not in the same extent, (ii) that the label given to each
cation. Table 3 summarizes how the model gives right, test sample is not necessarily the one of the SDG they are
wrong or inconclusive predictions. More than 90% of most related to, and (iii) that the more SDGs a paper title
the good predictions are obtained with the first (Top 1 - is related to, the lower score it will give at each one of
74.00%) and second (Top 2 - 17.79%) highest scores. The them individually. Coherently, the scores observed in the
average scoring pattern gives a relatively high value at "no coincidence" predictions of the positive samples
hapTop 1 (0.55-0.77) and drops significantly at every next pen to be ones with lowest scores. Also, the low average
prediction, scoring in the range 0.03-0.15 at Top 2 and scores registered in the false positive coincidences in Top
0.002-0.02 at Top 3. Nevertheless, the highest average 2 and Top 3 can be explained as those debatable cases
Top 2 (0.1514) and Top 3 (0.0214) scores correspond to that even with human observers reduce the agreement
right predictions obtained at second and third guesses. level to the previously mentioned 70.1%. Nevertheless,
Both right and wrong predictions at Top 1 score on aver- the following results appear to be relevant failures of the
age around 0.77, meaning that the model is particularly model worth to be analysed in detail:
mistaken in the wrong predictions.</p>
      <p>Deepening one step further, the test dataset has a par- • The 63.02% of false positive coincidences with the
ticular set of samples; those that have been labeled pos- highest scores (0.7795) at Top 1 prediction.
itively or negatively by more than one expert. They • The 12 bad predictions or false positive
coincicould be referred as "strong true samples" or "strong false dences on strong false samples.
samples" if they are either positive or negative labeled • The high rate of undetected true positives on SDG
samples. In opposition, we call the test samples labeled 6 and SDG 17.
by a single person as "weak true samples" or "weak false
samples". Table 4 shows the results of the test on these Appendix A shows several examples of these failures.
particular samples. The model has been able to classify Regarding the false positives in weak false samples, the
correctly all the strong true samples with a particularly wrong guesses are absolutely arguable and may fall in
high average Top 1 score of over 0.88, but at the same the side of the measured roughly 30% of human
disagreetime has classified incorrectly around 30% (12) of the ment level, with the exception of the example of SDG
strong false samples. 17 "Partnership for the goals" with the title "Tuple-based
semantic and structural mapping for sustainable
interoperability" not objectively relatable with this SDG. When
it comes to the false positives related to strong false
samples, that have happened exclusively for samples of the
SDG 3 "Good health and well-being", we can observe
several possible reasons for the failures like:
• Debatable or arguable labelling.
• A possible tendency of the model to relate tobacco
with health (SDG 3), and a tendency of experts not
to do it when the paper titles refer to its economic
dimensions.
• A dificulty of the model to distinguish between
animal health and human health.</p>
      <p>In the case of the undetected positive samples of SDG
6 "Clean water and sanitation" and SDG 17 "Partnership
for the goals", all cases appear to be very debatable. An
explanation may be that in these cases the titles of the
papers do not describe properly the contents of the paper,
or even may be misleading, but the experts have labeled
the papers not by their title but by their content. For
instance, the paper titled "Local renewable energy
cooperatives: revolution in disguise?", may be related with
the SDG 6 "Clean water and sanitation", but the title itself
suggests it may be more related to SDG 7 "Afordable and
clean energy" as the model predicts, or the paper titled
"Sustainability of small water supplies: Lessons from a
Brazilian program (SESP/FSESP)" may of course be
related to SDG 17 "Partnership for the goals" but the title
suggests it may be mainly related to SDG 6 "Clean water
and sanitation" as the model predicts.</p>
      <p>These phenomena are most likely related to the
evident overlaps that exist between the SDGs. Figures 3
and 4 depict the co-occurrence and confusion matrices of
the test. The co-occurrence matrix plots all Topk-3
predictions of the model on the positive samples of the test
dataset. Generally the model predicts more frequently
the right SDG, but we can also observe that SDG 15 "Life
on land" is remarkably more predicted than the other
SDGs, followed by SDG 12 "Responsible consumption
and production", SDG 1 "No poverty" and SDG 8 "Decent
work and economic growth". SDG 15 "Life on land" is
even more prevalent than the labeled SDG in the case
of SDG 6 "Clean water and sanitation", SDG 13 "Climate
action", SDG 14 "Life below the water" and SDG 17
"Partnership for the goals". The confusion matrix plots whe
wrong predictions of the model. In this case most
freThis work ofers initial experimental evidences that using
detailed descriptions of the main classes that shape an
specific domain has the potential to benefit Text
Classification. All the experiments reported have been developed
classifying automatically scientific papers to UN SDGs.</p>
      <p>The use of class descriptions may reduce significantly
or even eliminate the need to develop hand-labeled
samples for training NLP models, reducing drastically the
development cost. Depending on the availability on
descriptions of classes we recommend:
quently mistaken prediction is for SDG 12 "Responsible
consumption and production", followed in this case by
the same SDGs outstanding in the co-occurrence matrix.</p>
      <p>According to these results, the SDGs that most overlap
with the rest are the SDG 15 "Life on land", the SDG 12
"Responsible consumption and production", SDG 1 "No
poverty" and SDG 8 "Decent work and economic growth".</p>
      <p>This may mean that these SDGs are the ones that most
diversely may impact the UN 2030 Agenda for Sustainable
Development, what could be an excellent bonus insight
ofered by the model, but once again, this may be related
only to the diferent quality of the descriptions of each
SDG, and for sure a question worth to be further studied.
6. Conclusions and future work
for zero-shot exclusively with class descriptions,
have the potential to outperform conventional
classifiers fine-tuned on PLMs with thousands of
hand labeled samples.
• In the case of both detailed descriptions and
labeled samples available: Conventional PLM
classifiers fine-tuned with a combination of class
descriptions and labeled samples have the potential
to reduce the need of labeling by an order of
magnitude, being able to establish a new SOTA in our
case study.
• On the contrary, on a pure zero-shot approach,
in cases with only a single keyword or
description sentence available per class, the classical
prompted keyword classification seems to be
better than any similar description sentence based
classifier.
language. Also, not all descriptive sentences ofer the
same improvement potential: single sentences describing
the whole class (SDG titles) and collection of single
sentences describing each one a particular relevant aspect
of the class (SDG targets) contribute the most.</p>
      <p>Finally, the results of this initial experimental study
suggest the following future lines of research:
• Extending the study to further specific domain</p>
      <p>NLP applications to generate further evidence
about the potential benefits of using class
descriptions and grasp its limitations.
• Apply the use of class descriptions in methods
more sophisticated than the conventional NLP
approaches applied in this work to validate or
refuse the hypothesis that advanced NLP
techniques like generative LLMs and QA tasks may
also benefit from them.
• Deep dive in what makes a description good for</p>
      <p>NLP applications and explore how advanced
description development and improvement
techniques can contribute.</p>
      <p>Gold SDG
# Headline</p>
      <p>Good health
3 and wellbeing</p>
      <p>Quality
4 education</p>
      <p>Gender
5 equality</p>
      <p>Clean water
6 and sanitation</p>
      <p>Decent work
8 and economic
growth</p>
      <p>Industry,
9 innovation and
infrastructure</p>
      <p>Sustainable
11 cities and
communities</p>
      <p>Responsble
12 consumption</p>
      <p>and production
# Paper title
False postives in weak false samples</p>
      <p>Food insecurity and efectiveness of behavioral
1 interventions to reduce blood pressure
, New York City, 2012-2013</p>
      <p>Global governance for facilitating access to
2 medicines: Role of world health organization</p>
      <p>Equipping Preservice Elementary Teachers
3 for Data Use in the Classroom</p>
      <p>How to study varieties of opposition to
gender+ equality in Europe?: Lessons from this
4 book, conceptual building blocks, and puzzles
to address</p>
      <p>RETRACTED ARTICLE: Comparative advantage
5 analysis for water utilization in Hubei province
based on NRCA model</p>
      <p>A study on factors afecting the youth employment
6 rate: Focusing on data from 31 cities and counties</p>
      <p>in Gyeonggi-do, South Korea
7</p>
      <p>Analysis of the inclusions in 38Si7 spring steel
with fatigue failure</p>
      <p>The development and transition of urban walking
8 grey space in China, based on a unique model "</p>
      <p>Langpeng"</p>
      <p>Corporate sustainability in emerging markets:
9 Insights from the practices reported by the</p>
      <p>Brazilian retailers</p>
      <p>Sensitivity analysis with the regional climate model
10 COSMO-CLM over the CORDEX-MENA domain 13 Climate action</p>
      <p>Rainforest tourism, conservation and management:
11 Challenges for sustainable development 15 Life on land</p>
      <p>Capitalizing on Criminal Accomplices:
12 Considering the Relationship between</p>
      <p>Co-ofending and Illegal Earnings</p>
      <p>Tuple-based semantic and structural mapping Partnership
13 for a sustainable interoperability 17 for the goals
False positives in strong false samples # Headline</p>
      <p>Nature, scope and use of economic evaluation of</p>
      <p>Good health
14 healthcare programmes: With special reference 3 and wellbeing
to Pakistan</p>
      <p>Endovascular Aortic Repair for Thoracic Good health
15 Aortic Injuries 3 and wellbeing</p>
      <p>Comparison between online and ofline price of Good health
16 tobacco products using novel datasets 3 and wellbeing</p>
      <p>An Assessment of the Forward-Looking Hypothesis Good health
17 of the Demand for Cigarettes 3 and wellbeing</p>
      <p>Mycobacterium marinum infection in fish and man:</p>
      <p>Good health
18 Epidemiology, pathophysiology and management; 3 and wellbeing</p>
      <p>a review
Undetected SDG 6 and SDG 17 true positives # Headline</p>
      <p>An exploration of the boundaries of ‘community’ in</p>
      <p>Clean water
19 community renewable energy projects: Navigating 6</p>
      <p>and sanitation
between motivations and context</p>
      <p>Typology of future clean energy communities: An Clean water
20 exploratory structure, opportunities, and challenges 6 and sanitation</p>
      <p>A review of renewable energy investment</p>
      <p>Partnership
21 in the BRICS countries: History, models, 17 for the goals
problems and solutions</p>
      <p>Sustainability of small water supplies: Lessons Partnership
22 from a brazilian program (SESP/FSESP) 17 for the goals</p>
      <p>Peace, justice
16 and institutions</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          , Attention is all you need,
          <year>2017</year>
          . arXiv:
          <volume>1706</volume>
          .
          <fpage>03762</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roth</surname>
          </string-name>
          ,
          <article-title>Benchmarking zero-shot text classification: Datasets, evaluation and entailment approach</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLPIJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>3914</fpage>
          -
          <lpage>3923</lpage>
          . URL: https://aclanthology.org/D19-1404. doi:
          <volume>10</volume>
          .18653/ v1/
          <fpage>D19</fpage>
          -1404.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Vanderfeesten</surname>
          </string-name>
          , E. Spielberg,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gunes</surname>
          </string-name>
          ,
          <article-title>Survey data of "Mapping Research Output to the Sustainable Development Goals (SDGs)"</article-title>
          ,
          <year>2020</year>
          . URL: https: //doi.org/10.5281/zenodo.3813230. doi:
          <volume>10</volume>
          .5281/ zenodo.3813230.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/ N19-1423. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Narasimhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Salimans</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Improving language understanding by generative pre-training</article-title>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Min</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Sulem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P. B.</given-names>
            <surname>Veyseh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. H.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Sainz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Agirre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Heinz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roth</surname>
          </string-name>
          ,
          <article-title>Recent advances in natural language processing via large pre-trained language models: A survey</article-title>
          ,
          <year>2021</year>
          . arXiv:
          <volume>2111</volume>
          .
          <fpage>01243</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>A survey on text classification: From traditional to deep learning</article-title>
          ,
          <source>ACM Transactions on Intelligent Systems and Technology (TIST) 13</source>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Han,
          <article-title>Weaklysupervised neural text classification</article-title>
          ,
          <source>in: Proceedings of the 27th ACM International Conference on Information and Knowledge Management</source>
          , CIKM '18,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2018</year>
          , pp.
          <fpage>983</fpage>
          -
          <lpage>992</lpage>
          . URL: https: //doi.org/10.1145/3269206.3271737. doi:
          <volume>10</volume>
          .1145/ 3269206.3271737.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Han,
          <article-title>Text classification using label names only: A language model self-training approach</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>9006</fpage>
          -
          <lpage>9017</lpage>
          . URL: https: //aclanthology.org/
          <year>2020</year>
          .emnlp-main.
          <volume>724</volume>
          . doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .emnlp-main.
          <volume>724</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Chang,</surname>
          </string-name>
          <article-title>FastClass: A time-eficient approach to weakly-supervised text classification</article-title>
          ,
          <source>in: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Abu Dhabi, United Arab Emirates,
          <year>2022</year>
          , pp.
          <fpage>4746</fpage>
          -
          <lpage>4758</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          . emnlp-main.
          <volume>313</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Thangaraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sivakami</surname>
          </string-name>
          ,
          <article-title>Text classification techniques: A literature review</article-title>
          ,
          <source>Interdisciplinary Journal of Information, Knowledge, and Management</source>
          <volume>13</volume>
          (
          <year>2018</year>
          )
          <fpage>117</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          , Sentence-BERT:
          <article-title>Sentence embeddings using Siamese BERT-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>3982</fpage>
          -
          <lpage>3992</lpage>
          . URL: https://aclanthology.org/ D19-1410. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1410.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Schopf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Braun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Matthes</surname>
          </string-name>
          ,
          <article-title>Evaluating unsupervised text classification: Zero-shot and similaritybased approaches</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2211</volume>
          .
          <fpage>16285</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Chai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Description based text classification with reinforcement learning</article-title>
          ,
          <source>in: International Conference on Machine doi:10</source>
          .5281/zenodo.4883250.
          <string-name>
            <surname>Learning</surname>
          </string-name>
          ,
          <year>2020</year>
          . [25]
          <string-name>
            <given-names>F.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vanderfeesten</surname>
          </string-name>
          , Evaluation on accu-
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dietze</surname>
          </string-name>
          ,
          <article-title>Generating ef- racy of mapping science to the United Nations' Susfective label description for label-aware sentiment tainable Development Goals (SDGs) of the Aurora classification</article-title>
          ,
          <source>Expert Syst. Appl</source>
          .
          <volume>213</volume>
          (
          <year>2023</year>
          ).
          <source>URL: SDG queries</source>
          ,
          <year>2021</year>
          . URL: https://doi.org/10.5281/ https://doi.org/10.1016/j.eswa.
          <year>2022</year>
          .
          <volume>119194</volume>
          . doi:10. zenodo.4964606. doi:
          <volume>10</volume>
          .5281/zenodo.4964606. 1016/j.eswa.
          <year>2022</year>
          .
          <volume>119194</volume>
          . [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Vanderfeesten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Jaworek</surname>
          </string-name>
          , L. Keßler, AI for map-
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>X.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ruan</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . Zhao,
          <article-title>ping multi-lingual academic papers to the United Enhancing robust text classification via category Nations' Sustainable Development Goals (SDGs)</article-title>
          , description, 2022 IEEE International Conference
          <year>2022</year>
          . URL: https://doi.org/10.5281/zenodo.6487606.
          <article-title>on Data Mining (ICDM) (</article-title>
          <year>2022</year>
          )
          <fpage>151</fpage>
          -
          <lpage>160</lpage>
          . doi:
          <volume>10</volume>
          .5281/zenodo.6487606.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hou</surname>
          </string-name>
          , [27]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jaworeck</surname>
          </string-name>
          ,
          <string-name>
            <surname>SDG BERT</surname>
          </string-name>
          <article-title>- Multi-language MultiY</article-title>
          . Min,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Yang, label BERT model for classifying texts to SusY</article-title>
          . Chen,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <surname>X.</surname>
          </string-name>
          <article-title>Tang, tainable Development Goals (SDGs) based on AuZ</article-title>
          . Liu, P. Liu,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>A survey of large rora SDG Query Model v5</article-title>
          ,
          <year>2022</year>
          . URL: https: language models,
          <year>2023</year>
          . arXiv:
          <volume>2303</volume>
          .18223. //doi.org/10.5281/zenodo.7304547. doi:
          <volume>10</volume>
          .5281/
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>B. ayabalasingham</surname>
          </string-name>
          , R. Boverhof, K. Agnew, zenodo.7304547. L. Klein, Identifying research supporting the united [28]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , nations sustainable development goals,
          <year>2019</year>
          . URL:
          <article-title>BERT: pre-training of deep bidirectional transhttps://elsevier.digitalcommonsdata.com/datasets/ formers for language understanding</article-title>
          ,
          <source>CoRR 87txkw7khs/1</source>
          . doi:
          <volume>10</volume>
          .17632/87txkw7khs.1. abs/
          <year>1810</year>
          .04805 (
          <year>2018</year>
          ). URL: http://arxiv.org/abs/
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>M.</given-names>
            <surname>Rivest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kashnitsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bédard-Vallée</surname>
          </string-name>
          ,
          <year>1810</year>
          .
          <volume>04805</volume>
          . arXiv:
          <year>1810</year>
          .04805. D. Campbell,
          <string-name>
            <given-names>P.</given-names>
            <surname>Khayat</surname>
          </string-name>
          , I. Labrosse, S. Pinheiro, [29]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          , A. MoHenrique; Provençal,
          <string-name>
            <given-names>G.</given-names>
            <surname>Roberge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>James</surname>
          </string-name>
          , Im- hamed,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer,
          <article-title>BART: proving the scopus and aurora queries to identify denoising sequence-to-sequence pre-training for research that supports the united nations sustain- natural language generation, translation, and comable development goals (sdgs)</article-title>
          <year>2021</year>
          ,
          <year>2021</year>
          . URL: prehension, CoRR abs/
          <year>1910</year>
          .13461 (
          <year>2019</year>
          ). URL: http: https://elsevier.digitalcommonsdata.com/datasets/ //arxiv.org/abs/
          <year>1910</year>
          .13461. arXiv:
          <year>1910</year>
          .
          <volume>13461</volume>
          . 9sxdykm8s4/. doi:
          <volume>10</volume>
          .17632/9sxdykm8s4.4.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Roberge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kashnitsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>James</surname>
          </string-name>
          ,
          <article-title>Elsevier 2022 sustainable development goals A. Examples of wrong predictions (sdg) mapping, 2022</article-title>
          . URL: https://elsevier. digitalcommonsdata.com/datasets/6bjy52jkm9. doi:
          <volume>10</volume>
          .17632/6bjy52jkm9.1.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>D.</given-names>
            <surname>Science</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wastl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Porter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Draux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hook</surname>
          </string-name>
          ,
          <source>Contextualizing sustainable development research</source>
          ,
          <year>2020</year>
          . URL: https://digitalscience. ifgshare.com/articles/report/Contextualizing_ Sustainable_Development_Research/12200081. doi:
          <volume>10</volume>
          .6084/m9.figshare.
          <volume>12200081</volume>
          .
          <year>v2</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>C.</given-names>
            <surname>Armitage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lorenz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mikki</surname>
          </string-name>
          ,
          <article-title>Replication data for: Mapping scholarly publications related to the Sustainable Development Goals: Do independent bibliometric approaches get the same results</article-title>
          ?,
          <year>2020</year>
          . URL: https://doi.org/10.18710/98CMDR. doi:
          <volume>10</volume>
          . 18710/98CMDR.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <article-title>Mapping research to the sustainable development goals (sdgs</article-title>
          ),
          <year>2023</year>
          . URL: https://www. researchsquare.com/article/rs-2544385/v2. doi:
          <volume>10</volume>
          .21203/rs.3.rs-
          <volume>2544385</volume>
          /v2.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Vanderfeesten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Otten</surname>
          </string-name>
          , E. Spielberg,
          <article-title>Search Queries for "Mapping Research Output to the Sustainable Development Goals (SDGs)"</article-title>
          <source>v5.0.2</source>
          ,
          <year>2020</year>
          . URL: https://doi.org/10.5281/zenodo.4883250.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <source>2 Zero hunger 2 3</source>
          <volume>4 5 6 8 9 3 3 3 3 7 7 7 11 12 13 15 16 17 14 3 12 8 10 8 11 12 1 8 15 12 1 12 15 8 12 8 3 11 8 12 Top</volume>
          <source>1 Top 2 Top 3 Top 1 Top 2 Top 3</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>