=Paper= {{Paper |id=Vol-2936/paper-20 |storemode=property |title=Query-Focused Extractive Summarisation for Finding Ideal Answers to Biomedical and COVID-19 Questions |pdfUrl=https://ceur-ws.org/Vol-2936/paper-20.pdf |volume=Vol-2936 |authors=Diego Molla,Urvashi Khanna,Dima Galat,Vincent Nguyen,Maciej Rybinski |dblpUrl=https://dblp.org/rec/conf/clef/MollaKGNR21 }} ==Query-Focused Extractive Summarisation for Finding Ideal Answers to Biomedical and COVID-19 Questions== https://ceur-ws.org/Vol-2936/paper-20.pdf
Query-Focused Extractive Summarisation for
Finding Ideal Answers to Biomedical and COVID-19
Questions
Macquarie University Participation at BioASQ Synergy and BioASQ9b Phase B

Diego Mollá1,2 , Urvashi Khanna1 , Dima Galat1 , Vincent Nguyen2,3 and
Maciej Rybinski3
1
  Macquarie University, Australia
2
  CSIRO Data61, Australia
3
  Australian National University, Australia


                                         Abstract
                                         This paper presents Macquarie University’s participation to the BioASQ Synergy Task, and BioASQ9b
                                         Phase B. In each of these tasks, our participation focused on the use of query-focused extractive sum-
                                         marisation to obtain the ideal answers to medical questions. The Synergy Task is an end-to-end question
                                         answering task on COVID-19 where systems are required to return relevant documents, snippets, and
                                         answers to a given question. Given the absence of training data, we used a query-focused summarisa-
                                         tion system that was trained with the BioASQ8b training data set and we experimented with methods
                                         to retrieve the documents and snippets. Considering the poor quality of the documents and snippets
                                         retrieved by our system, we observed reasonably good quality in the answers returned. For phase B of
                                         the BioASQ9b task, the relevant documents and snippets were already included in the test data. Our
                                         system split the snippets into candidate sentences and used BERT variants under a sentence classifica-
                                         tion setup. The system used the question and candidate sentence as input and was trained to predict the
                                         likelihood of the candidate sentence being part of the ideal answer. The runs obtained either the best
                                         or second best ROUGE-F1 results of all participants to all batches of BioASQ9b. This shows that using
                                         BERT in a classification setup is a very strong baseline for the identification of ideal answers.

                                         Keywords
                                         BioASQ, Synergy, query-focused summarisation, Biomedical, COVID-19, BERT




1. Introduction
Supervised approaches to query-focused summarisation have the inherent problem of the
paucity of annotated data. This problem has been highlighted, for example, by [1], and the
biomedical domain is no exception. The BioASQ Challenge provides annotated data for multiple
tasks, including question answering [2]. While small in comparison with other data sets (the

CLEF 2021 – Conference and Labs of the Evaluation Forum, September 21–24, 2021, Bucharest, Romania
" Diego.Molla-Aliod@mq.edu.au (D. Mollá); Urvashi.Khanna@mq.edu.au (U. Khanna); dima.galat@gronade.com
(D. Galat); Vincent.Nguyen@anu.edu.au (V. Nguyen); maciek.rybinski@data61.csiro.au (M. Rybinski)
~ https://researchers.mq.edu.au/en/persons/diego-molla-aliod (D. Mollá); https://ngu.vin (V. Nguyen);
https://people.csiro.au/R/M/maciek-rybinski (M. Rybinski)
 0000-0003-4973-0963 (D. Mollá); 0000-0003-2345-5596 (U. Khanna); 0000-0003-1787-8090 (V. Nguyen)
                                       © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings
                  http://ceur-ws.org
                  ISSN 1613-0073       CEUR Workshop Proceedings (CEUR-WS.org)
training data set for BioASQ9b contains 3,742 questions), there may be enough to train or
fine-tune systems that have been pre-trained with other data sets. The problem of paucity of
annotated data, however, becomes critical for urgent tasks on new domains such as question
answering on biomedical papers related to COVID-19. In early 2021, BioASQ organised the
Synergy task where systems are required to develop various stages of an end-to-end question
answering system. In particular, given a question phrased in plain English, participating systems
were expected to retrieve relevant documents from the CORD-19 collection [3] and relevant
snippets. Optionally, the systems could complete the final stage of question answering by
returning exact and/or ideal answers. There was no annotated training data available for this
very specific task.
   This paper describes our contribution to the BioASQ Synergy task and phase B of the
BioASQ9b challenge.1 For the BioASQ Synergy task, we use a system that has been trained on
the BioASQ8b training data, whereas for phase B of the BioASQ9b challenge we explore the
use of Transformer architectures. In particular, we integrate BERT variants and fine-tune them
with the BioASQ9b training data.
   Prior work reports the success of BERT architectures for various tasks, by simply adding
a task-specific layer and fine-tuning the system [4]. BERT has also been used for finding
the ideal answers in BioASQ. For example, [5] used BERT in both an unsupervised sentence
cosine similarity setup and a supervised sentence regression setup, and [6] compared the use of
BERT embeddings with word2vec embeddings in a setup that directly modelled the interaction
between sentence embeddings of the question and the candidate sentence, and incorporated
sentence position. In our participation in BioASQ9b Phase B, we experimented with a simpler
architecture compared with [6], and obtained results that were among the top participating
systems2 . These good results suggest that the internal Transformer-based architecture of BERT
suffices to model the interaction between the question and the candidate sentence.
   This paper is structured as follows. Section 2 describes our contribution to the Synergy task.
Section 3 describes our participation in BioASQ9b. Section 4 summarises and concludes this
paper.


2. Synergy
Our contribution to the Synergy task focused on leveraging the use of a pre-trained question
answering system. In particular, we used one of the systems proposed by [6], which was
trained on the BioASQ8b training data, and was designed as a classifier that identified whether
a candidate sentence was part of the ideal answer.
   Figure 1 shows the architecture of the question answering system. This corresponds to the
system referred to as “NNC” by [6]. The input consists of a question, a candidate sentence, and
the candidate sentence position. The system uses Word2Vec trained on PubMed data to obtain
the word embeddings of the question and the sentence. These word embeddings are converted
to sentence embeddings through a layer of bi-directional LSTM chains. The interaction between

    1
      Code associated with this paper is available at https://github.com/dmollaaliod/bioasq-synergy-public and
https://github.com/dmollaaliod/bioasq9b-public.
    2
      As ranked by preliminary ROUGE results provided by the BioASQ organisers at the time of writing this paper
 sentence position
  sentence           word embeddings       sentence embeddings



                                       BiLSTM
                                                                            relu       sigmoid
                                                                                         ∫︀
             word2vec
  question




                                       BiLSTM              ×

                                                       interaction

Figure 1: Architecture of the question answering system used for the Synergy task.


the question and the sentence embeddings is modelled by applying element-wise multiplication.
The result of this multiplication is concatenated to the sentence embeddings and the sentence
positions. There is an intermediate hidden layer with dropout, followed by the final classification
layer. The loss function is binary cross-entropy, and the target labels (0 or 1) were generated
based on the ROUGE-F1 score of the candidate sentence with respect to the corresponding ideal
answer.
   The hyperparameters of the system are: Number of epochs=10; batch size=1024; dropout=0.7;
hidden layer size = 50; embeddings size=100; sentence length clipped to 300 tokens.
   The following sequence of steps was used to generate the candidate sentences that were fed
as input to the question answering system:

    1. Obtain the list of candidate documents, sorted by relevance. For this, we used the search
       API provided by the organisers of the BioASQ Synergy task. We also experimented with
       the use of sentence BERT fine-tuned with the BioASQ data as described in Section 2.1.
    2. Split the documents into sentences and select and rank the most relevant sentences. For
       this, we experimented with various methods described in Section 2.2. The resulting
       sentences were used as candidate sentences to be processed by the question answering
       system.

2.1. Document Retrieval
We experimented with three different approaches to document retrieval. These are listed below
and named for further reference in this paper.

DocAPI Most of our runs used the search API provided by the organisers of the BioASQ
Synergy task. In preliminary experiments, we observed that the default results returned by the
API were correlated with the cosine similarity with the question. We therefore concluded that
the API returned the results ranked by some sort of relevance. Consequently, we selected the
top 𝑛 documents, where 𝑛 depended on the round number (50 for round 1, and 100 for every
subsequent round).

DocNIR(untuned) We submitted one run based on the Neural Index Retrieval (NIR) method-
ology outlined in [7]. This document retrieval method combines a traditional inverted index
with a neural index of the document collection. Specifically, the document relevance scores for
each query are obtained by interpolating the normalised BM25 scores (so, the relevance score
based on the use of a traditional inverted index) with a cosine similarity score between the
neural representations of the query and the document. The neural representations are obtained
via a sBERT [8] model pre-trained on the target corpus and fine-tuned on a natural language
inference task.3

DocNIR(tuned) In round 4, we also experimented with document retrieval based on the use
of sBERT [8] fine-tuned with the BioASQ data. The retrieval model is, in essence, similar to
the NIR method outlined above. The main difference is that the sBERT model is additionally
fine-tuned on the target task training data (specifically, the relevance feedback available from
the previous rounds). Another notable difference is that we used the neural component only to
re-score (still using the BM25 for interpolation) the top-200 documents retrieved by the BM25
model for each query (making it, effectively, a re-ranker).
   The final runs used the top 10 documents, after removing those that were in previous feedback.

2.2. Snippet Retrieval
We experimented with several approaches to identify and rank the relevant snippets as described
below.

SnipCosine Our baseline snippet retrieval system was based on the tf.idf cosine similarity
between the question and the input document sentences. In particular, each document retrieved
by the document retrieval system (after removing false positives as indicated by the feedback
form previous rounds) was split into sentences (using NLTK’s sentence tokeniser). Then, for
each document, the top 3 sentences were extracted. To identify the top 3 sentences, we used
cosine similarity between the tf.idf vector of the question and the candidate sentence. For each
document, the top 3 sentences were ranked by order of occurrence in the document (not by
order of similarity). These sentences were then collated by order of document relevance.

SnipQA Our second approach used the BioASQ8b question answering system (Figure 1) to
rank the document input sentences. The rationale for this approach was that the BioASQ8b
question answering system had been trained to score sentences based on their likelihood of
being part of the ideal answer, and we wanted to know whether such a system could be used,
without fine-tuning, as a snippet re-ranker. As with the baseline system, the documents were
split into sentences. These sentences were then scored using the question-answering system,


   3
       We used the “manueltonneau/clinicalcovid-bert-nli” model from the huggingface transformers repository.
Table 1
Number of sentences selected, for each question type
                                     Summary    Factoid   Yesno   List
                                n         6        2        2      3


and the top 3 sentences per document were selected and ranked by order of occurrence. The
resulting sentences were then collated by order of document relevance.

SnipSBERT A third approach was based on the use of sBERT [8], trained for passage re-
trieval. Using the full CORD-194 dataset we have tried to retrieve the most relevant snippets by
minimising a cosine distance between a question and a sentence in the dataset. Two variants
were implemented: SnipSBERT(a) used the output of the Synergy API and returned the top 3
snippets, whereas SnipSBERT(b) searched the CORD-19 data directly and returned the top 3 or
top 5 snippets.
   The final runs used the top 10 snippets, after removing those that were in previous feedback.

2.3. Answer Generation
In all of our experiments, answer generation used the same process as illustrated in Figure 1
to conduct query-focused extractive summarisation. In particular, as in the original paper
[6], given a question, sentence, and sentence position, the system predicted the probability
that the sentence has high ROUGE-F1 score with the ideal answer. We obtained the relevant
sentences using the methods for snippet retrieval detailed in Section 2.2. Then, irrelevant
sentences (as indicated by feedback from previous rounds) were removed. The position of the
remaining sentences was indicated by their order after the snippet retrieval stage and after
removing irrelevant sentences. With this information, the question answering system returned
the sentence score. The answer was obtained by selecting the top 𝑛 sentences, and sorting them
by order of appearance in the list of snippets. The value of 𝑛 depended on the question type as
listed in Table 1.

2.4. Results of the Synergy Task
All runs submitted to the Synergy Task use the same approach to generate the ideal answers
(Section 2.3) and we experimented with combinations of the approaches to retrieve the doc-
uments (Section 2.1) and the snippets (Section 2.2). The specific set up of each run, and the
results, are detailed in Tables 2, 3, and 4.
  In document retrieval (Table 2), the NIR retrieval approaches outperformed the baseline that
used the API provided by the BioASQ organisers. Also, we observed best results when the
document retrieval system was tuned with the BioASQ data. Having said this, compared with
the other submissions to the Synergy tasks, the document retrieval systems performed poorly,
especially on rounds 2 to 4.

   4
       https://www.semanticscholar.org/cord19
Table 2
Document retrieval results of the submission to Synergy.                   Metric:   F1.   Legend:    DocAPI1 ;
DocNIR(untuned)2 ; DocNIR(tuned)3 .
                           Run         Round 1     Round 2      Round 3      Round 4
                           Best          0.3457      0.3237       0.2628       0.2375
                           Median        0.2474      0.2387       0.1810       0.1839
                           Worst         0.0802      0.0560       0.0179       0.0168
                           MQ-1         0.24741     0.16541      0.09731      0.10531
                           MQ-2         0.24741     0.16541      0.09731      0.10531
                           MQ-3                     0.16541      0.09731      0.10531
                           MQ-4                     0.16541                   0.15102
                           MQ-5                                               0.17623

Table 3
Snippet retrieval results of the submission to Synergy. Metric:F1. Legend: DocAPI→SnipCosine1 ;
DocAPI→SnipQA2 ; DocNIR(untuned)→SnipCosine3 ; DocNIR(tuned)→SnipCosine4 .
                           Run         Round 1     Round 2      Round 3      Round 4
                           Best          0.2712      0.1885       0.2026       0.1909
                           Median        0.2021      0.1634       0.1645       0.1461
                           Worst         0.0396      0.0204       0.0037       0.0078
                           MQ-1         0.14141     0.07041      0.04621      0.06401
                           MQ-2         0.13802     0.07062      0.04622      0.06572
                           MQ-3                     0.07092      0.04732      0.06342
                           MQ-4                     0.06952                   0.07983
                           MQ-5                                               0.09124


   In snippet retrieval, some runs used the output of DocAPI, others used the output of Doc-
NIR(untuned and tuned). We experimented with snippet re-ranking using SnipCosine and
SnipQA as detailed in Table 3. None of the runs used SnipSBERT because of problems meeting
the format requirements.5 We observed variability of results in the runs that used the sequence
DocAPI→SnipQA due to the undeterministic nature of the question answering module. Overall,
all results were very similar, and comparatively worse than the results of other runs. In fact,
our runs were near the bottom of the leaderboard in rounds 2 to 4.
   In ideal answer generation, the input to the question answering module used a sequence of
document retrieval followed by snippet retrieval as detailed in Table 4. We also included runs
that used SnipSBERT for snippet retrieval. Considering the poor results of the snippet retrieval
stage, the ideal answer results were relatively good and they were approximately around the
median of all submissions. This gives some indication that the question answering system,
trained on medical data but not on data containing COVID-19, was relatively robust. Even
though the absolute values of the ROUGE-S1 F1 scores were rather low, the average scores of
the human evaluation of most of our runs were above 3 in a scale from 0 to 4.
    5
     The Synergy task requires all snippets to include the character offsets. However, our implementation did not
provide this information.
Table 4
Ideal answer results of the submission to Synergy.    Metrics: ROUGE-SU F1 | Aver-
age human evaluation.      Legend:    DocAPI→SnipCosine→QA1 ; DocAPI→SnipQA→QA2 ;
DocNIR(untuned)→SnipCosine→QA3 ; DocNIR(tuned)→SnipCosine→QA4 ; SnipSBERT(a)→QA5 ;
SnipSBERT(b)→QA6 .
                Run        Round 1            Round 2           Round 3           Round 4
                Best                     0.0749 | 3.672    0.1170 | 4.185    0.1254 | 3.662
                Median                   0.0565 | 3.127    0.0883 | 3.517    0.0857 | 3.157
                Worst                    0.0096 | 0.667    0.0181 | 0.750    0.0221 | 0.705
                MQ-1                    0.05671 | 3.015   0.08831 | 3.517   0.09711 | 3.140
                MQ-2                    0.05652 | 2.965   0.09262 | 3.542   0.09122 | 3.157
                MQ-3                    0.04365 | 2.670   0.04676 | 3.062   0.05156 | 2.982
                MQ-4                    0.05006 | 3.047                     0.08573 | 3.190
                MQ-5                                                        0.07574 | 3.060

 sentence position

                      word embeddings         sentence embeddings

                                                                                  relu        sigmoid
  sentence




                                                                                                 ∫︀
                                           Mean



               BERT
  question




Figure 2: Architecture of the question answering system used for BioASQ 9b, Phase B.


3. BioASQ9b Phase B
The system that participated in BioASQ9b Phase B focused on the use of BERT-based architec-
tures for query-focused extractive summarisation. The experiments reported by [6] indicated
that replacing Word2Vec with BERT in the system of Figure 1 only gave a minor improvement of
the results. Subsequent (unpublished) experiments also appeared to indicate that using BERT as
an end-to-end system, without adding the multiplication layer between question and sentence,
plus the addition of the sentence position for the final classification layer, leads to similar or
better results. This motivated us to experiment with the use of BERT in the architecture shown
in Figure 2. The new architecture is a simplification to that of Figure 1, where most of the
computation, including determining the interaction between the question and the sentence, is
carried out by BERT. We experimented with several BERT variants as described in Section 3.1.
As with the system of Figure 1 and the system by [6], the system performs extractive summari-
sation and it is trained to predict whether the candidate sentence has a high ROUGE-SU4 F1
score with the ideal answer. In particular, the label of the training set was 1 if the sentence
was among the 5 sentences with highest ROUGE-SU4 F1 score, and 0 otherwise. The final ideal
answer is obtained by selecting the top 𝑛 sentences, and these sentences are presented in order
of appearance in the input snippets. The value of 𝑛 is as shown in Table 1.
   The question and sentence were fed to BERT in the standard approach defined by the creators
of BERT [4]. In particular, the question and sentence were input as two separate text segments
in the following order: first the “[CLS]” special token, then the question, then the sentence
separator “[SEP]”, and finally the candidate sentence.
   Instead of passing the embedding of the “[CLS]” special token to the classification layer, we
decided to use the embeddings of the tokens forming the candidate sentence. These embeddings
were mean pooled in order to obtain the sentence embeddings.

3.1. BERT Variants
We experimented with the following BERT variants. All of these variants were based on models
made available by the Huggingface transformers repository6 .

BERT        We used huggingface’s model “bert-base-uncased”.

BioBERT Given the medical nature of BioASQ, we tried BioBERT, which uses the same archi-
tecture as BERT base, and has been fine-tuned with PubMed articles [9]. We used huggingface’s
model “monologg/biobert_v1.1_pubmed”.

DistilBERT DistilBERT’s architecture is a reduced version of BERT, which has been trained
to replicate the soft predictions made by BERT [10]. The resulting system is faster to train, and
reportedly nearly as accurate as BERT. We used huggingface’s model “distilbert-base-uncased”.

ALBERT ALBERT uses parameter reduction techniques that allow faster training and with
lower memory consumption. This enables the use of larger numbers of transformer layers and
larger embedding sizes [11]. We used huggingface’s model “albert-xxlarge-v2”.

ALBERT-SQuAD This variant of ALBERT has been fine-tuned with data from SQuAD, a
well-known data set for question answering systems in the context of reading comprehension
[12]. We used huggingface’s model “mfeb/albert-xxlarge-v2-squad2”.

ALBERT-QA This final variant of ALBERT was obtained using ALBERT-SQUAD as a starting
point (using huggingface’s model “mfeb/albert-xxlarge-v2-squad2”). Then, the model was
fine-tuned by adding a SQuAD-style question answering classification layer and trained on
the BioASQ training set, using the exact answers as labels. For this fine-tuning stage, only
factoid questions were used. The system that implemented this fine-tuning is one of the systems
described by [13].
   6
       https://huggingface.co/transformers/
Table 5
Results of 10-fold cross-validation using the BioASQ9b training data. Metric: ROUGE SU4 F1.
                                     Number of Parameters
               System                     Full    Trained        Epochs     Dropout      SU4-F1
               BERT                  109,520,791       38,551           8          0.8    0.2779
               BioBERT               108,348,823       38,551           1          0.7    0.2798
               DistilBERT             66,401,431       38,551           1          0.6    0.2761
               ALBERT                222,800,535      204,951           5          0.5    0.2866
               ALBERT-SQuAD          222,800,535      204,951           5          0.7    0.2846
               ALBERT-QA             222,800,535      204,951           5          0.4   0.2875


   In all of our experiments, we froze all BERT layers and only trained the hidden and classifica-
tion layers. The reason for this decision was that, in preliminary experiments with unfrozen
BERT layers, we observed the catastrophic forgetting effect where all the pre-trained informa-
tion was lost, and decided to leave the study of fine-tuning strategies of the BERT layers for
further work.
   Table 5 shows the results of 10-fold cross-validation on the BioASQ9b training data. The table
also shows the values of the differing hyperparameters of the best systems as found through grid
search. The hyperparameters common to all systems were: batch size=32; hidden layer size=50;
sentence length clipped to 250 tokens. Overall, all results are similar, but we can observe that
BioBERT outperforms BERT, in line with most prior work (but in contrast with [6]). We can
also observe an improvement of the results of the three ALBERT variants. This is possibly due
to the larger architecture sizes. The fact that ALBERT-QA has a slightly better result than the
other ALBERT variants is encouraging.

3.2. Submission Results to BioASQ 9b Phase B
The runs submitted to BioASQ9b Phase B used all the BERT variants described in Section 3.1
except ALBERT-SQuAD.
   The preliminary evaluation results, as reported in the BioASQ website, are shown in Table 6.7
For each batch, our runs ranked among the top participating systems. In fact, ALBERT-QA
was the top run of batch 3. This demonstrates that a straightforward use of BERT is a very
strong baseline. As expected, BioBERT outperformed BERT. The experiments with ALBERT and
ALBERT-QA in batches 4 and 5, however, were not as good as expected given our cross-validation
results.


4. Summary and Conclusions
We have presented Macquarie University’s contribution to the BioASQ Synergy task and
BioASQ9b Phase B (Ideal Answers).


    7
      Note that the results reported in the BioASQ website (http://bioasq.org) may change in the future after the
test data is enriched with further annotations.
Table 6
Preliminary results of the submissions to BioASQ9b, Phase B.
                                                           ROUGE-SU4
            Run        System         Batch 1   Batch 2      Batch 3 Batch 4      Batch 5
            Best                       0.3410    0.3974        0.3266   0.4402     0.3893
            Median                     0.2536    0.1990        0.2647   0.3388     0.2666
            Worst                      0.1154    0.1186        0.1017   0.0886     0.1331
            MQ-1       BERT            0.3032     0.3560      0.3057     0.3585    0.3511
            MQ-2       BioBERT         0.3103     0.3615      0.3265     0.3612   0.3733
            MQ-3       DistilBERT      0.3007    0.3753       0.3204    0.3681     0.3711
            MQ-4       ALBERT         0.3205      0.3676      0.3100     0.3560    0.3570
            MQ-5       ALBERT-QA                  0.3610     0.3266      0.3559    0.3589


   For the synergy task, we have experimented with a question answering module that was
designed for, and trained with, the data from BioASQ8b. Due to the need to produce an end-to-
end system, we tried various baseline document and snippet retrieval systems. Overall, despite
the poor general quality of the document and snippet retrieval systems, the results of our
submissions indicate that the question answering component can generalise well to questions
related to COVID-19. Further work will focus on improving the quality of the document and
snippet retrieval components.
   The synergy task was organised in multiple rounds such that feedback from previous rounds
was available for subsequent rounds in some questions. Our system incorporated this feedback
only in a trivial manner, simply by removing documents or snippets that were identified as
known negatives. There has been research on relevance feedback since at least 1971 [14], which
could be incorporated into the system. More recent approaches, such as using a twin neural
network with a contrastive loss [15], may work here.
   The contribution to BioASQ9b Phase B focused on the use of BERT variants within a query-
focused extractive summarisation setting. The architecture concatenates the question and
candidate sentence as two separate text segments, very much as is done in question-answering
approaches with BERT, and the system is trained as a sentence classification system. We observe
that such a simple architecture is a very strong baseline. Further work will focus on exploring
further variants of BERT, and on enhancing the pre-training and fine-tuning stages.


Acknowledgments
This research was undertaken with the assistance of resources and services from the National
Computational Infrastructure (NCI), which is supported by the Australian Government.
  Research by Vincent Nguyen is supported by the Australian Research Training Program and
the CSIRO Postgraduate Scholarship.


References
 [1] Y. Xu, M. Lapata, Text summarization with latent queries, 2021. arXiv:2106.00104.
 [2] G. Tsatsaronis, G. Balikas, P. Malakasiotis, I. Partalas, M. Zschunke, M. R. Alvers, D. Weis-
     senborn, A. Krithara, S. Petridis, D. Polychronopoulos, Y. Almirantis, J. Pavlopoulos,
     N. Baskiotis, P. Gallinari, T. Artieres, A. Ngonga, N. Heino, E. Gaussier, L. Barrio-Alvers,
     M. Schroeder, I. Androutsopoulos, G. Paliouras, An overview of the bioasq large-scale
     biomedical semantic indexing and question answering competition, BMC Bioinformatics
     16 (2015) 138. URL: http://www.biomedcentral.com/content/pdf/s12859-015-0564-6.pdf.
     doi:10.1186/s12859-015-0564-6.
 [3] L. L. Wang, K. Lo, Y. Chandrasekhar, R. Reas, J. Yang, D. Burdick, D. Eide, K. Funk, Y. Katsis,
     R. M. Kinney, Y. Li, Z. Liu, W. Merrill, P. Mooney, D. A. Murdick, D. Rishi, J. Sheehan,
     Z. Shen, B. Stilson, A. D. Wade, K. Wang, N. X. R. Wang, C. Wilhelm, B. Xie, D. M. Raymond,
     D. S. Weld, O. Etzioni, S. Kohlmeier, CORD-19: The COVID-19 open research dataset,
     in: Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020, Association for
     Computational Linguistics, Online, 2020. URL: https://www.aclweb.org/anthology/2020.
     nlpcovid19-acl.1.
 [4] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, BERT: Pre-training of deep bidirectional
     transformers for language understanding, in: Proceedings of the 2019 Conference of
     the North American Chapter of the Association for Computational Linguistics: Human
     Language Technologies, Volume 1 (Long and Short Papers), Association for Computational
     Linguistics, Minneapolis, Minnesota, 2019, pp. 4171–4186. URL: https://www.aclweb.org/
     anthology/N19-1423. doi:10.18653/v1/N19-1423.
 [5] J.-C. Han, R. T.-H. Tsai, NCU-IISR: Using a pre-trained language model and logistic
     regression model for bioasq task8b phase b, in: L. Cappellato, C. Eickhoff, N. Ferro,
     A. Névéol (Eds.), Working Notes of CLEF 2020 — Conference and Labs of the Evaluation
     Forum, Thessaloniki, 2020. URL: http://ceur-ws.org/Vol-2696/paper_72.pdf.
 [6] D. Mollá, C. Jones, V. Nguyen, Query focused multi-document summarisation of biomedical
     texts, in: L. Cappellato, C. Eickhoff, N. Ferro, A. Névéol (Eds.), Working Notes of CLEF
     2020 — Conference and Labs of the Evaluation Forum, Thessaloniki, 2020. URL: http:
     //ceur-ws.org/Vol-2696/paper_119.pdf.
 [7] V. Nguyen, M. Rybinski, S. Karimi, Z. Xing, Pandemic literature search: Finding information
     on COVID-19, in: Proceedings of the 18th Annual Workshop of the Australasian Language
     Technology Association, Australasian Language Technology Association, Virtual, 2020, pp.
     92–97. URL: https://www.aclweb.org/anthology/2020.alta-1.11.
 [8] N. Reimers, I. Gurevych, Sentence-BERT: Sentence embeddings using siamese BERT-
     networks, in: Proceedings of the 2019 Conference on Empirical Methods in Natural
     Language Processing, Hong Kong, 2019, p. 3982.
 [9] J. Lee, W. Yoon, S. Kim, D. Kim, S. Kim, C. H. So, J. Kang, BioBERT: a pre-trained
     biomedical language representation model for biomedical text mining, Bioinformatics
     36 (2019) 1234–1240. URL: https://doi.org/10.1093/bioinformatics/btz682. doi:10.1093/
     bioinformatics/btz682.
[10] V. Sanh, L. Debut, J. Chaumond, T. Wolf, DistilBERT, a distilled version of BERT: smaller,
     faster, cheaper and lighter, in: 33rd Conference on Neural Information Processing Systems
     (NeurIPS 2019), 2019. arXiv:1910.01108.
[11] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, R. Soricut, ALBERT: A lite BERT
     for self-supervised learning of language representations, in: Proceedings of the 8th
     International Conference on Learning Representations, Virtual, 2020. URL: https://iclr.cc/
     virtual_2020/poster_H1eA7AEtvS.html.
[12] P. Rajpurkar, R. Jia, P. Liang, Know what you don’t know: Unanswerable questions for
     SQuAD, in: Proceedings of the 56th Annual Meeting of the Association for Computational
     Linguistics (Volume 2: Short Papers), Association for Computational Linguistics, Mel-
     bourne, Australia, 2018, pp. 784–789. URL: https://www.aclweb.org/anthology/P18-2124.
     doi:10.18653/v1/P18-2124.
[13] U. Khanna, D. Mollá, Transformer-based language models for factoid question answering
     at BioASQ9b, in: G. Faggioli, N. Ferro, A. Joly, M. Maistro, F. Piroi (Eds.), Working Notes
     of CLEF 2021 — Conference and Labs of the Evaluation Forum, Bucharest, Romania, 2021.
[14] J. J. Rocchio, Relevance feedback in information retrieval, in: G. Salton (Ed.), The SMART
     Retrieval System — Experiments in Automatic Document Processing, Prentice Hall, 1971,
     pp. 313–323.
[15] R. Hadsell, S. Chopra, Y. LeCun, Dimensionality reduction by learning an invariant
     mapping, in: 2006 IEEE Computer Society Conference on Computer Vision and Pattern
     Recognition (CVPR’06), volume 2, 2006, pp. 1735–1742. doi:10.1109/CVPR.2006.100.