<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Assembling four Open Web Search Components</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Linda Erben</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Hampel</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Malte-Christian Kuns</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vincent Melisch</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Per Natzschka</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wilhelm Pertsch</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lina Razouk</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Reiner Stolle</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Robert Thomas Thoss</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tuan Giang Trinh</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julius Gonsior</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anja Reusch</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Technische Universität Dresden, Dresden Database Research Group</institution>
          ,
          <addr-line>Dresden</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Technische Universität Dresden</institution>
          ,
          <addr-line>Dresden</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>In this work, we present the submission of TU Dresden to WOWS 2024. Four student teams assembled different approaches for Genre Classification, Text Snippet Extraction, Query Expansion, and Text Features. Each implemented component integrates seamlessly into the open web search ecosystem. We present each approach alongside a short evaluation of possible use cases, and hope that our submission will contain viable building blocks for future research to be build on top.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Information Retrieval</kwd>
        <kwd>Open Web Search</kwd>
        <kwd>Genre Classification</kwd>
        <kwd>Text Snippet Extraction</kwd>
        <kwd>Query Expansion</kwd>
        <kwd>Text Features</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Genre Classificaion</title>
      <p>
        The goal of genre classification [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is to categorize documents into the intent of the document
itself. Objectives for a website could include: (1) making sales (like in an online store), (2)
providing information (like in a course website for a university), (3) sharing personal experiences
(like in a personal blog), etc. Genre classification facilitates effective document filtering in
ranking based on the search query in conjunction with existing query intent classifiers that
differentiate search queries as informational (e.g. "What is IR?"), transactional (e.g., "I want
to buy a PlayStation"), or navigational (e.g., "take me to log in for my university course") [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Relevant websites with a matching genre should be ranked higher if a query indicates an intent.
      </p>
      <sec id="sec-2-1">
        <title>This open web search component 1 examines techniques for classifying text documents into</title>
        <p>their respective categories, employing rule-based and machine-learning methodologies. We
compare three classification strategies with a focus on high precision.</p>
        <sec id="sec-2-1-1">
          <title>2.1. Methods</title>
          <p>
            2.1.1. Rule-Based Classifier
The rule-based classifier makes use of a vocabulary list of relevant terms per genre. Comparing
the intersection between terms in the genre-specific vocabulary lists, and the terms in the
document, the most probable category is the one with the highest intersection. We first remove
stop words and subsequently extract the 75 most frequent terms that we compare to the
vocabulary lists to classify the genre. We use Snorkel AI [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] for implementation.
          </p>
          <p>
            The rule-based classifier can be adapted to a precision-oriented method, where the most
probable genre needs to be better than a threshold compared to the second most probable genre,
otherwise the classification result is abstain.
2.1.2. Multi-Layer Perceptron Classifier
As a typical Machine Learning based method a neural network was used for classification. As
features the web pages were converted into a tf-idf vectorspace. We use the Python library
scikitlearn [
            <xref ref-type="bibr" rid="ref5">5</xref>
            ] for the implementation of the Multi-Layer Perceptron classifier. After an empirical
hyperparameter search a neural network using a single hidden layer of 50 neurons, ReLU
activation function, stochastic gradient descent in the Adam variant using momentum for
optimizations, and a constant learning rate of 0.001 was used.
          </p>
        </sec>
        <sec id="sec-2-1-2">
          <title>2.2. Experiments</title>
          <p>
            2.2.1. Dataset
For evaluation we used the Genre-KI-04 dataset [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ]. This includes vocabulary lists, and the
following classification categories: articles, discussion, download, help, link lists, portrait (non
private), portrait (private), and shop. Details about the genres can be found in the original paper.
1https://github.com/tira-io/workshop-on-open-web-search-tu-dresden-01
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Text Snippet Extraction</title>
      <p>
        Since sophisticated neural ranking models such as cross-encoders generally require a lot of
computational effort, a customary retrieval pipeline first retrieves a number of (e.g., 1000)
documents using a fast but imprecise retrieval method and then re-ranks those documents using
a more precise weighting model [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Cross-encoder as introduced by Nogueira and Cho [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] are
an example for the latter, which are used to calculate scores for query-document pairs. Apart
from their comparatively high computational cost, cross-encoders have another disadvantage–
their limited input size. This weakness is typically mitigated by truncating the document once
the maximum number of input tokens is reached. The problem of this procedure is that content
which is not in the beginning of a document is not taken into account by cross-encoders. As a
result, the ranking of documents may be biased towards those that address the query early on.
      </p>
      <p>In this part, we therefore present a simple method of extracting a number of snippets, i.e.,
smaller chunks of the document which fit in the cross-encoder as an additional component
in a larger retrieval pipeline. Instead of simply truncation documents after a fixed number of
tokens, we search for the most relevant passages (ranked snippets) in the document. These
ranked snippets are used for the cross-encoder with the goal of a more precise ranking. We
show the benefits of this method on two exemplary datasets which contain long documents.</p>
      <sec id="sec-3-1">
        <title>3.1. Methodology</title>
        <p>The re-ranking process with ranked snippets consists of five steps. An example of those steps
for the re-ranking of n = 3 documents (d1, d2, and d3) is shown alongside the explanation.</p>
        <p>
          First, we subdivide all n documents into snippets. The maximum length of those snippets
may be chosen arbitrarily–we defaulted to 250 tokens which is the passage size used by Dalton
et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. The actual length of the snippets may vary since the division process aims to retain
context by not separating sentences. For example, we may start with three documents d1, d2,
d3. After the first step, each of these documents is divided into several snippets: s1s12 . . . sl11 ,
1
s21s22 . . . sl22 , s31s32 . . . sl33 where sij denotes the j-th snippet of document di for j ∈ { 1, . . . , li }
and li is the number of snippets of di.
        </p>
        <p>In Step 2, we pre-rank all extracted snippets in relation to the query. To accomplish this,
we view the set of all snippets of a document as a corpus. From this corpus, we can create a
ranking for the query using one of the following weighting models: Term frequency (TF), BM25
or PL2. We do not use cross-encoder for the pre-ranking of documents, because there may be a
multitude of snippets per document depending on document length and therefore ranking all
snippets using a cross-encoder can drastically slow down the re-ranking process. After this
pre-ranking step, our example snippets might be ranked in the following way: s33 &gt; s2 &gt; s2 &gt; s3
2 4 2
&gt; s31 &gt; s1 &gt; s3 &gt; s1 &gt; s2 &gt; s1 &gt; . . . .</p>
        <p>3 4 1 3 5</p>
        <p>In Step 3, we can obtain the top k relevant snippets of each document from the pre-ranking,
which are later ranked using a cross-encoder. This step ensures that the cross-encoder only needs
to rank n · k snippets for n documents instead of all snippets. In order to reduce computational
cost, we defaulted to k = 3. In our example, this step results in the following selection:
! s13, s11, s15 ", ! s2, s24, s23 ", ! s3, s32, s31 ". Here, s34 is not selected as one of the top snippets of
2 3
d3 since it is the (k + 1)-th snippet of d3 in the ranking despite being ranked relatively high.</p>
        <p>In Step 4, the top k snippets of all documents are ranked using a cross-encoder (CE). That
way, similar to Step 2, we can more accurately deduce which snippets best match the query–but
now the ranking is more precise since we used a CE instead of the simple weighting models
used in Step 2. An examplary ranking for our snippets might be: s24 &gt; s33 &gt; s22 &gt; s15 &gt; . . . . The
final document ranking ensues from this snippet ranking in Step 5, i.e., the document that
provided the best snippet is ranked first. Our example documents are therefore ranked in the
following way: d2 &gt; d3 &gt; d1. It should be noted that the goal of this section is to rank documents
with regard to a query, and not only passages. Therefore, the result is a ranking of documents.
Details on our implementation can be found in Appendix B.1.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Evaluation</title>
        <p>In this section, we conduct tests to study the possible improvements of our cross-encoder
re-ranking of top k snippets. As baselines, we use BM25 and the dense retriever MonoT5.
All further ranking is performed on the top 20 documents retrieved by these two systems.
We evaluate the re-ranking with our TF-ranked snippets. For this, we load the previously
saved top 3 snippets for each document. To re-rank the documents, we follow the “weakest
link” principle, selecting the minimum TF score among the top 3 snippets. This results in the
methods BM25+TF-SP and MonoT5+TF-SP. We denote by +CE that the 3 snippets are further
re-ranked by a cross-encoder. In addition, we compare the performance of these systems to the
cross-encoder’s performance when only evaluating the first snippet of each document (which
resembles the naïve application of a cross-encoder). These results are denoted by BM25+CE
and MonoT5+CE.</p>
        <p>
          To measure the performance of the approaches, we utilize normalized discounted
cumulative gain at 10 (NDCG@10) and mean reciprocal rank (MRR). We conduct our tests on the
ClueWeb12 [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and ClueWeb09 [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] datasets, which differ in document size: ClueWeb12 has
an average document size of 5641.7 tokens, and ClueWeb09 has an average document size of
1132.6 tokens.
        </p>
        <p>BM25</p>
        <p>MonoT5</p>
        <p>BM25+TF-SP</p>
        <p>MonoT5+TF-SP</p>
        <p>BM25+TF-SP+CE
MonoT5+TF-SP+CE</p>
        <p>BM25+CE
MonoT5+CE</p>
        <p>BM25</p>
        <p>MonoT5</p>
        <p>BM25+TF-SP</p>
        <p>MonoT5+TF-SP</p>
        <p>BM25+TF-SP+CE
MonoT5+TF-SP+CE</p>
        <p>BM25+CE</p>
        <p>MonoT5+CE
0
0.2 0.4
Performance
0.6
0
0.2 0.4
Performance
0.6
(a) ClueWeb09 (2011)
(b) ClueWeb12 (2013)</p>
        <p>The results for the two datasets are plotted in Fig. 2. Our approach of cross-encoder re-ranking
with TF-pre-ranked snippets achieves the best performance in both metrics across all our tested
datasets (see Appendix B.2 for diagrams of other evaluated datasets). The impact of our
TFranked snippet pre-selection is relatively high on ClueWeb12 with long documents, while it
is more marginal on ClueWeb09. This highlights the importance of snippet pre-selection for
longer documents. ClueWeb09 consists of approximately 6 snippets, and ClueWeb12 consists of
approximately 23 snippets per document. We assume for our naïve snippet generation approach
that information is equally spread throughout a document. A cross-encoder taking the first
snippet as input is assumed to capture more relevant information of a document with a size
that is closer to the cross-encoder input size. This also explains why MonoT5 scores better
on the shorter dataset, especially in comparison to BM25, since MonoT5 also suffers from a
limited input size. This proves that there is a need to address the problem of limited input
size, especially in large documents like those in ClueWeb12. That information is not always
equally spread over a document, like we assumed for our snippet generation, can be concluded
when comparing Figs. 4b and 4c. This raises the need of a more advanced approach for snippet
generation.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Summary</title>
        <p>Overall, our results show that selecting top-k pre-ranked snippets is a viable approach to tackle
the problem of input size restrictions on Transformer-based retrieval systems. Especially,
crossencoders can benefit from this approach since they are inefficient on large documents. Further
testing to edge out efficiency and reduce context loss with snippets will be required. Also,
it would be beneficial to test multiple pre-ranking systems and values of k for top-k snippet
selection. The code for this part can be found in the accompanying repository2.
4. Query Expansion and User Query Variants using Large</p>
        <p>
          Language Models
Query Expansion and User Query Variants are two common methods to increase the recall of an
IR system [
          <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
          ]. Both methods are based on modifying the query to include more related
keywords, thereby causing the IR system to score relevant documents higher. In addition to
conventional techniques such as the Kullback-Leibler Divergence (KL) [
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ] or Relevance
Model 3 (RM3) [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], recent approaches have embraced the utilization of Large Language Models
(LLMs). In this part, we employ various prompts to generate improved and expanded queries
using LLMs [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ], in particular, GPT-3.53, Llama2 [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] and FLAN-UL2 [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>4.1. Methodology</title>
        <p>
          LLMs have previously been in use for the task of query expansion and studies have been
conducted using various methods and language models [
          <xref ref-type="bibr" rid="ref16 ref20 ref21 ref22">16, 20, 21, 22</xref>
          ]. Wang et al. [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] employ
query2doc, a method where the LLM generates a document for a given query, which is then
used as Pseudo-Relevance Feedback (PRF). Jagerman et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] follow a similar approach
but extend the experiments to include alternative LLMs and additional prompt types. All
2https://github.com/tira-io/workshop-on-open-web-search-tu-dresden-02
3https://platform.openai.com/docs/models/gpt-3-5-turbo
        </p>
        <p>Model
Llama 2
FLAN-UL2
GPT-3.5 Turbo</p>
        <p>Temperature
min. Tokens
max. Tokens</p>
        <p>Quantization</p>
        <p>
          Parameters
previous work demonstrates improvements across different datasets. In order to weigh the
original query more heavily, multiple concatenations of the original query q with a single
instance of the LLM’s output may be used [
          <xref ref-type="bibr" rid="ref16 ref21">16, 21</xref>
          ]. The resulting expanded query is of the form
q′ = concat({q} ∗ n, LLMout), where n is the number of times q is concatenated with itself,
and LLMout is the LLM-generated version of q. We adopt this approach in our work with n = 5
and employ modified versions of the prompt types suggested by Jagerman et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]: Chain
of Thoughts (CoT) where the model is prompted to document its thought process, Query to
Expansion with Zero-Shot prompting (Q2E/ZS) where the model should reformulate the query
directly, and Query to Expansion with Few-Shot prompting (Q2E/FS), where three examples
for the desired query format are provided to the model. For the exact prompt format used,
see Appendix C.3. It should be noted that the prompt for Q2E/ZS differs between the models.
While GPT-3.5 and FLAN were prompted to generate five similar queries, Llama was asked to
answer the query. Apart from this difference, the prompts for in all experiments are similar and
comparable.
        </p>
        <p>Initially, the query, along with the prompt, is fed into the LLM, and its response is concatenated
with the original query (n = 5). For evaluation, the Recall@1000 metric of the original and
modified queries is compared on the given dataset using BM25. The specific LLMs in use are</p>
        <sec id="sec-3-4-1">
          <title>GPT-3.545, Llama 2 [18] and FLAN-UL2 [19]. Llama 2 and FLAN-UL2 were run locally. Table 1</title>
          <p>
            shows the model configurations that we used in our experiments. The temperature values were
chosen empirically in a way such that model outputs are roughly similar. The lower and upper
token limitations prevent generation edge cases such as empty responses or endless output,
while still allowing for expressive responses. Local models had to fit GPU memory constraints.
Hence, we had to employ the quantized versions of the models. We conducted experiments for
the prompt types presented above: CoT, Q2E/FS, and Q2E/ZS. While FLAN-UL2 and GPT-3.5 can
be prompted without further changes, Llama 2 requires the chat-prompt to follow a pre-defined
format, our version of which can be found in the project’s repository6. We utilize BM25 as the
retrieval system in the default configuration of the Tira-framework [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ]. The query expansion
baselines consist of an unmodified BM25, BM25 with Kullback-Leibler Divergence (KL) and
BM25 with RM3.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>4.2. Evaluation</title>
        <p>
          We measure the recall, which is aggregated over 18 datasets per model, and per prompt type. The
datasets cover a range of diverse topics and were provided as part of TIRA [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] / TIREx [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. The
aggregated results can be observed in Table 4.2. Avgeasy excludes evidently (cord19, longeval)
and presumably (medline) difficult datasets. This highlights the difficulties LLMs experience
on specific datasets, especially domain-specific ones: Excluding those, CoT+OpenAI GPT-3.5
Turbo (GPT) now performs 0.03 points better than baseline models. Note that the two Avg
rows cannot be compared to one another, as baseline scores have also shifted due to the
exclusion of generally low-performing datasets. Detailed results for each dataset can be found
in Appendix C.1. For our query expansion approaches, it is evident that the choice of prompt
has a large impact on recall performance. The combination of CoT and GPT consistently yields
the highest recall in absolute numbers. However, with other prompt types such as Q2E/ZS
and Q2E/FS, GPT also frequently achieves the highest recall per dataset, albeit less frequently
compared to CoT. In this regard, our results are consistent with those reported in [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Although
CoT generally performs the best, it exhibits poorer results than the baselines in datasets such as
cord19 or the longeval datasets. In these cases, Q2E/ZS and Q2E/FS emerge as better choices,
but are still commonly outperformed by the baseline models.
        </p>
        <p>Q2E/FS exhibits less convincing effectiveness, presumably because it mimics the relatively
short responses of example queries through the Q2E/FS method, resulting in short queries with
few new keywords. Q2E/ZS behaves similarly. Although the responses of the LLMs are longer
compared to Q2E/FS, as the LLMs do not conform to the rather short examples, the generated
responses are overall less extensive than those of CoT, likely resulting in inferior effectiveness.
Considering the longeval datasets and cord19, it is evident that they contain either very general
or highly specific queries. In the case of nonspecific queries, there is a risk that they may be
muddled by the consequently more general, and in the case of CoT, extensive responses from
the LLMs. This effect might potentially be reversed by conveying the user intent to the LLM,
indicating whether, for instance, in the case of the query "car," the user intends to buy one or
have it repaired. With domain-specific queries, it is plausible that models were trained with
insufficient knowledge on the subject, resulting in subpar effectiveness.</p>
        <p>
          While our main evaluation is conducted using recall@1000, we also evaluated nDCG@10.
4https://platform.openai.com/docs/models/gpt-3-5-turbo
5https://platform.openai.com/docs/api-reference/
6https://github.com/tira-io/workshop-on-open-web-search-tu-dresden-03
Flesch Reading Ease [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]
Flesch Kincaid grade level [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]
Gunning FOG [
          <xref ref-type="bibr" rid="ref29 ref30">29, 30</xref>
          ]
SMOG Index [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ]
Automated Readability Index [
          <xref ref-type="bibr" rid="ref29 ref32">29, 32</xref>
          ]
Coleman-Liau Index [
          <xref ref-type="bibr" rid="ref33">33</xref>
          ]
LIX [
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]
The results for this metric are detailed in Table 4 in the appendix. Overall, our conclusions
for nDCG are similar to those for recall. The generations for each model and each prompt are
publicly available in our repository7.
        </p>
      </sec>
      <sec id="sec-3-6">
        <title>4.3. Summary</title>
        <p>In this part, we generated different versions of query expansions using three LLMs and three
prompt templates. We were able to demonstrate that LLMs are capable of improving the recall of
user queries. The combination of the prompt CoT alongside GPT proves to be the most promising,
improving recall scores by up to 15%. Future research could focus on further templates for using
the generated expansions since we only evaluated the qqqqq, response-format.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Text Features</title>
      <p>Text Features are quantified metrics describing syntactic or semantic features of natural language.
An example is the readability of a text, useful for returning user-dependent search results. A
search engine targeted to school children should return results with a high readability score,
whereas a search engine with domain experts as target audience will also include texts with
low readability scores. Additionally, this could be used to filter out noisy websites.</p>
      <p>
        This Open Web Search component8 incorporates two tools for computing text features,
namely Textstat [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] and textdescriptives [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] from spaCy [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. SpaCy’s Text Feature analysis
is more comprehensive than the one in Textstat, but is less efficient. Per design of the pipeline
approach of SpaCy many things are computed in the background, from which only a few are
required for the calculation of the text features. This overhead results in a longer runtime which
should be considered. Table 3 displays the implemented text features.
7https://github.com/tira-io/workshop-on-open-web-search-tu-dresden-03/tree/main/src/generated
8https://github.com/tira-io/workshop-on-open-web-search-tu-dresden-04
      </p>
      <p>Additional contributions besides the integration of the text features components include
examining a potential correlation between text features and documents evaluated as relevant
by ranked retrieval models. For easier exploration of the document corpus we provide an
interactive Jupyter Notebook showing correlation graphs between Ranked Retrieval and Text
Features, applicable to arbitrary datasets, as well as the analysis of correlations between Ranked
Retrieval and Text Measures.</p>
      <sec id="sec-4-1">
        <title>5.1. Evaluation</title>
        <p>To verify the capability to differentiate between levels of Readability, unit tests were used.
The test data consists of multi-sentence snippets from web pages. These were categorized by
difficulty in the following categories (including the amount of test documents): children (3),
teenagers (3), academic (3), and simple language (2), depending on what demographic the source
was directed to. Initial tests involved a project member assessing the reading level of excerpts
and comparing their assessments to the classifications provided by the automated measures,
thus proving correct usage of the used text feature libraries at least for the readability scores.</p>
        <p>Compared to human assessments, the automated Text Measures often overestimated the
reading level, possibly failing to capture the complexities of human reading abilities within
their respective indexes. Large-scale dataset computations further highlighted the discrepancies
between predicted and human-classified reading levels, corroborating these findings. Despite
the observed differences in assessment, the data suggested an inverse proportional relationship
between comprehension levels and readability measures.</p>
      </sec>
      <sec id="sec-4-2">
        <title>5.2. Experiment Design</title>
        <p>
          The experiments were run on the "antique/test" [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ] dataset from the ir_datasets collection [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ].
Based on TIRA [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ] ranked retrieval models were used to create top-10 results.
5.2.1. Correlations between ranked retrieval and text feature readability
A primary objective of our project was to investigate whether the ranking of relevant documents
by ranked retrieval models correlates to document Readability.
5.2.2. Readability of Top 10
First we looked at the top 10 retrieved documents for all queries across multiple retrieval models,
the resulting distributions are displayed in Figure 3. The majority of results, assessed using
the Flesch Reading Ease, indicates comprehension levels at or below an eighth-grade level,
implying a high degree of readability. The high degree of readability was consistently observed
across multiple retrieval models. Compared to the overall readability across all documents in
a collection, we found that some retrieval models like SBERT or MonoT5 indeed result in a
higher readability in the retrieved documents compared to the rest of the corpora, suggesting a
potential relationship between relevancy and readability, whereas other retrieval models such
as BM25 do not share this characteristic.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>7. Acknowledgments</title>
      <p>We would like to express our gratitude to the Open Search Foundation for organizing the WOWS
2024 and especially Maik Fröbe, who supported us and our student teams in organizing and
conducting our Hackathon which made this submission possible.</p>
      <p>In addition, the authors gratefully acknowledge the computing time made available to them
on the high-performance computer at the NHR Center of TU Dresden. This center is jointly
supported by the Federal Ministry of Education and Research and the state governments
participating in the NHR (www.nhr-verein.de/unsere-partner).</p>
    </sec>
    <sec id="sec-6">
      <title>A. Hackathon</title>
      <p>The paper’s work was carried out by students from TU Dresden as part of a one-week hackathon.
The workshop was open to students in the Computer Science program and related fields, and
they could earn ECTS credit points for lab work. The hackathon was advertised on the mailing
lists of the beginner Information Retrieval courses from the past three years. Interested students
could fill a survey indicating their preferred timeframe for the hackathon.</p>
      <p>After a date was decided, 10 students signed up for the hackathon, three from the Bachelor’s
program and seven from the Master’s program. The university supervisors prepared four topics,
which were advertised beforehand, and the students signed up for their preferred topics. The
text features topic was designated to the 3 Bachelor students. The Master students were provided
with a peer-reviewed research paper as additional material, which they were required to read
and understand before the hackathon.</p>
      <p>On the first day of the hackathon, an invited member of the Open Web Search project provided
a brief introduction to the Open Web Search ecosystem and TIRA/TIREx. Following this, the
teams worked on their components, with supervisors providing guidance through daily
checkins. On the fifth and final day of the hackathon, a short presentation from each team was held.
Following the hackathon, the students were requested to prepare a report on their work, which
served as the basis for this paper.</p>
      <p>In retrospective, the short amount of time, one week, motivated the students to work diligently
on their project. However, at the end of the week, the students had several open ideas for future
work which they could not finish in time. Therefore, more time, even a few days more, might
be beneficial for the next iteration of the hackathon. The size of the group ranged between two
and three members. The small group size facilitated the organization within each group and
kept the management overhead small. The topics of the hackathon were aligned with the basics
gained during the Information Retrieval course, but required also reading additional literature
and research.</p>
    </sec>
    <sec id="sec-7">
      <title>B. Text Snippet Extraction</title>
      <sec id="sec-7-1">
        <title>B.1. Implementation</title>
        <p>
          To implement the described re-ranking steps, we utilized several Python libraries, detailed
below to facilitate reproducibility. For snippet extraction in Step 1, we adapted the
SpacyPassageChunker class from the corpus_processing package, as provided by Dalton et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], to
allow for variable snippet sizes. The class requires spaCy [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ]; we used version 3.3.0 for our
implementation. The snippet pre-ranking in Step 2 was implemented using PyTerrier [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ],
version 0.10.0. For Step 4 we utilized ms-marco-MiniLM-L-6-v2 which has been published on
HuggingFace.co [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ]. To embed the model into our project, we used the transformers library [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ],
version 4.38.2, and the PyTorch library [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ], version 2.2.0. The results of the preparation steps
are accessible via TIRA [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] / TIREx [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ].
        </p>
      </sec>
      <sec id="sec-7-2">
        <title>B.2. Results on other evaluated datasets</title>
        <p>BM25
MonoT5
tn BM25 + TF-Snippets
em MonoT5 + TF-Snippets
ire BM25 + TF-Snippets + CE
xp MonoT5 + TF-Snippets + CE
E BM25 + CE (naive)
MonoT5 + CE (naive)</p>
        <p>BM25
MonoT5
tn BM25 + TF-Snippets
em MonoT5 + TF-Snippets
ire BM25 + TF-Snippets + CE
xp MonoT5 + TF-Snippets + CE
E BM25 + CE (naive)
MonoT5 + CE (naive)</p>
        <p>BM25
MonoT5
tn BM25 + TF-Snippets
em MonoT5 + TF-Snippets
ire BM25 + TF-Snippets + CE
xp MonoT5 + TF-Snippets + CE
E BM25 + CE (naive)
MonoT5 + CE (naive)</p>
        <p>BM25
MonoT5
tn BM25 + TF-Snippets
em MonoT5 + TF-Snippets
ire BM25 + TF-Snippets + CE
xp MonoT5 + TF-Snippets + CE
E BM25 + CE (naive)
MonoT5 + CE (naive)
0
(d) Clueweb12 (2014)
0.6
0.7
0.6
0.7
0.8</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>Query Expansion</title>
      <sec id="sec-8-1">
        <title>Detailed Results for Recall</title>
      </sec>
      <sec id="sec-8-2">
        <title>Detailed Results for nDCG</title>
      </sec>
      <sec id="sec-8-3">
        <title>C.3. Prompts</title>
        <p>Method Prompt
f’Answer the following query:’
f’’
CoT f’{q}’
f’’
f’Give the rationale before answering.’
f’For every query, suggest a similar query:’
f’’
f’Original query: How to tie a windsor knot?’
f’Similar query: Instructions for tying a windsor knot’
f’’
f’Original query: How is the weather tomorrow morning?’
Q2E/FS f’Similar query: Weather tomorrow morning’
f’’
f’Original query: Simple vegan cooking recipes’
f’Similar query: What are some delicious and simple vegan cooking recipes?’
f’’
f’Original Query: {q}’
f’Similar Query:’
f’Suggest 5 queries that are similar to the following query:’
Q2E/ZS f’’</p>
        <p>f’Query: {q}’
Q2E/ZS</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Farzana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Granitzer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Hendriksen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hiemstra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zerhoudi</surname>
          </string-name>
          , 1st International Workshop on Open Web Search (WOWS),
          <source>in: Advances in Information Retrieval. 46th European Conference on IR Research (ECIR</source>
          <year>2024</year>
          ), Lecture Notes in Computer Science, Springer,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] S. Meyer zu Eissen,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Genre classification of web pages</article-title>
          , in: S. Biundo,
          <string-name>
            <given-names>T.</given-names>
            <surname>Frühwirth</surname>
          </string-name>
          , G. Palm (Eds.),
          <source>KI 2004: Advances in Artificial Intelligence</source>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2004</year>
          , pp.
          <fpage>256</fpage>
          -
          <lpage>269</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A. Z.</given-names>
            <surname>Broder</surname>
          </string-name>
          ,
          <article-title>A taxonomy of web search</article-title>
          ,
          <source>SIGIR Forum 36</source>
          (
          <year>2002</year>
          )
          <fpage>3</fpage>
          -
          <lpage>10</lpage>
          . URL: https: //doi.org/10.1145/792550.792552. doi:
          <volume>10</volume>
          .1145/792550.792552.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ratner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Bach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ehrenberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fries</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ré</surname>
          </string-name>
          , Snorkel:
          <article-title>Rapid training data creation with weak supervision</article-title>
          ,
          <source>in: Proceedings of the VLDB endowment. International conference on very large data bases</source>
          , volume
          <volume>11</volume>
          , NIH Public Access,
          <year>2017</year>
          , p.
          <fpage>269</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pedregosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vanderplas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cournapeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brucher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Perrot</surname>
          </string-name>
          , E. Duchesnay,
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nogueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          , Passage Re-ranking
          <string-name>
            <surname>with</surname>
            <given-names>BERT</given-names>
          </string-name>
          ,
          <year>2020</year>
          . URL: http://arxiv.org/abs/
          <year>1901</year>
          . 04085. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1901</year>
          .
          <volume>04085</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dalton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Callan</surname>
          </string-name>
          ,
          <source>TREC CAsT</source>
          <year>2019</year>
          :
          <article-title>The Conversational Assistance Track Overview</article-title>
          ,
          <year>2020</year>
          . URL: http://arxiv.org/abs/
          <year>2003</year>
          .13624. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>2003</year>
          .
          <volume>13624</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Callan</surname>
          </string-name>
          ,
          <article-title>The lemur project and its clueweb12 dataset</article-title>
          ,
          <source>in: Invited talk at the SIGIR 2012 Workshop on Open-Source Information Retrieval</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Callan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yoo</surname>
          </string-name>
          ,
          <string-name>
            <surname>L</surname>
          </string-name>
          . Zhao,
          <source>Clueweb09 data set</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J. J. R.</given-names>
            <surname>Jr</surname>
          </string-name>
          .,
          <article-title>Relevance feedback in information retrieval. The SMART retrieval system: experiments in automatic document processing (</article-title>
          <year>1971</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Kor</surname>
          </string-name>
          <article-title>$age, To see, or not to see-is that the query?</article-title>
          ,
          <source>in: Proceedings of the 14th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          ,
          <year>1991</year>
          , pp.
          <fpage>134</fpage>
          -
          <lpage>141</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. R</surname>
          </string-name>
          . Kor$age, E. Rasmussen,
          <article-title>Query improvement in information retrieval using genetic algorithms-a report on the experiments of the trec project</article-title>
          ,
          <source>in: Proceedings of the Text REtrieval Conference (TREC-1)</source>
          ,
          <year>1993</year>
          , pp.
          <fpage>31</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kullback</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Leibler</surname>
          </string-name>
          ,
          <source>On Information and Sufficiency</source>
          ,
          <source>The Annals of Mathematical Statistics</source>
          <volume>22</volume>
          (
          <year>1951</year>
          )
          <fpage>79</fpage>
          -
          <lpage>86</lpage>
          . URL: https://doi.org/10.1214/aoms/1177729694. doi:
          <volume>10</volume>
          .1214/ aoms/1177729694.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Raiber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kurland</surname>
          </string-name>
          ,
          <article-title>Kullback-leibler divergence revisited</article-title>
          ,
          <source>in: Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval</source>
          , ICTIR '17,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2017</year>
          , p.
          <fpage>117</fpage>
          -
          <lpage>124</lpage>
          . URL: https: //doi.org/10.1145/3121050.3121062. doi:
          <volume>10</volume>
          .1145/3121050.3121062.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>V.</given-names>
            <surname>Lavrenko</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Relevance-based language models</article-title>
          ,
          <source>in: ACM SIGIR Forum</source>
          , volume
          <volume>51</volume>
          , ACM New York, NY, USA,
          <year>2017</year>
          , pp.
          <fpage>260</fpage>
          -
          <lpage>267</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jagerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhuang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bendersky</surname>
          </string-name>
          ,
          <article-title>Query expansion by prompting large language models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2305</volume>
          .
          <fpage>03653</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Alaofi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gallagher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sanderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Scholer</surname>
          </string-name>
          , P. Thomas,
          <article-title>Can generative llms create query variants for test collections? an exploratory study</article-title>
          ,
          <source>in: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '23,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , p.
          <fpage>1869</fpage>
          -
          <lpage>1873</lpage>
          . URL: https://doi.org/10.1145/3539618.3591960. doi:
          <volume>10</volume>
          .1145/3539618.3591960.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Albert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almahairi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Babaei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bashlykov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhargava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhosale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bikel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Blecher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Ferrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Cucurull</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Esiobu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fernandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fuller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Goswami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hartshorn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hosseini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Inan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kardas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kerkez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Khabsa</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Kloumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Korenev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Koura</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Liskovich</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Mao</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Martinet</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Mihaylov</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Mishra</surname>
            , I. Molybog,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Poulton</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Reizenstein</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Rungta</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Saladi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Schelten</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Silva</surname>
            ,
            <given-names>E. M.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Subramanian</surname>
            ,
            <given-names>X. E.</given-names>
          </string-name>
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Taylor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>J. X.</given-names>
          </string-name>
          <string-name>
            <surname>Kuan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Yan</surname>
            , I. Zarov,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Fan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kambadur</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Narang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Stojnic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Edunov</surname>
          </string-name>
          ,
          <source>T. Scialom, Llama</source>
          <volume>2</volume>
          :
          <article-title>Open foundation and fine-tuned chat models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2307</volume>
          .
          <fpage>09288</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dehghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. Q.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Garcia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. W.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bahri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zheng</surname>
          </string-name>
          , et al.,
          <article-title>Ul2: Unifying language learning paradigms</article-title>
          ,
          <source>in: The Eleventh International Conference on Learning Representations</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Claveau</surname>
          </string-name>
          ,
          <article-title>Neural text generation for query expansion in information retrieval</article-title>
          ,
          <source>in: WI-IAT 2021 - 20th IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology, Proceedings of the WI-IAT Conference</source>
          , IEEE, Melbourne, Australia,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . URL: https://hal.science/hal-03494692. doi:
          <volume>10</volume>
          .1145/3486622.3493957.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <article-title>Query2doc: Query expansion with large language models</article-title>
          ,
          <source>in: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>9414</fpage>
          -
          <lpage>9423</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Qu</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>W. X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>She</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-R.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Rocketqav2: A joint training method for dense passage retrieval and passage re-ranking</article-title>
          ,
          <source>in: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>2825</fpage>
          -
          <lpage>2835</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Reimer</surname>
          </string-name>
          , S. MacAvaney,
          <string-name>
            <given-names>N.</given-names>
            <surname>Deckers</surname>
          </string-name>
          , S. Reich, J.
          <string-name>
            <surname>Bevendorff</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>The information retrieval experiment platform</article-title>
          ,
          <source>in: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR '23</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <year>2023</year>
          . URL: http://dx.doi.org/10.1145/3539618.3591888. doi:
          <volume>10</volume>
          . 1145/3539618.3591888.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolyada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Grahm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elstner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Loebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Continuous Integration for Reproducible Shared Tasks with TIRA.io</article-title>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval. 45th European Conference on IR Research (ECIR</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>241</lpage>
          . URL: https://link.springer.com/chapter/10.1007/ 978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>20</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Reimer</surname>
          </string-name>
          , S. MacAvaney,
          <string-name>
            <given-names>N.</given-names>
            <surname>Deckers</surname>
          </string-name>
          , S. Reich, J.
          <string-name>
            <surname>Bevendorff</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>The Information Retrieval Experiment Platform</article-title>
          , in: H.
          <string-name>
            <surname>-H. Chen</surname>
            ,
            <given-names>W.-J. E.</given-names>
          </string-name>
          <string-name>
            <surname>Duh</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-H. Huang</surname>
            ,
            <given-names>M. P.</given-names>
          </string-name>
          <string-name>
            <surname>Kato</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Poblete (Eds.),
          <source>46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR</source>
          <year>2023</year>
          ), ACM,
          <year>2023</year>
          , pp.
          <fpage>2826</fpage>
          -
          <lpage>2836</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3539618.3591888. doi:
          <volume>10</volume>
          .1145/ 3539618.3591888.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bansal</surname>
          </string-name>
          , textstat, https://github.com/textstat/,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Olsen</surname>
          </string-name>
          , K. Enevoldsen,
          <article-title>TextDescriptives: A Python package for calculating a large variety of metrics from text</article-title>
          ,
          <source>Journal of Open Source Software</source>
          <volume>8</volume>
          (
          <year>2023</year>
          )
          <article-title>5153</article-title>
          . URL: https://joss.theoj.org/papers/10.21105/joss.05153. doi:
          <volume>10</volume>
          .21105/joss.05153.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>M.</given-names>
            <surname>Honnibal</surname>
          </string-name>
          , I. Montani,
          <string-name>
            <given-names>S. Van</given-names>
            <surname>Landeghem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Boyd</surname>
          </string-name>
          , spaCy: Industrial-strength
          <source>Natural Language Processing in Python (</source>
          <year>2020</year>
          ). doi:
          <volume>10</volume>
          .5281/zenodo.1212303.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Kincaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. P.</given-names>
            <surname>Fishburne</surname>
          </string-name>
          <string-name>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Rogers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. S.</given-names>
            <surname>Chissom</surname>
          </string-name>
          ,
          <article-title>Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel (</article-title>
          <year>1975</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bogert</surname>
          </string-name>
          ,
          <article-title>In defense of the fog index</article-title>
          ,
          <source>The Bulletin of the Association for Business Communication</source>
          <volume>48</volume>
          (
          <year>1985</year>
          )
          <fpage>9</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>G. H.</given-names>
            <surname>Mc</surname>
          </string-name>
          <string-name>
            <surname>Laughlin</surname>
          </string-name>
          ,
          <article-title>Smog grading-a new readability formula</article-title>
          ,
          <source>Journal of reading 12</source>
          (
          <year>1969</year>
          )
          <fpage>639</fpage>
          -
          <lpage>646</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>R.</given-names>
            <surname>Senter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <source>Automated readability index</source>
          ,
          <source>Technical Report, Technical report, DTIC document</source>
          ,
          <year>1967</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>M.</given-names>
            <surname>Coleman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Liau</surname>
          </string-name>
          ,
          <article-title>A computer readability formula designed for machine scoring</article-title>
          .,
          <source>Journal of Applied Psychology</source>
          <volume>60</volume>
          (
          <year>1975</year>
          )
          <fpage>283</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <surname>J. Anderson</surname>
          </string-name>
          ,
          <article-title>Lix and rix: Variations on a little-known readability index</article-title>
          ,
          <source>Journal of Reading</source>
          <volume>26</volume>
          (
          <year>1983</year>
          )
          <fpage>490</fpage>
          -
          <lpage>496</lpage>
          . URL: http://www.jstor.org/stable/40031755.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>H.</given-names>
            <surname>Hashemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aliannejadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Antique: A non-factoid question answering benchmark</article-title>
          ,
          <source>in: ECIR</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Downey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goharian</surname>
          </string-name>
          ,
          <article-title>Simplified data wrangling with ir_datasets</article-title>
          , in: F. Diaz,
          <string-name>
            <given-names>C.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Suel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Castells</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Jones</surname>
          </string-name>
          , T. Sakai (Eds.),
          <source>SIGIR '21: The 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Virtual Event, Canada,
          <source>July 11-15</source>
          ,
          <year>2021</year>
          , ACM,
          <year>2021</year>
          , pp.
          <fpage>2429</fpage>
          -
          <lpage>2436</lpage>
          . URL: https://doi.org/10.1145/3404835.3463254. doi:
          <volume>10</volume>
          .1145/3404835.3463254.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolyada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Grahm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elstner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Loebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Continuous Integration for Reproducible Shared Tasks with TIRA.io</article-title>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval. 45th European Conference on IR Research (ECIR</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>241</lpage>
          . URL: https://link.springer.com/chapter/10.1007/ 978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>20</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -28241-6_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>I.</given-names>
            <surname>Montani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Honnibal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Boyd</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. V.</given-names>
            <surname>Landeghem</surname>
          </string-name>
          , H. Peters, spaCy: Industrial-strength
          <source>Natural Language Processing in Python</source>
          ,
          <year>2023</year>
          . URL: https://zenodo.org/records/10009823. doi:
          <volume>10</volume>
          .5281/zenodo.10009823.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>C.</given-names>
            <surname>Macdonald</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tonellotto</surname>
          </string-name>
          ,
          <article-title>Declarative experimentation in information retrieval using pyterrier</article-title>
          ,
          <source>in: Proceedings of the 2020 ACM SIGIR on International Conference on Theory of Information Retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>161</fpage>
          -
          <lpage>168</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Louf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , P. von Platen, C. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. Le</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gugger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Drame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lhoest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rush</surname>
          </string-name>
          , Transformers:
          <article-title>State-of-the-Art Natural Language Processing</article-title>
          , in: Q. Liu, D. Schlangen (Eds.),
          <source>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>38</fpage>
          -
          <lpage>45</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          . emnlp-demos.6. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .emnlp-demos.
          <volume>6</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paszke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bradbury</surname>
          </string-name>
          , G. Chanan,
          <string-name>
            <given-names>T.</given-names>
            <surname>Killeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gimelshein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antiga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Desmaison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kopf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>DeVito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Raison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tejani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chilamkurthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Steiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bai</surname>
          </string-name>
          , S. Chintala,
          <article-title>PyTorch: An Imperative Style, High-Performance Deep Learning Library</article-title>
          , in: H.
          <string-name>
            <surname>Wallach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Beygelzimer</surname>
          </string-name>
          , F.
          <string-name>
            <surname>d'Alché Buc</surname>
          </string-name>
          , E. Fox, R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          <volume>32</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2019</year>
          , pp.
          <fpage>8024</fpage>
          -
          <lpage>8035</lpage>
          . URL: http://papers.neurips.cc/ paper/9015-pytorch
          <article-title>-an-imperative-style-high-performance-deep-learning-library</article-title>
          .pdf.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>