<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Seismic: Eficient and Efective Retrieval over Learned Sparse Representation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>SebastianBruch</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Franco MariaNardin</string-name>
          <email>francomaria.nardini@isti.cn</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cosimo Rull</string-name>
          <email>cosimo.rulli@isti.cnr</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>RossanoVenturin</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pinecone</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ISTI-CNR</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Italy</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Pisa</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Learned sparse representations form an attractive class of contextual embeddings for text retrieval thanks to their efectiveness and interpretability. Retrieval over sparse embeddings remains challenging due to the distributional diferences between learned embeddings and term frequency-based lexical models of relevance, such as BM25. Recognizing this challenge, recent research trades of exactness for eficiency, moving to approximate retrieval systems. In this wo1,rwke propose a novel organization of the inverted index that enables fast yet efective approximate retrieval over learned sparse embeddings. Our approach organizes inverted lists into geometrically-cohesive blocks, each equipped with a summary vector. During query processing, we quickly determine if a block must be evaluated using the summaries. Experiments on theSplade andE-Splade embeddings on theMs Marco andNQ datasets show that our approach is up to21× time faster than the winning (graph-based) submissions to the BigANN Challenge.</p>
      </abstract>
      <kwd-group>
        <kwd>Learned sparse representations</kwd>
        <kwd>maximum inner product search</kwd>
        <kwd>inverted index</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Learned Sparse Retrieval (LSR2), 3[
        <xref ref-type="bibr" rid="ref4 ref5 ref6">, 4, 5, 6</xref>
        ] repurposes Large Language Models to encode
an input intosparse embeddings, a vector in an inner product space where each dimension
corresponds with a term in the model’s vocabulary. LSR models are of pivotal interest as
they i) compete withdense retrieval models that encode text into dense vectors in terms of
efectiveness [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref7 ref8 ref9">7, 8, 9, 10, 11, 12, 13</xref>
        ], ii) tend to generalize better to out-of-domain da1t4a,s6e]t, s [
iii) areinterpretable by design [
        <xref ref-type="bibr" rid="ref1 ref6">6, 1</xref>
        ]. The straightforward usage of standard inverted index
for sparse embeddings is hindered by the statistical properties of the weights learned by LSR,
which do not conform to the assumptions under which popular inverted index-based retrieval
algorithms operat1e5[, 16, 17]. Hence, many recent solutions give up on exact search to boost
the eficiency of the search algorithm15[, 18], taking a leaf out of the Approximate Nearest
Neighbor (ANN) literatur1e9][. As a clear example, the 2023 BigANN Challe1ngaet NeurIPS
dedicated a track to learned sparse embeddings. Inspired by BigANN, we present a novel
ANN algorithm that we caSelilsmic (Spilled Clusetring ofInverted Lists witShummaries for
Maximum Inner Prodcut Search) and that admits efective and eficient retrieval over learned
sparse embeddings. Our solution (Sect2io) nuses in a new way two familiar data structures:
1This contribution is an extended abstract of Betraulc.h[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
the inverted and the forward index. We extend the inverted index by introducing a novel
organization of inverted lists into geometrically-cohesive blocks. Each block is equipped with a
“sketch,” serving as saummary of the vectors contained in it. The summaries allow us to skip
over a large number of blocks during retrieval and save substantial compute. Our experimental
evaluation (Sectio3n)shows thatSeismic outperforms the state-of-the-art competitors up to
21× on theSplade andE-Splade embeddings on theMs Marco andNQ datasets.
2. Methodology
2:
4:
5:
7:
 ← { | 
all ∈
      </p>
      <p>{ , }=1
Algorithm 1: Indexing.</p>
      <sec id="sec-2-1">
        <title>Input:  : sparse vectors iℝn ;</title>
        <p>: Maximum inverted list length;
 : Maximum number of blocks per
inverted list;
 : Fraction of the overall importance
preserved by each summary.</p>
        <p>Result: Seismic index.</p>
        <p>1: for  ∈ {1, … , } do
() ≠ 0,  () ∈  }</p>
      </sec>
      <sec id="sec-2-2">
        <title>3: Sort  in decreasing order b y for</title>
        <p>ℐ</p>
        <p>← { ,1 ,  ,2 , … ,  , }</p>
      </sec>
      <sec id="sec-2-3">
        <title>Cluster ℐ into partitions,</title>
        <p>6: for 1 ≤  ≤  do
8: return ℐ , { , } ∀, 
 , ←  -mass subvector o(f
, )</p>
        <sec id="sec-2-3-1">
          <title>Algorithm 2: Query processing</title>
        </sec>
        <sec id="sec-2-3-2">
          <title>Input:  : query; : number of results;</title>
          <p>cut: query entries considered;
heap_factor: correction factor for
summary inner productℐ;  ’s and
 , ’s: inverted lists and summaries .</p>
          <p>Result: A Heap with the
top</p>
          <p>documents.
1:  cut ← the topcut entries of
2: Heap← ∅
3: for  ∈  cut do
4:
5:
6:
7:
8:
9:
10:
for   ∈ ℐ do
 ← ⟨,</p>
          <p>, ⟩
if  &lt; hHeeaapp_.fmacinto()r then</p>
          <p>continue {Skip the block}
for  ∈   do
 = ⟨,</p>
          <p>ForwardIndex[]⟩</p>
          <p>UpdateHeap(Heap, p, d)
11: return Heap</p>
          <p>
            The design ofSeismic relied both on an inverted index and a forward iSnediesmx.ic uses
an organization of the inverted index that blends tosgtaettichearnddynamic pruning. The
documents pinpointed by the inverted index are then evaluated using the forward index. The
data structure and the indexing / query processing algorithm are described in detail below.
Static Pruning. Seismic heavily relies on the concentration of importance property discussed
by Bruchet al. [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. The property shows that a small subset of the most important coordinates of
the sparse embedding of a query and document vector can be used to efectively approximate
their inner product. Concretsetalyti,c pruning means that for coordin a,twee build its inverted
list by gathering al∈l
          </p>
          <p>whose  ≠ 0. We then sort the inverted list ’bsyvalue in decreasing
order (breaking ties arbitrarily), so that the document -wthhocsoeordinate has the largest
value appears at the beginning of the list. We then prune the inverted list by keeping at most
the first  entries for a fixed —our first hyper-parameter. We denote the resulting inverted list
for coordina teby ℐ .</p>
          <p>Blocking of Inverted Lists. Seismic also introduces a novel blocking strategy on inverted lists.
It partitions each inverted list insmtoall blocks—our second hyper-parameter. The rationale
behind a blocked organization of an inverted list is to group together documentsismtihlaart are
so as to facilitatdeyanamic pruning strategy.</p>
          <p>A clustering algorithm is used to partition the documents whose ids are present in an inverted
list into clusters. Each cluster is then turned into one block, consisting of the id of documents
whose vectors belong to the same cluster. Conceptually, each block is “atomic” in the following
sense: if the dynamic pruning algorithm decides we must visit a ballolctkh,e documents in that
block are fully evaluated. We note that any geometrical (supervised or unsupervised) clustering
algorithm may be readily used. We use a shallow vari2a0n]tof[K-Means; see the original
paper for more detail1s][.</p>
          <p>Per-block summary Vectors. Seismic leverages the concept osfuammary vector to determine
whether a block should be evaluated. A summa ry-diismensional vector built with the idea
to upper-bound the full inner product attainable by documents in a block. In other words, the
 -th coordinate of the summary vector of a block would contain the m a xiammuomng the
documents in that block. More precisely, our summary func∶ti2o n → ℝ takes a block
 from the universe of all bloc2ks, and produces a vector who s-eth coordinate is simply
()  = max∈   . This summary is conservative: its inner product with the query is no less
than the inner product between the query and any of its doc u⟨, m(e)⟩n≥t: ⟨, ⟩ for all
 ∈  and an arbitrary que r.y</p>
          <p>The number of non-zero entries in summary vectors grows quickly with the block size,
increasing the memory footprint and the search timSeeisomfic. To this end, we prune()
by keeping only it s -mass subvector. See the original work for the definition-moafss
subvector1[]. That, , is our third and last indexing hyper-parameter. We further reduce the
size of summaries by applying scalar quantization after min-max scaling, employing only a
single byte for each value.</p>
          <p>Indexing. We summarize the discussion above in Algorit1h.mWhen indexing a collection
 ⊂ ℝ  , for every coordinat∈e{1, … , } , we form its inverted list, recording only the document
identifiers (Line2). We then sort the list in decreasing order of values3)(,Lained apply static
pruning by keeping, for each inverted list, tehleements with the largest value (L4i)n.Wee
then apply clustering to the inverted list to derive atblmoocskts (Line5). Once documents
are assigned to the blocks, we then build the block summary using the procedure described
earlier (Lin7e).</p>
          <p>Query Processing. Algorithm2 shows the query processing logic Sineismic. We select
a subset of the query coordinatceust (Line 1), sorted by magnitude, and (b) define a novel
dynamic pruning strategy (Lin5e–s7) that allows to skip blocks in the inverted lists of the
coordinates incut. Seismic adopts a coordinate-at-a-time traversal3()Loinfethe inverted
index. For each coordina t∈e cut, it evaluates the blocks using their summary. The documents
within a block are evaluated further if the approximation with the summary is greater than a
fraction of the minimum inner product in the MHeina-p, using the Forward Index. A document
whose inner product is greater than the minimum score in thHeeMaipni-s inserted into the
heap (UpdateHeap).</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experiments</title>
      <p>
        Experimental Setup. We experiment on two publicly-available dataMsestsM:arco v1
Passage [21] and Natural QuestionNsQ() from Beir [22]. We evaluateSeismic on embeddings
generated usinSgplade [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] andE-Splade. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>We compareSeismic with five state-of-the-art retrieval solutions. In this manuscript, we only
report the comparison against the best competitors, namely the winning solutions of the “Sparse
Track” at the 2023 BigANN Challenge at NeurGIPrSa,ssRMA andPyAnn. See the original
work for the complete compariso1n].[We compare the methods using mean query latency
( sec.) and accuracy, i.e., the percentage of true nearest neighbors recalled in the returned set.
We implementedSeismic in Rust2. We conduct experiments on a server equipped with one
Intel i9-9900K CPU, clock ra3t.e60 GHz and64 GiB of RAM, with single-threaded execution.
Results Table1 details retrieval performance in terms of average per-query latency at various
accuracy cutS.eismic consistently outperforGmrsassRMA andPyAnn by a substantial margin,
ranging from2.6× (Splade on Ms Marco) to21.6× (E-Splade on Ms Marco) depending on
the level of accuracy. In fact, as accuracy increases, the latency gap Sbeitswmeicenand the
two graph-based methods widens. This gap is much larger when query vectors are sparser,
such as withE-Splade embeddings. That is because, when queries are highly sparse, inner
products between queries and documents become smaller, reducing the eficacy of a greedy graph
traversal. As one data poiPnytA, nn overE-Splade embeddings ofMs Marco visits roughly
40,000 documents to reac9h7% accuracy, whereaSseismic evaluates jus2t,198 documents.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions and Future Work</title>
      <p>This paper presentSseismic, a novel approach for eficient and efective retrieval over sparse
learned representations. Our solution outperforms the state-of-art graph-based solutions for
eficient sparse retrieval up to a factor21o×f on theSplade andE-Splade embeddings on
theMs Marco dataset. As future work, we intend to explore the application of compression
techniques for inverted lis2t3s][to further reduce the size of inverted and forward indexes.
2Our code is publicly availablehattps://github.com/TusKANNy/seism.ic
[14] S. Bruch, S. Gai, A. Ingber, An analysis of fusion functions for hybrid retrieval, ACM</p>
      <p>Transactions on Information Systems 42 (2023).
[15] S. Bruch, F. M. Nardini, A. Ingber, E. Liberty, An approximate algorithm for maximum
inner product search over streaming sparse vectors, ACM Transactions on Information
Systems 42 (2023).
[16] J. Mackenzie, A. Trotman, J. Lin, Wacky weights in learned sparse representations and the
revenge of score-at-a-time query evaluation, 2a0r2X1i.v:2110.11540.
[17] M. Crane, J. S. Culpepper, J. Lin, J. Mackenzie, A. Trotman, A comparison of
document-at-atime and score-at-a-time query evaluation, in: Proceedings of the 10th ACM International
Conference on Web Search and Data Mining, 2017, pp. 201–210.
[18] S. Bruch, F. M. Nardini, A. Ingber, E. Liberty, Bridging dense and sparse maximum inner
product search, 2023a.rXiv:2309.09013.
[19] S. Bruch, Foundations of Vector Retrieval, Springer Nature Switzerland, 2024.
[20] F. Chierichetti, A. Panconesi, P. Raghavan, M. Sozio, A. Tiberi, E. Upfal, Finding near
neighbors through cluster pruning, in: Proceedings of the Twenty-Sixth ACM
SIGMODSIGACT-SIGART Symposium on Principles of Database Systems, 2007, pp. 103–112.
[21] T. Nguyen, M. Rosenberg, X. Song, J. Gao, S. Tiwary, R. Majumder, L. Deng, Ms marco: A
human generated machine reading comprehension dataset (2016).
[22] N. Thakur, N. Reimers, A. Rücklé, A. Srivastava, I. Gurevych, BEIR: A heterogeneous
benchmark for zero-shot evaluation of information retrieval models, in: 35th Conference
on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2),
2021.
[23] G. E. Pibiri, R. Venturini, Techniques for inverted index compression, ACM Computing
Surveys 53 (2021) 125:1–125:36.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bruch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Nardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rulli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Venturini</surname>
          </string-name>
          ,
          <article-title>Eficient inverted indexes for approximate retrieval over learned sparse representations</article-title>
          ,
          <source>in: Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          , p.
          <fpage>152</fpage>
          -
          <lpage>162</lpage>
          .
          <year>1d0o</year>
          .i:
          <volume>1145</volume>
          /3626772.3657769.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Nardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Perego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tonellotto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goharian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Frieder</surname>
          </string-name>
          ,
          <article-title>Expansion via prediction of importance with contextualization</article-title>
          ,
          <source>in: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1573</fpage>
          -
          <lpage>1576</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Formal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Piwowarski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clinchant</surname>
          </string-name>
          , Splade:
          <article-title>Sparse lexical and expansion model for ifrst stage ranking</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>2288</fpage>
          -
          <lpage>2292</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Formal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lassance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Piwowarski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clinchant</surname>
          </string-name>
          , Splade v2:
          <article-title>Sparse lexical and expansion model for information retrieval, 2a02r1X</article-title>
          .iv:
          <volume>2109</volume>
          .
          <fpage>10086</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Formal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lassance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Piwowarski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clinchant</surname>
          </string-name>
          ,
          <article-title>From distillation to hard negative sampling: Making sparse neural ir models more efective</article-title>
          ,
          <source>in: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>2353</fpage>
          -
          <lpage>2359</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lassance</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Clinchant</surname>
          </string-name>
          ,
          <article-title>An eficiency study for splade models</article-title>
          ,
          <source>in: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>2220</fpage>
          -
          <lpage>2226</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. F.</given-names>
            <surname>Nogueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          ,
          <article-title>Pretrained Transformers for Text Ranking: BERT and Beyond</article-title>
          ,
          <source>Synthesis Lectures on Human Language Technologies</source>
          , Morgan &amp; Claypool Publishers,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Oguz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Min</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Edunov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          , W.-t. Yih,
          <article-title>Dense passage retrieval for open-domain question answering</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>6769</fpage>
          -
          <lpage>6781</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.-F.</given-names>
            <surname>Tang</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bennett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ahmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Overwijk</surname>
          </string-name>
          ,
          <article-title>Approximate nearest neighbor negative contrastive learning for dense text retrieval</article-title>
          ,
          <source>in: International Conference on Learning Representations</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Santhanam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Saad-Falcon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Potts</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Zaharia, ColBERTv2: Efective and eficient retrieval via lightweight late interaction</article-title>
          ,
          <source>in: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>3715</fpage>
          -
          <lpage>3734</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaharia</surname>
          </string-name>
          ,
          <article-title>Colbert: Eficient and efective passage search via contextualized late interaction over bert</article-title>
          ,
          <source>in: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>48</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Nardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rulli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Venturini</surname>
          </string-name>
          ,
          <article-title>Eficient multi-vector dense retrieval with bit vectors</article-title>
          ,
          <source>in: Advances in Information Retrieval</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>