<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>CRAWLDoc: A Dataset for Robust Ranking of Bibliographic Documents</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fabian Karl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ansgar Scherp</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universität Ulm</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>32</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>Publication databases rely on accurate metadata extraction from diverse web sources, yet variations in web layouts and data formats present challenges for metadata providers. This paper introduces CRAWLDoc, a new method for contextual ranking of linked web documents. Starting with a publication's URL, such as a digital object identifier, CRAWLDoc retrieves the landing page and all linked web resources, including PDFs, ORCID profiles, and supplementary materials. It embeds these resources, along with anchor texts and the URLs, into a unified representation. For evaluating CRAWLDoc, we have created a new, manually labeled dataset of 600 publications from six top publishers in computer science. Our method CRAWLDoc demonstrates a robust and layout-independent ranking of relevant documents across publishers and data formats. It lays the foundation for improved metadata extraction from web documents with various layouts and formats. Our source code and dataset can be accessed at https://github.com/FKarl/CRAWLDoc.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Scholarly Dataset</kwd>
        <kwd>Bibliographic Metadata</kwd>
        <kwd>Information Retrieval</kwd>
        <kwd>Language Model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Databases such as Web of Science, Crossref, and DBLP are crucial academic resources of bibliographic
information. Identifying high-quality metadata sources about new publications is essential for these
services. While there are methods and tools for extracting bibliographic metadata [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], these are
typically restricted to a single document like a PDF. Currently, many potential web sources and content
that may contain valuable metadata are underutilized. This is due to source heterogeneity of web
layouts, document types, and formats, including full texts, publication PDFs, publisher landing pages,
ORCIDs, and other web content.
      </p>
      <p>
        We consider the example of DBLP, the de facto main metadata provider in computer science. The
main strategy for integrating publisher-provided metadata is to implement publisher-specific wrappers,
an approach that is time-consuming and requires maintenance whenever the publisher changes its
website [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Thus, an automated service is needed to systematically search for bibliographic metadata
sources across multiple web documents. Often, bibliographic information cannot be found on a single
website, e. g., the publication’s landing page, necessitating to harvest linked documents and identifying
those relevant to the publication. Identifying relevant linked documents is challenging because two web
documents with similar layouts and text can refer to diferent papers with paper-specific components
like titles, authors, and afiliations. Another challenge is the heterogeneity of web data. Important
documents can be in HTML or other formats like PDF. Another reason is that using wrappers or APIs
relies on crawling publisher websites, which is expensive to maintain [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        We propose a novel retrieval system CRAWLDoc (Contextual RAnking of Web-Linked Documents),
see Figure 1, that can automatically identify relevant data sources from diverse web sources. Input is a
Digital Object Identifier (DOI) of a publication, which is provided by publishers [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The web content
linked from this seed URI is harvested and analyzed. We identify relevant linked content referring to the
same paper as the DOI that may carry metadata. To this end, we embed the source document and linked
Web Scraping from Seed URI
      </p>
      <p>
        Neural Document Ranking
documents along with their associated anchor texts and URLs into a shared vector space and treat the
publication’s landing page as the query. A ranking is computed by the similarity between the landing
page embedding and the embeddings of linked documents, efectively identifying the most relevant
sources for metadata extraction. By embedding the content, we efectively address the challenge that
the web sources from which the data is extracted are highly diverse and vary in structure and format [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>We evaluate CRAWLDoc on a new dataset derived from DBLP, which comprises 600 publications from
the six largest computer science publishers. Our dataset is unique as it provides annotated relevancy
labels for all outgoing links from publication landing pages, along with bibliographic metadata, including
titles, years, authors’ names, and afiliations manually. Our experiments show that CRAWLDoc reliably
identifies relevant web documents based on a single seed document.</p>
      <p>A leave-one-out experiment shows that our system is robust w. r. t. the retrieval from websites
of various layouts from publishers that were excluded from the training dataset. In summary, our
contributions are:
• A document-as-query approach CRAWLDoc to determine relevant documents that encodes web
content of various formats, anchor text, and URIs in a single embedding space.
• Evaluating 600 publications from the six largest publishers in computer science.
• A robustness check by training on five publishers and testing on a held-out publisher.
• A new dataset of bibliographic metadata with author afiliations, along with relevancy information
for linked web documents.</p>
      <p>Below, we summarize related work. We introduce our CRAWLDoc metadata retrieval system in
Section 3. The experimental apparatus is described in Section 4. The results are described in Section 5
and discussed in Section 6, before we conclude and outline future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        We discuss research in neural information retrieval and layout-aware language models. Neural
Information Retrieval (NIR) is a prominent research area, utilizing neural networks to improve the
retrieval process. The landscape of NIR research has been extensively surveyed [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ], highlighting
the use of learned representations of queries and documents, commonly referred to as embeddings.
These embeddings capture semantic similarities that traditional information retrieval models might
overlook [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. The BERT model [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], although not specifically designed for information retrieval,
has profoundly impacted NIR [
        <xref ref-type="bibr" rid="ref12 ref13 ref14">12, 13, 14</xref>
        ]. BERT-based models such as CEDR [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] have achieved
impressive performance on various information retrieval benchmarks. The ColBERT model [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] introduced a
late interaction paradigm, enabling eficient and efective passage retrieval. ColBERT’s ability to balance
efectiveness and eficiency has made it a popular choice for large-scale retrieval tasks [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>Publish with us</p>
      <p>Help
View PDF
TeX Source
BibTeX
Related Papers
Keywords
&lt; prev | next &gt;
identifying
document</p>
      <p>
        Layout-infused language models consider both textual content and spatial layout.
LayoutLMv3 [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] exemplifies this concept by pre-training multimodal transformers with a unified text and
image masking objective. Another approach is DocLLM [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], which does not rely on expensive image
encoders but relies solely on bounding box information from optical character recognition (OCR).
LMDX [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] is a model-agnostic method to adapt arbitrary Large Language Models (LLMs) for document
information extraction. It extracts text with OCR and enriches it with layout information. The model
proposes an XML-style prompt for information extraction and trains a text-only LLM with text and
bounding boxes. Layout-infused LLMs can face challenges with layout distribution shifts. Chen et al.
[
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] note that model performance can degrade by up to 20 points in macro F1 score under layout
distribution shifts.
3.
      </p>
    </sec>
    <sec id="sec-3">
      <title>CRAWLDoc Metadata Retrieval</title>
      <p>We introduce CRAWLDoc (Contextual RAnking of Web-Linked Documents), a novel system for
identifying relevant bibliographic sources across web documents. Based on a seed URI, a DOI of a publication,
CRAWLDoc scrapes linked resources. Subsequently, the retrieved web documents in the form of H TML
or PDF are ranked using a Small Language Model (SLM) [22]. Our primary assumption is that all
necessary information can be found within a one-hop crawl of the landing page associated with the
DOI. This assumption is based on our observation that publishers present key bibliographic information
on the landing page or pages directly linked to it e. g., the PDF of the publication.</p>
      <p>Web Scraping from Seed URI The initial step of our system involves web scraping, starting with a
DOI as the input and progressing to the scraping of the corresponding web page. After this starting
point, all documents linked from the seed URI are retrieved, which may be formatted in H TML or
PDF. Both PDF and H TML files undergo a series of steps to extract the relevant text and its associated
bounding boxes to also capture layout information. For PDF documents, the text and its corresponding
bounding box coordinates are directly extracted from the file using the PDFMiner Python library. In
the case of H TML documents, the page is first rendered in a Firefox web browser (Version: 129.0.2)
to accurately present the content’s formatting and layout, and then the text and bounding boxes are
extracted. This information is then converted into a uniform textual JSON format and serves as the
input for the neural document ranking. Figure 2 illustrates the diferent steps to create our document
representation.</p>
      <p>Neural Document Ranking In the second step, we employ a SLM to create embeddings of the
documents along with their associated anchor texts and URLs. For each document, we construct a single
input representation by concatenating the anchor text, URL, and document content using a special
separator token ([SEP]). This representation is then embedded into a dense vector space. The document
originating from the DOI is embedded utilizing a query encoder, and all documents linked from the
landing page are embedded with the document encoder. A Maximum Inner Product Search (MIPS) is
performed with the embedding of the landing page and the embeddings of all scraped documents to
create a Contextual RAnking of Web-Linked Documents (CRAWLDoc) based on the landing page.</p>
      <p>
        We use the jina-embeddings-v2 model [23] as neural retriever. It is based on a BERT [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]
architecture and supports the symmetric bidirectional variant of ALiBi [24], allowing for a sequence
length of up to 81,921 tokens. Due to memory restrictions, we limit our experiments to the first 2,048
tokens. The neural retriever is trained using contrastive learning with the InfoNCE loss function [25].
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Apparatus</title>
      <p>Dataset We take a subset of bibliographies from the six largest publishers in the DBLP Computer
Science Bibliography dataset [26]. The publishers represent more than 80% of all publications listed in
DBLP. This ensures the dataset contains a representative set of layouts encountered in bibliographic
web content. We randomly select 100 publications for each publisher and split them into training,
validation, and test sets in an 80/10/10 ratio with equal per-publisher distribution.</p>
      <p>We obtained the metadata for each publication by manually retrieving the title, publication year, and
authors’ names and afiliations. We retrieved the landing page of each publication and labeled every
outgoing link on the landing page with a binary relevancy label. This label indicates whether the linked
website or document is about the same publication . By manually creating this dataset, we ensure a high
quality of the metadata and can accurately assess the document retrieval process in our proposed setup.</p>
      <p>To prevent artificial inflation of our performance metrics, we identified and removed any instances in
our test set where the landing page contained links to itself. The trivial nature of calculating document
similarity to itself would otherwise result in an unrepresentative boost in ranking performance.</p>
      <p>Our dataset consists of 600 publications with detailed metadata and 72,483 linked documents with
binary relevancy labels. Per publication, we have an average of 3.63 (SD: 2.10) authors, with an average
of 1.14 (SD: 0.41) afiliations per author. Furthermore, there is an average of 120.81 (SD: 76.52) linked
websites per landing page and an average of only 5.45 (SD: 2.99) relevant websites per publication.
To the best of our knowledge, we are the first to release a dataset that includes author afiliations as
mentioned in the publications. Additionally, we are the first to provide relevancy labels for linked
documents in the context of publication web data. For legal purposes, we are only able to publish the
labels and not the actual websites. However, we do ofer the source code for our procedure.</p>
      <p>The DBLP dataset is released under CC0 1.0 Public Domain Dedication license. Our annotations have
the same license.</p>
      <p>Procedure Our experimental procedure for document ranking involves fine-tuning a neural document
retriever using contrastive loss to improve document ranking. To ensure robust performance, we
evaluate the ranking capabilities on both in-distribution and out-of-distribution data. We optimize the
hyperparameters of the neural document retriever resulting in a learning rate of 3e-05, 32 accumulation
steps, and patience of 5, resulting in 16 epochs.</p>
      <p>Metrics To evaluate the ranking of the web documents, we employ several metrics. The Mean
Reciprocal Rank (MRR) evaluates the efectiveness of a retrieval system by considering the rank position
of the first relevant result. The MRR focuses on the first relevant document in the ranked list, i. e.,
it favors a relevant document in the highest position. In contrast, Mean Average Precision (MAP)
evaluates the precision of a retrieval system by averaging the precision scores at all ranks where
relevant documents are found and then averaging these scores over all queries. Normalized Discounted
Cumulative Gain (nDCG) [27] measures the usefulness of a document based on its position in the result
list, assuming that highly relevant documents are more useful when appearing earlier. We further
calculate the precision@k, recall@k, and F1@k which measure the proportion of relevant items in the
top  results.
.</p>
      <p>7
0.344
0.972
0.510</p>
      <p>2</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>The ranking metrics for identifying relevant linked documents are shown in Table 1. Overall, we achieve
a very high average ranking performance with MRR of 0.967, MAP of 0.987, and nDCG of 0.961. The
MRR, MAP, and nDCG values exhibit a consistently high level of performance for all six publishers,
except for MRR and nDCG on the Springer dataset. The MRR for IEEE, Elsevier, ACM, arXiv, and MDPI
all achieve the maximum score of 1.000, indicating that a relevant document is always placed at the top
position. To understand the impact of layout information on ranking performance, we conducted an
ablation study. The results without layout information showed slightly lower performance with a MRR
of 0.950, MAP of 0.976, and nDCG of 0.952.</p>
      <p>We have conducted a more detailed examination of the ranking performance with diferent cut-of
values  presented in Table 2. The recall increases with increasing values of , reaching 0.951 at  = 10.
Precision declines from 0.972 for  = 1 to 0.416 for  = 10. The F1@k score, which combines precision
and recall, reaches its highest value of 0.772 for  = 4 and  = 5.</p>
      <p>We have evaluated robustness using a leave-one-out strategy, training on all but one publisher and
testing on the left-out publisher. The results of the robustness analysis are shown in Table 3. We obtain
a high performance across all publishers, with an average MRR of 0.959, MAP of 0.968, and nDCG
of 0.961. This is less than one point for MRR and nDCG and less than two points for MAP compared
to using the full training dataset shown in Table 1. The results were particularly strong for IEEE and
arXiv, both achieving the maximum score of 1.000 for all three metrics. However, the performance was
slightly lower for Springer, consistent with the result on the full training set.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>Document Ranking Our system shows impressive overall ranking performance, with documents
ranked at the top being mostly relevant. This trend is seen when examining the evaluation of ranking
performance at various cutof values. We notice a sharp rise in recall@  for the first few documents
but only minor enhancements after around five documents. The decline in precision@  as  values
increase is a natural result considering that a publication has on average 5.45 relevant documents (see
Section 4). This is also reflected in the F1@  score, which is peaking at  = 4 and  = 5. Overall, the
results show that CRAWLDoc maintains a good balance between precision and recall with a cut-of
value of  = 5.</p>
      <p>
        Upon investigating cases where the model ranked irrelevant documents higher than relevant ones, we
could not identify a general error pattern. The errors we observed were predominantly paper-specific
rather than systematic. For example, errors occurred when the model ranked links from the references
section of a paper highly or when it assigned high ranks to diferent chapters of the same book. In
particular, Springer publications presented more special cases than other publishers in our dataset.
Robustness of Document Ranking Our model demonstrates strong robustness across diferent
publishers. While previous research, such as Chen et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], has identified challenges for layout-infused
LLMs when dealing with layout distribution shifts, our system shows consistent performance. This
is evidenced by nearly equivalent performance between in-distribution and out-of-distribution data,
suggesting efective generalization. Academic publishers often follow similar design patterns for their
publication pages, reducing the efective layout distribution shift between sources. Our robustness
evaluation considered six major publishers. The conventional nature of academic publication layouts
suggests our model likely generalizes to a broader range of publishers.
      </p>
      <p>Generalization and Threat to Validity The generalizability of our work refers to diferent publishers
based on a leave-one-out test (Table 3) in which we test the system for publishers on which it has
not been trained. Our robustness check demonstrates that a trained model can achieve comparably
good results on out-of-distribution data within our tested scope. This finding suggests that our model
has learned generalizable features of document relevance beyond the specific layouts and publishers
in our training data. Our approach of transforming diferent document formats (HTML and PDF)
into a uniform textual representation enhances its potential for generalization. This uniformity in
representation suggests applicability to other web and document-related tasks.</p>
      <p>While our study provides robust results, it is important to reflect on potential threats to validity. One
such threat is the limited scope of our investigation, which focuses on only six publishers, primarily
from the computer science field. However, the threat is reduced as these publishers represent more than
80% of computer science publications and provide various formats. Nevertheless, we acknowledge that
true generalizability to the remaining 20% of publications, which may exhibit greater variability in
their document layouts and metadata presentation, remains to be thoroughly tested in future work.
An additional possible risk is the presence of recency bias in our dataset, given that most publications
in DBLP are from more recent years. Nevertheless, we have found that older publications, including
papers as far back as 1967, in our test set achieve similar performance to more recent ones, which
eases this concern. This indicates that the performance of our model is not much influenced by the
publication year.</p>
      <p>Although there is no particular reason why other embedding models could not be used, our work
does not focus on finding optimal embedding models for the retrieval tasks. We use Jina embeddings
because they are widely used and have demonstrated strong results [23].</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion and Future Work</title>
      <p>Our Contextual RAnking of Web-Linked Documents (CRAWLDoc) retrieval system efectively identifies
relevant bibliographic documents across diverse web sources. The key scientific findings include robustly
identifying pertinent web documents and the system’s consistent performance across publishers with
varying web layouts. The insights presented in this study can potentially advance the management and
enrichment of comprehensive bibliographic databases.</p>
      <p>
        Although our model’s performance is already very strong, rerankers [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] could improve document
ranking accuracy. Future work could also explore alternative neural retriever setups like ColBERTv2 [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
and token-level representation of documents with MaxSim [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] instead of cosine similarity. In the next
steps, we plan to run diferent metadata extractor components and setups on the CRAWLDoc-ranked list
of web resources. Furthermore, we plan to evaluate CRAWLDoc in the context of the DBLP workflow.
      </p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>We thank Florian Reitz from DBLP for valuable feedback. The authors acknowledge support by the
state of Baden-Württemberg through bwHPC. This research is co-funded by the SmartER project (No.
515537520) of the DFG, German Research Foundation.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Writefull and Grammarly to check spelling and
grammar. They also used Writefull and Claude 3.5 Sonnet to improve the writing style. After using
these tools, the authors reviewed and edited the content as needed and take full responsibility for the
publication’s content.
ACL 2023, Toronto, Canada, July 9-14, 2023, Association for Computational Linguistics, 2023, pp.
13345–13360. URL: https://doi.org/10.18653/v1/2023.nfidings-acl.844. doi: 10.18653/V1/2023.</p>
      <p>FINDINGS-ACL.844.
[22] Z. Lu, X. Li, D. Cai, R. Yi, F. Liu, X. Zhang, N. D. Lane, M. Xu, Small language models: Survey,
measurements, and insights, CoRR abs/2409.15790 (2024). URL: https://doi.org/10.48550/arXiv.
2409.15790. doi:10.48550/ARXIV.2409.15790. arXiv:2409.15790.
[23] M. Günther, J. Ong, I. Mohr, A. Abdessalem, T. Abel, M. K. Akram, S. Guzman, G. Mastrapas,
S. Sturua, B. Wang, M. Werk, N. Wang, H. Xiao, Jina embeddings 2: 8192-token general-purpose
text embeddings for long documents, CoRR abs/2310.19923 (2023). URL: https://doi.org/10.48550/
arXiv.2310.19923. doi:10.48550/ARXIV.2310.19923. arXiv:2310.19923.
[24] O. Press, N. A. Smith, M. Lewis, Train short, test long: Attention with linear biases enables input
length extrapolation, in: The Tenth International Conference on Learning Representations, ICLR
2022, Virtual Event, April 25-29, 2022, OpenReview.net, 2022. URL: https://openreview.net/forum?
id=R8sQPpGCv0.
[25] A. van den Oord, Y. Li, O. Vinyals, Representation learning with contrastive predictive coding,</p>
      <p>CoRR abs/1807.03748 (2018). URL: http://arxiv.org/abs/1807.03748. arXiv:1807.03748.
[26] dblp Team, dblp computer science bibliography – Monthly Snapshot XML Release of April 2024,
2024. URL: https://doi.org/10.4230/dblp.xml.2024-04-01. doi:10.4230/dblp.xml.2024-04-01.
[27] K. Järvelin, J. Kekäläinen, Cumulated gain-based evaluation of IR techniques, ACM Trans. Inf.</p>
      <p>Syst. 20 (2002) 422–446. URL: http://doi.acm.org/10.1145/582415.582418. doi:10.1145/582415.
582418.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Grobid</surname>
          </string-name>
          , https://github.com/kermitt2/grobid, 2008-
          <fpage>2023</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Tkaczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Szostek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fedoryszak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Dendek</surname>
          </string-name>
          , L. Bolikowski,
          <article-title>CERMINE: automatic extraction of structured metadata from scientific literature</article-title>
          ,
          <source>Int. J. Document Anal. Recognit</source>
          .
          <volume>18</volume>
          (
          <year>2015</year>
          )
          <fpage>317</fpage>
          -
          <lpage>335</lpage>
          . URL: https://doi.org/10.1007/s10032-015-0249-8. doi:
          <volume>10</volume>
          .1007/S10032-015-0249-8.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Schenkel</surname>
          </string-name>
          ,
          <article-title>Integrating and exploiting public metadata sources in a bibliographic information system</article-title>
          , in: P. Mayr, I. Frommholz, G. Cabanac (Eds.),
          <source>Proceedings of the 7th International Workshop on Bibliometric-enhanced Information Retrieval (BIR</source>
          <year>2018</year>
          )
          <article-title>co-located with the 40th European Conference on Information Retrieval (ECIR</article-title>
          <year>2018</year>
          ), Grenoble, France, March
          <volume>26</volume>
          ,
          <year>2018</year>
          , volume
          <volume>2080</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>21</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-2080/paper2.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ley</surname>
          </string-name>
          , DBLP
          <article-title>- some lessons learned</article-title>
          ,
          <source>Proc. VLDB Endow</source>
          .
          <volume>2</volume>
          (
          <year>2009</year>
          )
          <fpage>1493</fpage>
          -
          <lpage>1500</lpage>
          . URL: http: //www.vldb.org/pvldb/vol2/vldb09-
          <fpage>98</fpage>
          .pdf.
          <source>doi:10.14778/1687553</source>
          .1687577.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yuan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Liu</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>C.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <article-title>Large language models for information retrieval: A survey</article-title>
          ,
          <source>CoRR abs/2308</source>
          .07107 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv. 2308.07107. doi:
          <volume>10</volume>
          .48550/ARXIV.2308.07107. arXiv:
          <volume>2308</volume>
          .
          <fpage>07107</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Rahman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Braylan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Dang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>McNamara</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Angert</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Banner</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Khetan</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>McDonnell</surname>
            ,
            <given-names>A. T.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>B. C.</given-names>
          </string-name>
          <string-name>
            <surname>Wallace</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Lease</surname>
          </string-name>
          ,
          <article-title>Neural information retrieval: A literature review</article-title>
          ,
          <source>CoRR abs/1611</source>
          .06792 (
          <year>2016</year>
          ). URL: http://arxiv.org/abs/ 1611.06792. arXiv:
          <volume>1611</volume>
          .
          <fpage>06792</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. B.</given-names>
            <surname>Croft</surname>
          </string-name>
          , X. Cheng,
          <article-title>A deep look into neural ranking models for information retrieval</article-title>
          ,
          <source>Inf. Process. Manag</source>
          .
          <volume>57</volume>
          (
          <year>2020</year>
          )
          <article-title>102067</article-title>
          . URL: https://doi.org/10.1016/j.ipm.
          <year>2019</year>
          .
          <volume>102067</volume>
          . doi:
          <volume>10</volume>
          .1016/J.IPM.
          <year>2019</year>
          .
          <volume>102067</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Rahman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Braylan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Dang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>McNamara</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Angert</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Banner</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Khetan</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>McDonnell</surname>
            ,
            <given-names>A. T.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>B. C.</given-names>
          </string-name>
          <string-name>
            <surname>Wallace</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Lease</surname>
          </string-name>
          ,
          <article-title>Neural information retrieval: A literature review</article-title>
          ,
          <source>CoRR abs/1611</source>
          .06792 (
          <year>2016</year>
          ). URL: http://arxiv.org/abs/ 1611.06792. arXiv:
          <volume>1611</volume>
          .
          <fpage>06792</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mitra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <article-title>An introduction to neural information retrieval</article-title>
          ,
          <source>Found. Trends Inf. Retr</source>
          .
          <volume>13</volume>
          (
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>126</lpage>
          . URL: https://doi.org/10.1561/1500000061. doi:
          <volume>10</volume>
          .1561/1500000061.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Abbasiantaeb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Momtazi</surname>
          </string-name>
          ,
          <article-title>Text-based question answering from information retrieval and deep neural network perspectives: A survey, WIREs Data Mining Knowl</article-title>
          . Discov.
          <volume>11</volume>
          (
          <year>2021</year>
          ). URL: https://doi.org/10.1002/widm.1412. doi:
          <volume>10</volume>
          .1002/WIDM.1412.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the</source>
          <year>2019</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis</article-title>
          , MN, USA, June 2-7,
          <year>2019</year>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://doi.org/10.18653/v1/n19-
          <fpage>1423</fpage>
          . doi:
          <volume>10</volume>
          .18653/v1/n19-
          <fpage>1423</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>X.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Retrieval of scientific documents based on HFS and BERT</article-title>
          , IEEE Access 9
          <article-title>(</article-title>
          <year>2021</year>
          )
          <fpage>8708</fpage>
          -
          <lpage>8717</lpage>
          . URL: https://doi.org/10.1109/ACCESS.
          <year>2021</year>
          .
          <volume>3049391</volume>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2021</year>
          .
          <volume>3049391</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. T. R.</given-names>
            <surname>Laskar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhuiyan</surname>
          </string-name>
          ,
          <article-title>Utilizing BERT for information retrieval: Survey, applications, resources, and challenges</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>56</volume>
          (
          <year>2024</year>
          )
          <volume>185</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>185</lpage>
          :
          <fpage>33</fpage>
          . URL: https://doi.org/10.1145/3648471. doi:
          <volume>10</volume>
          .1145/3648471.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          , É. Gaussier,
          <article-title>Intra-document block pre-ranking for bert-based long document information retrieval - abstract</article-title>
          , in: L.
          <string-name>
            <surname>Tamine</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Amigó</surname>
          </string-name>
          , J. Mothe (Eds.),
          <source>Proceedings of the 2nd Joint Conference of the Information Retrieval Communities in Europe (CIRCLE</source>
          <year>2022</year>
          ), Samatan, Gers, France,
          <source>July 4-7</source>
          ,
          <year>2022</year>
          , volume
          <volume>3178</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2022</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-3178/CIRCLE_2022_paper_27.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <string-name>
            <surname>N.</surname>
          </string-name>
          <article-title>Goharian, CEDR: contextualized embeddings for document ranking</article-title>
          , in: B.
          <string-name>
            <surname>Piwowarski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chevalier</surname>
            , É. Gaussier,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Maarek</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Nie</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Scholer</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>SIGIR</surname>
          </string-name>
          <year>2019</year>
          , Paris, France,
          <source>July 21-25</source>
          ,
          <year>2019</year>
          , ACM,
          <year>2019</year>
          , pp.
          <fpage>1101</fpage>
          -
          <lpage>1104</lpage>
          . URL: https://doi.org/10.1145/3331184.3331317. doi:
          <volume>10</volume>
          .1145/3331184.3331317.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaharia</surname>
          </string-name>
          ,
          <article-title>Colbert: Eficient and efective passage search via contextualized late interaction over BERT</article-title>
          , in: J. X.
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , X. Cheng, J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Murdock</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wen</surname>
          </string-name>
          , Y. Liu (Eds.),
          <source>Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval</source>
          ,
          <string-name>
            <surname>SIGIR</surname>
          </string-name>
          <year>2020</year>
          ,
          <string-name>
            <given-names>Virtual</given-names>
            <surname>Event</surname>
          </string-name>
          , China,
          <source>July 25-30</source>
          ,
          <year>2020</year>
          , ACM,
          <year>2020</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>48</lpage>
          . URL: https://doi.org/10.1145/3397271.3401075. doi:
          <volume>10</volume>
          .1145/3397271.3401075.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>K.</given-names>
            <surname>Santhanam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Saad-Falcon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Potts</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Zaharia, Colbertv2: Efective and eficient retrieval via lightweight late interaction</article-title>
          , in: M.
          <string-name>
            <surname>Carpuat</surname>
            , M. de Marnefe,
            <given-names>I. V. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ruíz</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the</source>
          <year>2022</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</article-title>
          ,
          <string-name>
            <surname>NAACL</surname>
          </string-name>
          <year>2022</year>
          , Seattle, WA,
          <string-name>
            <surname>United</surname>
            <given-names>States</given-names>
          </string-name>
          ,
          <source>July 10-15</source>
          ,
          <year>2022</year>
          , Association for Computational Linguistics,
          <year>2022</year>
          , pp.
          <fpage>3715</fpage>
          -
          <lpage>3734</lpage>
          . URL: https://doi.org/10.18653/v1/
          <year>2022</year>
          .naacl-main.
          <volume>272</volume>
          . doi:
          <volume>10</volume>
          .18653/V1/
          <year>2022</year>
          .NAACL-MAIN.
          <year>272</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <article-title>Layoutlmv3: Pre-training for document AI with unified text and image masking</article-title>
          , in: J.
          <string-name>
            <surname>Magalhães</surname>
            ,
            <given-names>A. D.</given-names>
          </string-name>
          <string-name>
            <surname>Bimbo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Satoh</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Sebe</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Alameda-Pineda</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>Jin</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Oria</surname>
          </string-name>
          , L. Toni (Eds.),
          <source>MM '22: The 30th ACM International Conference on Multimedia, Lisboa, Portugal, October 10 - 14</source>
          ,
          <year>2022</year>
          , ACM,
          <year>2022</year>
          , pp.
          <fpage>4083</fpage>
          -
          <lpage>4091</lpage>
          . URL: https://doi.org/10.1145/3503161.3548112. doi:
          <volume>10</volume>
          .1145/3503161.3548112.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Raman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sibue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ma</surname>
          </string-name>
          , P. Babkin,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kaur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nourbakhsh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Docllm: A layout-aware generative language model for multimodal document understanding</article-title>
          ,
          <source>CoRR abs/2401</source>
          .00908 (
          <year>2024</year>
          ). URL: https://doi.org/10.48550/arXiv.2401.00908. doi:
          <volume>10</volume>
          .48550/ARXIV. 2401.00908. arXiv:
          <volume>2401</volume>
          .
          <fpage>00908</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Perot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Luisier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Boppana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , N. Hua,
          <article-title>LMDX: language model-based document information extraction and localization</article-title>
          ,
          <source>CoRR abs/2309</source>
          .10952 (
          <year>2023</year>
          ). URL: https://doi.org/10.48550/arXiv.2309.10952. doi:
          <volume>10</volume>
          .48550/ARXIV. 2309.10952. arXiv:
          <volume>2309</volume>
          .
          <fpage>10952</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Stanovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Downey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <article-title>Are layout-infused language models robust to layout distribution shifts? A case study with scientific documents</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <article-title>Findings of the Association for Computational Linguistics:</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>