<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Italian Information Retrieval Workshop. September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Enhancing Ground Truth Creation for IR Through Web-Based Collaborative Annotation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ornella Irrera</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefano Marchesin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gianmaria Silvello</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Engineering, University of Padova</institution>
          ,
          <addr-line>via Grednigo 6/B, 35131, Padua</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>0</volume>
      <fpage>3</fpage>
      <lpage>05</lpage>
      <abstract>
        <p>Ground truth creation plays a key role in Information Retrieval (IR), serving as the foundation for building test collections used to train and evaluate retrieval systems. However, this process is often time-consuming and demands considerable efort from human experts. Evaluation campaigns like TREC and CLEF highlight the scale of this challenge, requiring extensive manual annotation to ensure the accuracy and reliability of the data. To reduce this workload and support assessors more efectively, we present Doctron. Doctron is a collaborative, web-based, containerized platform designed to simplify ground truth creation in IR. It supports both text and image annotation, including features like entity tagging and linking, passages annotation, graded labeling, and object detection. The platform enables team collaboration through role-based permissions and integrates Inter Annotator Agreement (IAA) metrics to ensure annotation consistency and quality.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Manual annotation</kwd>
        <kwd>Ground truth creation</kwd>
        <kwd>Evaluation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Creating ground truth datasets for large corpora is fundamental to Information Retrieval (IR), as they
are the foundation for training, evaluating, and enhancing search systems. Manual annotation – where
human assessors label documents – remains the standard approach, playing a key role in ensuring
the reliability and robustness of test collections that contribute to the progress in IR. In the context of
large-scale evaluation initiatives like Text REtrieval Conference (TREC) and Conference and Labs of
the Evaluation Forum (CLEF), the creation of high-quality annotated corpora is crucial for enabling
researchers to benchmark models and foster progress in the field [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        However, generating such datasets is a demanding and resource-intensive process involving multiple
stakeholders – such as domain experts, annotators, and project coordinators – in a complex workflow.
This process encompasses defining guidelines, selecting and configuring tools, preprocessing data,
performing annotations, resolving inconsistencies, assessing quality, and making revisions. Due to its
complexity, this workflow often becomes a bottleneck in IR projects [
        <xref ref-type="bibr" rid="ref13">13, 17</xref>
        ].
      </p>
      <p>
        Choosing an appropriate annotation tool can streamline the process and ease assessors’ work. In
recent years, several reviews have assessed the efectiveness of annotation tools, compared their
functionalities, and provided guidance to researchers in selecting the most appropriate tool for their
specific needs [
        <xref ref-type="bibr" rid="ref13 ref14 ref2">2, 13, 14</xref>
        ]. Some annotation tools are specifically designed for certain domains, addressing
the unique demands of particular research contexts. A significant number of these are tailored for the
biomedical field [
        <xref ref-type="bibr" rid="ref1 ref4 ref7 ref8">1, 4, 7, 8, 18</xref>
        ], typically supporting tasks such as document classification, Named Entity
Recognition (NER) and Named Entity Recognition and Linking (NER+L), and relation annotation, often
with integration of domain-specific ontologies. Conversely, general-purpose tools [
        <xref ref-type="bibr" rid="ref12 ref15 ref3 ref5 ref9">3, 5, 9, 12, 15, 16, 19–
21</xref>
        ] ofer high customizability to accommodate a wide range of annotation tasks. Nevertheless, most
annotation tools do not specifically address the needs of IR. They often lack essential features like
relevance assessments for topic-document pairs or passage-level annotations, and are dificult to be
properly configured. As a result, choosing and adapting a tool for IR is often complex, making it
i
ii
iii
iv
v
vi
vii
viii
ix
x
xi
impractical for non-expert users.
      </p>
      <p>
        To address the limitations of existing annotation tools, we introduce Doctron [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a web-based,
open-source, Dockerized platform designed for collaborative document annotation, with a strong
focus on IR. It supports topic-based and graded relevance annotation, ofering advanced features
such as passage-level annotation, object detection, and support for both textual and image data. The
platform enables eficient teamwork through role-based Inter Annotator Agreement ( IAA) metrics
to assess annotation consistency. Integration with the ir_dataset library [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] allows easy use and
re-annotation of standard IR test collections. Doctron implements an intuitive, customizable interface
accessible even to non-technical users, setting it apart from more complex tools. The platform is
available as a cloud service at https://doctron.dei.unipd.it/ (currently accessible providing username:
demo and password: demo), and can also be installed as a Docker container1 for local deployment.
      </p>
      <p>
        Doctron [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] was presented as a resource paper at SIGIR 2025.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Doctron</title>
      <p>Doctron is a web-based, platform-independent annotation tool designed for document-topic pair
annotation. Distributed as a Docker container, it enables easy deployment and ensures privacy by
allowing local installation. Its architecture follows a three-tier design: a data layer with a PostgreSQL
database storing documents, topics, and annotations; a business logic layer implemented with Django
2, handling core functionalities such as processing requests and interacting with the database; and a
presentation layer developed with React.js 3, providing an interactive platform for annotators.</p>
      <p>In Doctron, there are three user types: annotators, reviewers, and administrators. Annotators add
annotations to documents. Reviewers, with the highest expertise, have full access to annotators’ work
and can update annotations to ensure quality and consistency. Administrators manage the collection,
oversee annotators and reviewers, set annotation guidelines, and configure templates and settings. They
can modify and update all annotations, including those of reviewers, and track progress.</p>
      <sec id="sec-2-1">
        <title>1https://github.com/meta-doc-dev/DocTron 2https://www.djangoproject.com/ 3https://react.dev/</title>
        <p>The user interface of Doctron is presented in Figure 1. Upon login, users select an annotation template,
and the system loads the relevant collections and the last opened topic-document pair. The main header
1 provides access to the Home, Collections, and Statistics pages. The document header 2 displays key
identifiers and ofers navigation and annotation reset options. A multifunctional left sidebar 3 enables
quick access to features like role switching, document/topic lists, settings, annotation downloads, and
tutorials. The left panel 4 summarizes the user’s annotations for the current pair, allowing edits and
comments. The main area presents the topic details 5 and the document to annotate 6 , with the
layout adapting to both text and image-based content.</p>
        <p>Annoation templates. Annotators can annotate document-topic pairs in Doctron using seven
annotation templates. These templates include (i) graded labeling, where annotators assign labels (e.g.,
relevance) with a range of values (e.g., 0 to 3) to a document (textual or image) regarding a specific topic;
(ii) passage annotation, where annotators select specific passages within a textual document and assign
a graded label indicating its relevance to a topic; (iii) object annotation, which involves identifying and
labeling objects within images by selecting the object’s perimeter and assigning one or more graded
labels to the selected area; (iv) NER, where annotators identify mentions of entities in a text and label
them with predefined tags; (v) NER+L, which adds the step of linking identified entities to specific
entries in external knowledge bases like Wikipedia or Wikidata; (vi) relationship annotation, where
annotators identify relationships between a subject, predicate, and object, which may be represented by
ontological concepts, tags, or mentions; and (vii) fact annotation, which involves annotating factual
triples (subject, predicate, object), where all components are ontological concepts or tags, and none are
textual mentions from the document.</p>
        <p>
          Collections and Customization. A collection in Doctron includes a group of annotators (at least
one), a set of documents (either text or images), a set of topics (either text or images), and is associated
with an annotation template. Each user in Doctron has the ability to annotate multiple document
collections, which exist independently from one another. Documents in Doctron are schema-free and
can be uploaded in several formats including: JSON, CSV, TXT and PDF (thanks to the integration
with GROBID [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]) for textual documents, and JPG and PNG for image documents. Doctron supports
integration with ir_datasets [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and the PubMed REST API, allowing users to import documents
and topics from URLs or external sources with full customization. Annotation templates are fully
configurable, enabling the definition of custom tags, ontological concepts, and labels. Collections can go
through multiple annotation rounds and be set in either collaborative or competitive mode: the former
allows annotators to view other annotators’ work, the latter instead hides the identities of other users.
This ensures flexibility, as well as control over the annotation process.
        </p>
        <p>Collection Statistics. Doctron provides two levels of statistical reporting: individual and global.
Individual statistics give each annotator an overview of their work, including the number of
annotated and unannotated documents per topic, as well as annotation-specific metrics depending on the
template used (e.g., graded labels, identified passages, tagged entities, or extracted relations and facts).
Administrators and reviewers can also monitor these statistics to track annotator activity and spot
inconsistencies or documents that need to be reannotated. Global statistics aggregate this information
across all annotators in a collection, also listing which annotators worked on each document. To
evaluate annotation quality and consistency, Doctron includes IAA metrics – Cohen’s Kappa, Fleiss’s
Kappa, and Krippendorf’s Alpha – which are accessible by administrators and reviewers. These metrics
provide insight into annotation reliability by quantifying the degree of agreement among annotators.
Together, these features ofer a comprehensive view of annotation performance and ensure quality
control throughout the annotation process.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Evaluation</title>
      <p>
        We performed a qualitative evaluation where we compared Doctron functionalities to those provided by
other 9 tools: MetaTron [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], Doctag [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Doccano, LabelStudio, TeamTat [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], INCEpTION [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], brat [19],
TagTog [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], POTATO [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. We used three sets of criteria: technical, data-related, and functionality-based.
These criteria cover aspects such as ease of use, data integration, annotation capabilities, and support for
information retrieval workflows. As shown in Table 1, Doctron emerges as the most complete solution,
fulfilling 24 out of 25 criteria. It is the only tool that supports the combination of topic-document
annotation, passage-level annotation, and graded labeling, while also ofering TREC-like export and
IR-specific dataset integration. Although it lacks built-in automatic predictions, Doctron supports easy
integration of user-defined models, ensuring flexibility and customizability.
      </p>
      <p>
        We performed a quantitative analysis to assess the eficiency of five annotation tools – Doctron,
Doctag, INCEpTION, Doccano, and LabelStudio – by measuring the number of clicks and time required
for two tasks: (i) creating and configuring a document collection, and (ii) annotating 15 documents
using three typical IR-related templates: multilabel classification, passage annotation, and NER. While
the annotation phase yielded comparable results across tools – due to similar implementations of task
templates – significant diferences emerged in the setup phase. Doctron and Doctag proved more
eficient for creating collections, with fewer steps and a simpler setup—Doctron especially stood out
for its ease of use. In contrast, general-purpose tools like INCEpTION and LabelStudio required more
complex configurations to support IR-specific workflows. These findings show that general tools often
demand extra efort for IR tasks, while Doctron ofers a more integrated, ready-to-use solution. More
details can be found in the original Doctron paper [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Final remarks</title>
      <p>We presented Doctron, a portable, web-based annotation tool designed for the IR. It supports
collaborative workflows with role-based access, ofers a wide range of annotation templates (including graded
labeling, passage annotation, NER and NER+L), and integrates with PubMed and ir_dataset ensuring
easy and customizable collection setup. Doctron includes built-in IAA metrics and detailed statistics
to ensure annotation quality. Through qualitative and quantitative evaluations, we demonstrated
that Doctron stands out for its flexibility, rich feature set, and ease of use. As future work, we plan
to integrate automatic annotation support using pre-trained models for NER/NER+L and Language
Language Models (LLMs) for graded and passage-level annotation.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work is supported by the HEREDITARY Project, as part of the European Union’s Horizon Europe
research and innovation programme under Grant Agreement No GA 101137074.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>During the preparation of this work, the author did not use any AI tool.</title>
        <p>Proceedings of the The 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP
2022 - System Demonstrations, Abu Dhabi, UAE, December 7-11, 2022, pages 327–337. Association
for Computational Linguistics, 2022.
[16] T. Perry. Lighttag: Text annotation platform. In Heike Adel and Shuming Shi, editors, Proceedings of
the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations,
EMNLP 2021, Online and Punta Cana, Dominican Republic, 7-11 November, 2021, pages 20–27.</p>
        <p>Association for Computational Linguistics, 2021.
[17] J. Pustejovsky and A. Stubbs. Natural Language Annotation for Machine Learning: A guide to
corpus-building for applications. " O’Reilly Media, Inc.", 2012.
[18] D. Salgado, M. Krallinger, M. Depaule, E. Drula, A. V. Tendulkar, F. Leitner, A. Valencia, and
C. Marcelle. Myminer: a web application for computer-assisted biocuration and text annotation.</p>
        <p>Bioinformatics, 28(17):2285–2287, 2012.
[19] P. Stenetorp, S. Pyysalo, G. Topic, T. Ohta, S. Ananiadou, and J. Tsujii. brat: a web-based tool
for nlp-assisted text annotation. In EACL 2012, 13th Conference of the European Chapter of the
Association for Computational Linguistics, Avignon, France, April 23-27, 2012, pages 102–107. The
Association for Computer Linguistics, 2012.
[20] D. Wilby, T. Armakharm, I. Roberts, X. Song, and K. Bontcheva. GATE teamware 2: An open-source
tool for collaborative document classification annotation. In Proceedings of the 17th Conference
of the European Chapter of the Association for Computational Linguistics. EACL 2023 - System
Demonstrations, Dubrovnik, Croatia, May 2-4, 2023, pages 145–151. Association for Computational
Linguistics, 2023.
[21] J. Yang, Y. Zhang, L. Li, and X. Li. YEDDA: A lightweight collaborative text span annotation tool.</p>
        <p>In Fei Liu and Thamar Solorio, editors, Proceedings of ACL 2018, Melbourne, Australia, July 15-20,
2018, System Demonstrations, pages 31–36. Association for Computational Linguistics, 2018.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.M.</given-names>
            <surname>Cejuela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>McQuilton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ponting</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Marygold</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Stefancsik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. H.</given-names>
            <surname>Millburn</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Rost</surname>
          </string-name>
          .
          <article-title>tagtog: interactive and text-mining-assisted annotation of gene mentions in PLOS full-text articles</article-title>
          .
          <source>Database J. Biol. Databases Curation</source>
          ,
          <year>2014</year>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Colucci Cante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. D</given-names>
            <surname>'Angelo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. Di</given-names>
            <surname>Martino</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Graziano</surname>
          </string-name>
          .
          <article-title>Text annotation tools: A comprehensive review and comparative analysis</article-title>
          .
          <source>In International Conference on Complex, Intelligent, and Software Intensive Systems</source>
          , pages
          <fpage>353</fpage>
          -
          <lpage>362</lpage>
          . Springer,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B.</given-names>
            <surname>Di Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Marulli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Graziano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Lupi. Prettytags</surname>
          </string-name>
          :
          <article-title>An open-source tool for easy and customizable textual multilevel semantic annotations</article-title>
          .
          <source>In Complex, Intelligent and Software Intensive Systems - Proceedings of the 15th International Conference on Complex, Intelligent and Software Intensive Systems (CISIS-2021)</source>
          , Asan, Korea,
          <fpage>1</fpage>
          -
          <issue>3</issue>
          <year>July 2021</year>
          , volume
          <volume>278</volume>
          <source>of Lecture Notes in Networks and Systems</source>
          , pages
          <fpage>636</fpage>
          -
          <lpage>645</lpage>
          . Springer,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Giachelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Irrera</surname>
          </string-name>
          , and Silvello G.
          <article-title>Medtag: a portable and customizable annotation tool for biomedical documents</article-title>
          .
          <source>BMC Medical Informatics Decis. Mak.</source>
          ,
          <volume>21</volume>
          (
          <issue>1</issue>
          ):
          <fpage>352</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Giachelle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Irrera</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Silvello. Doctag</surname>
          </string-name>
          :
          <article-title>A customizable annotation tool for ground truth creation</article-title>
          .
          <source>In Advances in Information Retrieval - 44th European Conference on IR Research</source>
          , ECIR
          <year>2022</year>
          , Stavanger, Norway,
          <source>April 10-14</source>
          ,
          <year>2022</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , volume
          <volume>13186</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>288</fpage>
          -
          <lpage>293</lpage>
          . Springer,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>O.</given-names>
            <surname>Irrera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchesin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>SHami</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Silvello. Doctron</surname>
          </string-name>
          :
          <article-title>A web-based collaborative annotation tool for ground truth creation in ir</article-title>
          .
          <source>In Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '25)</source>
          ,
          <source>July 13-18</source>
          ,
          <year>2025</year>
          , Padua, Italy,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>O.</given-names>
            <surname>Irrera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchesin</surname>
          </string-name>
          , and 1a G.
          <article-title>Silvello. Metatron: advancing biomedical annotation empowering relation annotation and collaboration</article-title>
          . volume
          <volume>25</volume>
          , page 112,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R. Islamaj</given-names>
            <surname>Dogan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kwon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lu</surname>
          </string-name>
          .
          <article-title>Teamtat: a collaborative text annotation tool</article-title>
          . volume
          <volume>48</volume>
          , pages
          <fpage>W5</fpage>
          -
          <lpage>W11</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Klie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bugert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Boullosa</surname>
          </string-name>
          , R. E. de Castilho,
          <string-name>
            <surname>and I. Gurevych.</surname>
          </string-name>
          <article-title>The inception platform: Machine-assisted and knowledge-oriented interactive annotation</article-title>
          .
          <source>In COLING 2018, The 27th International Conference on Computational Linguistics: System Demonstrations, Santa Fe, New Mexico, August 20-26</source>
          ,
          <year>2018</year>
          , pages
          <fpage>5</fpage>
          -
          <lpage>9</lpage>
          . Association for Computational Linguistics,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lopez</surname>
          </string-name>
          . GROBID:
          <article-title>combining automatic bibliographic data recognition and term extraction for scholarship publications</article-title>
          .
          <source>In Research and Advanced Technology for Digital Libraries, 13th European Conference, ECDL</source>
          <year>2009</year>
          , Corfu, Greece,
          <source>September 27 - October 2</source>
          ,
          <year>2009</year>
          . Proceedings, volume
          <volume>5714</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>473</fpage>
          -
          <lpage>474</lpage>
          . Springer,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Downey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N</given-names>
            <surname>Goharian</surname>
          </string-name>
          .
          <article-title>Simplified data wrangling with ir_datasets</article-title>
          .
          <source>In SIGIR '21: The 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Virtual Event, Canada,
          <source>July 11-15</source>
          ,
          <year>2021</year>
          , pages
          <fpage>2429</fpage>
          -
          <lpage>2436</lpage>
          . ACM,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>H.</given-names>
            <surname>Nakayama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kubo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Taniguchi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>X.</given-names>
            <surname>Liang</surname>
          </string-name>
          . Doccano:
          <article-title>Text annotation tool for human</article-title>
          . Software available from https://github. com/doccano/doccano, 34,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Neves</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Seva</surname>
          </string-name>
          .
          <article-title>An extensive review of tools for manual annotation of documents</article-title>
          . Briefings in Bioinformatics,
          <volume>22</volume>
          (
          <issue>1</issue>
          ):
          <fpage>146</fpage>
          -
          <lpage>163</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M. L.</given-names>
            <surname>Neves</surname>
          </string-name>
          and
          <string-name>
            <given-names>U.</given-names>
            <surname>Leser</surname>
          </string-name>
          .
          <article-title>A survey on annotation tools for the biomedical literature</article-title>
          .
          <source>Briefings in Bioinformatics</source>
          ,
          <volume>15</volume>
          (
          <issue>2</issue>
          ):
          <fpage>327</fpage>
          -
          <lpage>340</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ananthasubramaniam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dedeloudis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sargent</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Jurgens</surname>
          </string-name>
          .
          <article-title>POTATO: the portable text annotation tool</article-title>
          . In Wanxiang Che and Ekaterina Shutova, editors,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>