<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of ImageCLEFmedical 2024 - Medical Visual Question Answering for Gastrointestinal Tract</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Steven Hicks</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Storås</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pål Halvorsen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Riegler</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vajira Thambawita</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>OsloMet - Oslo Metropolitan University</institution>
          ,
          <addr-line>Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>SimulaMet - Simula Metropolitan Center for Digital Engineering</institution>
          ,
          <addr-line>Oslo</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper provides details on the second edition of the Medical Visual Question Answering for the Gastrointestinal Tract (MedVQA-GI) challenge, which took place during ImageCLEF 2024. This year, we changed the task from visual question answering to the application of text-to-image models for the creation of synthetic medical images. There were two sub-tasks in this challenge. The first sub-task involved using prompts to generate realistic looking images from the gastrointestinal tract. The second sub-task focused on the technical aspects involved in the implementation of these models, and optimizing the prompts to generate realistic-looking images using a low number of tokens. Despite considerable interest in the task, the rate of submissions remained low, suggesting that participants may have encountered barriers or found the task too complex to complete.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Machine learning</kwd>
        <kwd>medical ai</kwd>
        <kwd>endoscopy</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The second edition of the Medical Visual Question Answering for the Gastrointestinal Tract
(MedVQAGI) challenge at ImageCLEF [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] introduces a new goal that focuses on the use of generative models of
text-to-image in medical diagnosis. This combines natural language processing and image generation
to potentially improve diagnostic processes in healthcare by providing more comprehensive datasets
that can be used for training machine learning models. In contrast to last year’s focus on a Visual
Question Answering (VQA) task that required retrieving images or masks from user questions, this
year’s overall goal was to use generative models to create synthetic medical images from textual inputs.
Participants were tasked with generating synthetic images using existing generative models developed
using a dataset derived from last year’s MedVQA-GI challenge [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Machine learning has been a common method used to identify lesions in gastrointestinal (GI) images [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref3 ref4 ref5 ref6 ref7 ref8 ref9">3,
4, 5, 6, 7, 8, 9, 10, 11, 12, 13</xref>
        ]. Traditionally, the emphasis in GI analysis has been on disease detection
from images or videos, focusing mostly on polyp detection [14, 15, 16, 17, 18, 19, 20]. Several challenges
have demonstrated consistent advancements in this field, including some challenges we have organized
in the past [21, 22, 23, 24, 25]. However, there has been a growing interest in extending the capabilities
of image analysis GI through the generation of synthetic images [26, 27]. This new focus aims to develop
models that generate realistic GI images that can be used in-place of real data. Such synthetic images
can be used to train medical professionals, refine diagnostic algorithms without the privacy concerns of
real patient data, and improve the interpretability and reliability of AI systems in clinical settings. To
this end, this year’s MedVQA-GI focuses on synthetic GI image generation. The dataset and the scripts
used to verify and evaluate submissions are available in our public GitHub repository1.
      </p>
      <p>The remainder of this paper is organized as follows. First, we start with an explanation of the creation
of the dataset, looking at how the data was collected and organized. Then, we discuss the specific
sub-tasks involved in the MedVQA-GI challenge and the evaluation methods used. Finally, we present
statistics on the participants and the results of the submitted runs.</p>
      <p>Generate an image containing a
polyp.</p>
      <p>Generate an image containing
text.</p>
      <p>Generate an image containing
oesophagitis.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Dataset</title>
      <p>The dataset used for this challenge is based on data developed for last year’s challenge, which was based
on the HyperKvasir dataset [28] and the Kvasir-Instrument dataset [29] datasets. Participants were
provided with a dataset consisting of 2, 000 image and text pairs, which was organized in a directory
containing the images and CSV files with prompts and connections to the image filenames. Example
images and corresponding prompts can be seen in Figure 1.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Task Description and Evaluation</title>
      <p>This year, participants could participate in two sub-tasks: Image Synthesis and Optimal Prompt
Generation. Participants could submit to either sub-task and were not limited to the number of submissions.</p>
      <sec id="sec-3-1">
        <title>3.1. Sub-task 1: Image Synthesis</title>
        <p>The first sub-task, Image Synthesis, involves using text-to-image generative models to construct a
comprehensive dataset of medical images from textual descriptions. This sub-task requires participants
to create accurate visual representations of various medical conditions described solely in text. For
example, with a description such as "An early-stage colorectal polyp," participants must generate an
image that precisely reflects the given text. Participants could use the development dataset, described
in Section 2, to develop their models.</p>
        <p>For the submission, each participant received a list of 5, 000 prompts. They were required to create
synthetic images based on these prompts and submit them to the organizers by email. Each submitted
image file was named according to the prompt’s index number from the list. The quality of the synthetic
images was assessed using two metrics: the Inception Score (IS) [30] and the Fréchet Inception Distance
(FID) [31]. These metrics evaluated how the synthetic images were compared with three distinct testing
datasets. The first data set consisted of images from the previous year’s MedVQA-GI challenge. The
second dataset was GastroVision [32], which is a newly released open-source collection that includes
8, 000 images obtained from various medical centers. The final data set used for the evaluation was a
combination of the first two datasets.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Sub-task 2: Optimal Prompt Generation</title>
        <p>The second sub-task, Optimal Prompt Generation, focuses on participants creating their own prompts
to generate images that meet specific medical imaging requirements. This sub-task asked participants
to tailor their prompt generation skills to produce images that accurately match a set of predefined
categories. These categories are designed to test the model’s ability to generate precise and clinically
relevant images based on the prompts. Participants had to devise prompts for:
• A prompt that generates an image containing n polyps.
• A prompt that generates a polyp in a specific region of the image.
• A prompt that generates a polyp of a specific type and size.
• A prompt that generates an image containing no findings from either the esophagus or large
bowel.
• A prompt that generates an image containing one of the following instruments: biopsy forceps,
metal clip, and tube.
• A prompt that generates an image containing one of the following anatomical landmarks: Z-line,</p>
        <p>Pylorus, Cecum.</p>
        <p>For evaluation, the efectiveness of each prompt was evaluated not only on the accuracy of the image
it produced but also on the conciseness of the prompt itself. Shorter and more precise prompts were
preferred, as they are more beneficial in clinical settings where clarity and eficiency are necessary.
Additionally, the generated images were subjected to the same quantitative evaluation metrics as in
sub-task 1, IS and FID, to ensure consistency in assessing the quality of images across diferent tasks.
This dual approach, which combined qualitative assessment of prompt efectiveness with quantitative
image quality metrics, provided a comprehensive assessment of participants’ proficiency in generating
both relevant prompts and high-quality synthetic images.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Participation and Results</title>
      <p>This section provides an overview of participation in the challenge and discusses the results submitted
by those who completed it.</p>
      <sec id="sec-4-1">
        <title>4.1. Participation</title>
        <p>In total, 22 teams signed up for the task, 2 teams submitted runs, and 2 teams submitted working notes
papers [33, 34]. Table 1 shows an overview of the participants and the number of submissions to each
sub-task alongside the number of participants from last year’s challenge. This year experienced a
noticeable decline in participation compared to last year’s challenge. One reason for this may be the
complexity of the task and the hardware and model requirements. Furthermore, future editions could
benefit from enhanced outreach and support mechanisms, such as tutorials or "getting started" scripts,
to broaden participant engagement and lower the entry barriers.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results</title>
        <p>4.2.1. MMCP Team
A total of six runs were submitted to sub-task 1, and no runs were submitted to sub-task 2. This section
gives an overview of the results of each run and briefly discusses the approach submitted by each
participant. The results can be seen in Table 2.</p>
        <p>Team MMCP’s approach was based on two methods: fine-tuning Kandinsky models and implementing
a Medical Synthesis with Difusion Model (MSDM). They fine-tune pre-trained Kandinsky models to
generate images from text prompts. In addition, they experimented with MSDM, showing improved
results over Kandinsky-based models. Example images of each sumbission can be seen in Figure 2. For
more information on their approach, please read their working notes paper [33].</p>
        <p>Generate an image Generate an image
not containing text. containing the z-line.</p>
        <p>Generate an image
containing tube.</p>
        <p>Generate an image
containing a polyp.</p>
        <p>Generate an image
from a colonoscopy.</p>
        <p>Generate an image
from a gastroscopy.
Team 2 used a diferent approach for each submission. The first approach did not generate synthetic
images, rather it retrieved images that closely related to the input prompt. To do this, they used a
Connecting text and images (CLIP) model. The second submission used a fine-tuned stable difusion
model that generated synthetic images. The third submission used a fine-tuned Low-Rank Adaptation
of Large Language Models (LoRA) model to generate images. This method uses LoRA to modify
preexisting stable difusion model to enable the production of high-quality images that closely align with
the input specifications. Example images of each submission can be seen in Figure 3.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Discussion</title>
        <p>The challenge results highlight several important insights and areas for further exploration. Firstly,
the performance across the two teams and runs varied. This variability underscores the complexity of
creating high-quality medical images. However, we found that the quality of the images did not always
correspond to the scores provided by the quantitative metrics, suggesting that we need more robust
synthetic image quality metrics specifically for medical images and their applications.</p>
        <p>Another notable finding was that there was some confusion surrounding generation of synthetic
images. One team submitted a run that retrieved "real" images that corresponded to the submitted
prompt. This deviated from the intended goal, as the main point was to generate synthetic images. This
highlights the need for clearer communication of the challenge requirements.</p>
        <p>Furthermore, reduced participation compared to last year indicates possible entry barriers that
may include the complexity of tasks or a lack of foundational resources for newcomers. Addressing
these barriers could involve providing more comprehensive datasets, detailed examples of successful
implementations, and potentially simplifying the challenge structure to attract a broader range of
participants.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Outlook</title>
      <p>This paper discussed the second edition of the MedVQA-GI challenge, which took place at ImageCLEF
in 2024. The challenge consisted of two sub-tasks centered on the generation of synthetic images in the
gastrointestinal tract. In the future, we plan on making a more robust task with more resources to get
started. Furthermore, we also want to merge the tasks from the first year with this year’s challenge to
keep the task more consistent.
doi:https://doi.org/10.1109/CBMS.2016.63.
[13] M. Riegler, K. Pogorelov, P. Halvorsen, T. de Lange, C. Griwodz, P. T. Schmidt, S. L. Eskeland,
D. Johansen, EIR - eficient computer aided diagnosis framework for gastrointestinal endoscopies,
in: Proceedings of the IEEE International Workshop on Content-Based Multimedia Indexing
(CBMI), 2016, pp. 1–6. doi:https://doi.org/10.1109/CBMI.2016.7500257.
[14] Y. Wang, W. Tavanapong, J. Wong, J. H. Oh, P. C. De Groen, Polyp-alert: Near real-time
feedback during colonoscopy, Computer Methods and Programs in Biomedicine 120 (2015) 164–179.
doi:https://doi.org/10.1016/j.cmpb.2015.04.002.
[15] D. Jha, P. H. Smedsrud, M. A. Riegler, D. Johansen, T. De Lange, P. Halvorsen, H. D. Johansen,
Resunet++: An advanced architecture for medical image segmentation, in: Proceedings of the
International Symposium on Multimedia (ISM), 2019, pp. 225–230. doi:https://doi.org/10.
1109/ISM46123.2019.00049.
[16] J. Bernal, A. Histace, M. Masana, Q. Angermann, C. Sánchez-Montes, C. Rodriguez, M. Hammami,
A. Garcia-Rodriguez, H. Córdova, O. Romain, G. Fernández-Esparrach, X. Dray, J. Sanchez, Polyp
detection benchmark in colonoscopy videos using gtcreator: A novel fully configurable tool for
easy and fast annotation of image databases, in: Proceedings of Computer Assisted Radiology and
Surgery (CARS), 2018. doi:https://hal.archives-ouvertes.fr/hal-01846141.
[17] Y. Guo, J. Bernal, B. J Matuszewski, Polyp segmentation with fully convolutional deep neural
networks—extended evaluation study, Journal of Imaging 6 (2020) 69.
[18] M. Min, S. Su, W. He, Y. Bi, Z. Ma, Y. Liu, Computer-aided diagnosis of colorectal polyps using
linked color imaging colonoscopy to predict histology, Scientific reports 9 (2019) 2881. doi: https:
//doi.org/10.1038/s41598-019-39416-7.
[19] N. M. Ghatwary, X. Ye, M. Zolgharni, Esophageal abnormality detection using densenet based
faster r-cnn with gabor features, IEEE Access 7 (2019) 84374–84385. doi:https://doi.org/10.
1109/ACCESS.2019.2925585.
[20] S. Shah, N. Park, N. E. H. Chehade, A. Chahine, M. Monachese, A. Tiritilli, Z. Moosvi, R. Ortizo,
J. Samarasena, Efect of computer-aided colonoscopy on adenoma miss rates and polyp detection:
a systematic review and meta-analysis, Journal of Gastroenterology and Hepatology 38 (2023)
162–176.
[21] S. Hicks, M. Riegler, P. Smedsrud, T. B. Haugen, K. R. Randel, K. Pogorelov, H. K. Stensland, D.-T.</p>
      <p>Dang-Nguyen, M. Lux, A. Petlund, T. de Lange, P. T. Schmidt, P. Halvorsen, Acm multimedia
biomedia 2019 grand challenge overview, in: Proceedings of the ACM International Conference
on Multimedia (ACM MM), 2019, pp. 2563–2567. doi:https://doi.org/10.1145/3343031.
3356058.
[22] K. Pogorelov, M. Riegler, P. Halvorsen, S. A. Hicks, K. R. Randel, D.-T. Dang-Nguyen, M. Lux,
O. Ostroukhova, T. De Lange, Medico multimedia task at mediaeval 2018, in: Proceeding of the
MediaEval Benchmarking Initiative for Multimedia Evaluation Workshop (MediaEval), 2018.
[23] M. Riegler, K. Pogorelov, P. Halvorsen, K. Randel, S. Eskeland, D.-T. Dang-Nguyen, M. Lux, C.
Griwodz, C. Spampinato, T. de Lange, Multimedia for medicine: the medico task at mediaeval 2017,
in: Proceeding of the MediaEval Benchmarking Initiative for Multimedia Evaluation Workshop
(MediaEval), 2017.
[24] J. Bernal, H. Aymeric, Miccai endoscopic vision challenge polyp detection and segmentation,
https://endovissub2017-giana.grand-challenge.org/home/, 2017. Accessed: 2017-12-11.
[25] S. Hicks, M. Riegler, P. Smedsrud, T. B. Haugen, K. R. Randel, K. Pogorelov, H. K. Stensland, D.-T.</p>
      <p>Dang-Nguyen, M. Lux, A. Petlund, T. de Lange, P. T. Schmidt, P. Halvorsen, Acm multimedia
biomedia 2019 grand challenge overview, in: Proceedings of the 27th ACM International
Conference on Multimedia, MM ’19, Association for Computing Machinery, New York, NY, USA, 2019, p.
2563–2567. URL: https://doi.org/10.1145/3343031.3356058. doi:10.1145/3343031.3356058.
[26] V. Thambawita, P. Salehi, S. A. Sheshkal, S. A. Hicks, H. L. Hammer, S. Parasa, T. d. Lange,
P. Halvorsen, M. A. Riegler, Singan-seg: Synthetic training data generation for medical image
segmentation, PLOS ONE 17 (2022) 1–24. URL: https://doi.org/10.1371/journal.pone.0267976.
doi:10.1371/journal.pone.0267976.
[27] D. Yoon, H.-J. Kong, B. S. Kim, W. S. Cho, J. C. Lee, M. Cho, M. H. Lim, S. Y. Yang, S. H. Lim,
J. Lee, J. H. Song, G. E. Chung, J. M. Choi, H. Y. Kang, J. H. Bae, S. Kim, Colonoscopic image
synthesis with generative adversarial network for enhanced detection of sessile serrated lesions
using convolutional neural network, Sci Rep 12 (2022) 261.
[28] H. Borgli, V. Thambawita, P. H. Smedsrud, S. Hicks, D. Jha, S. L. Eskeland, K. R. Randel, K. Pogorelov,
M. Lux, D. T. D. Nguyen, et al., Hyperkvasir, a comprehensive multi-class image and video dataset
for gastrointestinal endoscopy, Scientific data 7 (2020). doi: 10.1038/s41597-020-00622-y.
[29] D. Jha, S. Ali, K. Emanuelsen, S. A. Hicks, V. Thambawita, E. Garcia-Ceja, M. A. Riegler, T. de Lange,
P. T. Schmidt, H. D. Johansen, D. Johansen, P. Halvorsen, Kvasir-instrument: Diagnostic and
therapeutic tool segmentation dataset in gastrointestinal endoscopy, in: Proceeedings of the
International COnference on MultiMedia Modeling (MMM), 2021, pp. 218–229. doi:10.1007/
978-3-030-67835-7_19.
[30] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, X. Chen, Improved techniques for
training gans, Advances in neural information processing systems 29 (2016).
[31] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, S. Hochreiter, Gans trained by a two time-scale
update rule converge to a local nash equilibrium, Advances in neural information processing
systems 30 (2017).
[32] D. Jha, V. Sharma, N. Dasu, N. K. Tomar, S. Hicks, M. Bhuyan, P. K. Das, M. A. Riegler, P. Halvorsen,
T. de Lange, U. Bagci, Gastrovision: A multi-class endoscopy image dataset for computer aided
gastrointestinal disease detection, in: ICML Workshop on Machine Learning for Multimodal
Healthcare Data (ML4MHD 2023), 2023.
[33] M. Chaychuk, Mmcp team at imageclefmed 2024 task on image synthesis: Difusion models for
text-to-image generation of colonoscopy images, in: CLEF2024 Working Notes, CEUR Workshop
Proceedings, CEUR-WS.org, Grenoble, France, 2024.
[34] E.-P. Oluwafemi Ojonugwa, M. Rahman, F. Khalifa, Advancing ai-powered medical image synthesis:
Insights from medvqa-gi challenge using clip, fine-tuned stable difusion, and dream-booth + lora,
in: CLEF2024 Working Notes, CEUR Workshop Proceedings, CEUR-WS.org, Grenoble, France,
2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.-M. Drăgulinescu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Rückert</surname>
            ,
            <given-names>A. Ben</given-names>
          </string-name>
          <string-name>
            <surname>Abacha</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            , L. Bloch,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Brüngel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Idrissi-Yaghir</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Schäfer</surname>
            ,
            <given-names>C. S.</given-names>
          </string-name>
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>T. M.</given-names>
          </string-name>
          <string-name>
            <surname>Pakull</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Damm</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Bracke</surname>
            ,
            <given-names>C. M.</given-names>
          </string-name>
          <string-name>
            <surname>Friedrich</surname>
            ,
            <given-names>A.-G.</given-names>
          </string-name>
          <string-name>
            <surname>Andrei</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Prokopchuk</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Karpenka</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radzhabov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Kovalev</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macaire</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Schwab</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Lecouteux</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Esperança-Rodier</surname>
            , W.-W. Yim,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Fu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Yetisgen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          <string-name>
            <surname>Hicks</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Thambawita</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Storås</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Halvorsen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Heinrich</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kiesel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
          </string-name>
          , Overview of ImageCLEF 2024:
          <article-title>Multimedia retrieval in medical, socialmedia and recommender systems applications</article-title>
          , in: CLEF2024 Working Notes, CEUR Workshop Proceedings, CEUR-WS.org, Grenoble, France,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Hicks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Storås</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          , T. de Lange,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          ,
          <article-title>Overview of imageclefmedical 2023 - medical visual question answering for gastrointestinal tract</article-title>
          ,
          <source>in: CLEF2023 Working Notes, CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Thessaloniki, Greece,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Spadaccini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Iannone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Maselli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jovani</surname>
          </string-name>
          , V. T. Chandrasekar, G. Antonelli,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Areia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dinis-Ribeiro</surname>
          </string-name>
          , et al.,
          <article-title>Performance of artificial intelligence in colonoscopy for adenoma and polyp detection: a systematic review and meta-analysis</article-title>
          ,
          <source>Gastrointestinal endoscopy 93</source>
          (
          <year>2021</year>
          )
          <fpage>77</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Alammari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Oh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Tavanapong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. C. De Groen</surname>
          </string-name>
          ,
          <article-title>Classification of ulcerative colitis severity in colonoscopy videos using cnn</article-title>
          ,
          <source>in: Proceedings of the ACM International Conference on Information Management and Engineering (ACM ICIME)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>139</fpage>
          -
          <lpage>144</lpage>
          . doi:https://doi.org/10.1145/3149572.3149613.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Linder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Turkki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nordling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Kovanen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Verrill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Walliander</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lundin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Haglund</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lundin</surname>
          </string-name>
          ,
          <article-title>Deep learning based tissue analysis predicts outcome in colorectal cancer</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>8</volume>
          (
          <year>2018</year>
          )
          <article-title>3395</article-title>
          . URL: http://dx.doi.org/10.1038/s41598-018-21758-3. doi: https: //doi.org/10.1038/s41598-018-21758-3.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mori</surname>
          </string-name>
          , S.-e. Kudo,
          <string-name>
            <given-names>M.</given-names>
            <surname>Misawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Saito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ikematsu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hotta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ohtsuka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Urushibara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kataoka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ogawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Maeda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Takeda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nakamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ichimasa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kudo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Hayashi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wakamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ishida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Inoue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Itoh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Oda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Mori</surname>
          </string-name>
          ,
          <article-title>Real-Time Use of Artificial Intelligence in Identification of Diminutive Polyps During Colonoscopy: A Prospective Study</article-title>
          ,
          <source>Annals of Internal Medicine</source>
          <volume>169</volume>
          (
          <year>2018</year>
          )
          <fpage>357</fpage>
          -
          <lpage>366</lpage>
          . doi:https://doi.org/10.7326/M18-0249.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pogorelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. L.</given-names>
            <surname>Eskeland</surname>
          </string-name>
          , T. de Lange,
          <string-name>
            <given-names>C.</given-names>
            <surname>Griwodz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Randel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. K.</given-names>
            <surname>Stensland</surname>
          </string-name>
          , D.-T. DangNguyen,
          <string-name>
            <given-names>C.</given-names>
            <surname>Spampinato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <article-title>A holistic multimedia system for gastrointestinal tract disease detection</article-title>
          ,
          <source>in: Proceedings of the ACM on Multimedia Systems Conference (MMSYS)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>112</fpage>
          -
          <lpage>123</lpage>
          . doi:https://doi.org/10.1145/3193740.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Histace</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Romain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Dray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Granado</surname>
          </string-name>
          ,
          <article-title>Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer</article-title>
          ,
          <source>International Journal of Computer Assisted Radiology and Surgery</source>
          <volume>9</volume>
          (
          <year>2014</year>
          )
          <fpage>283</fpage>
          -
          <lpage>293</lpage>
          . doi:https://doi.org/10.1007/s11548-013-0926-3.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V. L.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Hammer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. D.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <article-title>An extensive study on cross-dataset bias and evaluation metrics interpretation for machine learning applied to gastrointestinal tract abnormality classification</article-title>
          ,
          <source>ACM Transactions on Computing for Healthcare</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <article-title>Doubleu-net: A deep convolutional neural network for medical image segmentation</article-title>
          ,
          <source>in: Proceeding of the International Symposium on Computer Based Medical Systems (CBMS)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Angermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bernal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sánchez-Montes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hammami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Fernández-Esparrach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Dray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Romain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. J.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Histace</surname>
          </string-name>
          ,
          <article-title>Towards real-time polyp detection in colonoscopy videos: Adapting still frame-based methodologies for video sequences analysis</article-title>
          ,
          <source>in: Proceedings of Computer Assisted and Robotic Endoscopy and Clinical Image-Based Procedures (CARE CLIP)</source>
          , volume
          <volume>10550</volume>
          , Springer,
          <year>2017</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Pogorelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. T.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Griwodz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Johansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. L.</given-names>
            <surname>Eskeland</surname>
          </string-name>
          , T. de Lange,
          <article-title>Gpu-accelerated real-time gastrointestinal diseases detection</article-title>
          ,
          <source>in: Proceedings of the International Symposium on Computer-Based Medical Systems (CBMS)„ IEEE</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>185</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>