<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Run-time verification of robot-assisted surgery using visual input</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kristina Gogoladze</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Romy van Jaarsveld</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Natasha Alechina</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ronald de Jong</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yasmina Al Khalil</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gino Kuiper</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Brian Logan</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jelle P. Ruurda</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Open Universiteit</institution>
          ,
          <addr-line>Heerlen</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Technical University Eindhoven</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universiteit Utrecht</institution>
          ,
          <addr-line>Utrecht</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University Medical Centre Utrecht</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>University of Aberdeen</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>We present a preliminary study of the feasibility of using run-time verification to provide intraoperative warnings during robot-assisted surgery for cancer. Robot-assisted surgery ofers several advantages over traditional surgery. However, it can be challenging for surgeons, particularly if they lack experience in the procedure being performed. We report the results of a survey to elicit from surgeons which clinically relevant properties should be monitored, and present a proof-of-concept evaluation of the feasibility of monitoring using the video image available to the surgeon as input, which suggests that real-time monitoring for violations of clinically relevant properties is possible.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Robot-assisted surgery</kwd>
        <kwd>Clinically relevant properties</kwd>
        <kwd>Image segmentation</kwd>
        <kwd>Run-time verification</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In this paper, we present a preliminary study of the use of run-time verification in robot-assisted surgery
for cancer. Oncological operations are among the most complex surgical procedures, as the tumor and
chemoradiotherapy can afect surrounding tissues. Changes in anatomy and tissue consistency can cause
surgeons to have more dificulty operating efectively and eficiently. Robot-Assisted Minimally Invasive
Surgery is an increasingly utilized method to treat operable cancer. Robot-assisted surgery ofers
several advantages over traditional surgery, including an enlarged three-dimensional view, articulated
instruments, and improved ergonomics [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. However, for a surgeon to become proficient in
robotassisted surgery and operate without supervision requires considerable experience; for example, to
become proficient in esophageal cancer surgery typically requires experience of 20 to 80 cases. Moreover,
during surgery, the surgeon must take into account many tasks at the same time to achieve the best
outcome for the patient. For surgeons in training, the mental and physical load of these tasks requires a
great deal of energy and time [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. We hypothesize that intraoperative warning systems based on
runtime verification could prevent unwanted actions and thus possibly prevent complications, improving
patient outcomes.
      </p>
      <p>
        The objective of this study is to investigate the feasibility of warning systems based on run-time
verification for intraoperative use. We focus on surgery performed using the da Vinci Surgical System
robot (dVSS), which is widely used in robot-assisted surgery. The manufacturers of the dVSS, Intuitive
Surgical, do not make the interface to the sensors of the robot available to end users or researchers.
Some previous studies have been conducted using the da Vinci Research Kit (dVRK) which does allow
access to the robot’s sensors [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. However, the dVRK only runs on out-of-date hardware that is no
longer used in surgery. In this study, we therefore use the video image available to the surgeon during
surgery as input. The key contributions of this paper are twofold: we report the results of a survey
to elicit from surgeons which clinically relevant properties should be monitored; and we present a
proof-of-concept evaluation of the feasibility of monitoring for one of the properties using the video
image available to the surgeon as input.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Clinically Relevant Properties</title>
      <p>A key problem in providing intraoperative warnings is to determine which properties should be
monitored. In this section, we present the results of a survey to gauge whether surgeons would
consider such warnings useful, and, if so, which clinically relevant properties should be monitored.
The survey was conducted during an international surgical course on Minimally Invasive Gastrectomy
and Esophagectomy. The properties contained in the survey are therefore grounded in surgery for
gastrectomy and esophagectomy, but are also relevant to other forms of robot-assisted minimally
invasive surgery. In conjunction with surgical experts from University Medical Center Utrecht we
formulated eleven diferent properties relevant to preventing postoperative complications that could
give rise to intraoperative warnings.</p>
      <p>A total of 32 surgeons with difering levels of experience from novice to expert took part in the survey,
including surgical fellows and consulting surgeons. The respondents were stratified into two groups
for subgroup analysis: 7 experts (defined as someone who had performed more than 100 RAMIEs), and
25 trainees (who had performed fewer than 100 RAMIEs).</p>
      <p>The surgeons were asked the following questions. The results were descriptively analyzed and the
average response is shown after each question.</p>
      <p>1. A surgical warning system could monitor and warn about the intraoperative events. Do you
think you would experience benefit from such a warning system? [Average score 4.4 (where 1
was no benefit, 5 definitely benefit).]
2. Please indicate which events you believe are relevant (you may select multiple options). [The
percentage of participants who selected the property is shown after each property.]
• Excessive force on tissue [81%]
• Activated energy source too near to vital structure (risk of thermal injury) [75%]
• Irregular tool movement (e.g., sudden or unsure motion of instrument) [75%]
• Thermal instrument out of camera view is activated [63%]
• Surgical material left in body at end of procedure (e.g., gauze, needle) [66%]
• Inadequate suture (e.g., knot tied is not square, incorrect needle position, direction of suture)
[44%]
• Non-addressed bleeding (camera moves away from active bleeding) [41%]
• Instrument collision with vital structure [75%]
• Moving instrument out of camera view [34%]
• Poor tissue stabilization (misgrasping tissue, excessive grasping) [19%]
• Incorrect use of camera (not focused on working area) [9%]</p>
      <p>Interestingly, subgroup analysis shows that, on average, expert surgeons selected 6.4 events (range
4-9), whereas trainees selected 5.6 events (range 2-11), suggesting that experts could see benefit in more
warnings than trainees.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Proof-of-Concept: Surgical Instrument Detection</title>
      <p>Without access to sensor information from the dVSS, it is not possible to monitor for events such as
“Excessive force on tissue” or “Activated energy source too near to vital structure”. However, with
input from a video segmentation system that can reliably identify surgical instruments and anatomical
structures, we can monitor for events such as “Irregular tool movement”, “Non-addressed bleeding”,
“Instrument collision with vital structure”, and “Moving instrument out of camera view”. While possible
in principle, detecting such events is challenging given the current state of the art in image segmentation.
For a proof-of-concept study, we therefore focus on the simplest case of monitoring for instruments
moving out of the camera view.</p>
      <p>The proof-of-concept system consists of an image processing component and a run-time verification
component connected by a simple Python-based client.</p>
      <sec id="sec-3-1">
        <title>3.1. Instrument Segmentation</title>
        <p>
          Recent advances in surgical instrument segmentation allow accurate real-time identification of
instruments in minimally invasive surgery. The particular model used for in the proof-of-concept is based on
the large-scale surgical foundation model introduced by Jaspers et al. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. We used the CAFormer-S18 [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]
variant of their work, pre-trained with self-supervised learning on the SurgeNetXL dataset. This is
a diverse collection of more than 4.7 million surgical video frames drawn from public, private, and
curated YouTube sources [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. This extensive pre-training supports strong feature learning in the domain,
enhancing performance in downstream tasks such as instrument detection. The model was fine-tuned
using the RAMIE dataset in de Jong et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] to ensure strong performance in our intraoperative
thoracoscopic data. The RAMIE data set consists of 879 annotated frames extracted from thoracoscopic
robot-assisted minimally invasive esophagectomy (RAMIE) procedures, collected from 32 patients. Each
frame is densely labeled across 12 classes, including several anatomical structures such as the aorta and
the airways, and four surgical instruments (forceps, suction &amp; irrigation, hook, and vessel sealer). The
annotations were performed by trained researchers and validated by an experienced surgeon.
        </p>
        <p>For our experiments, we extracted only the output segmentation maps for the four surgical
instruments, as the anatomy was not relevant for the monitored property. The presence of each instrument
in each frame is computed as follows:
(1)
(2)
 = ∑︁ ∑︁ (, ),</p>
        <p>where (, ) is the segmentation map for class , defined over all pixel coordinates (, ). The value
of (, ) is 1 if the pixel at (, ) is predicted as class , and 0 otherwise. Thus,  represents the total
number of pixels predicted as class . The binary presence indicator  for class  is defined as:
 =
{︃1, if  &gt; 
0, otherwise
for  ∈ {1, 2, 3, 4}
where  is the pixel count threshold used to determine the presence of the instrument associated with
class . In our experiments,  was set to 50, as this empirically led to stable predictions. If the number
of pixels predicted as class  exceeds the threshold  , the instrument is considered present ( = 1);
otherwise, it is considered absent ( = 0).</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Run-time Verification</title>
        <p>
          For the run-time verification component, we employed the Numerical Runtime Verification (NuRV) tool
[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], specifically the "NuRV orbit" executable, designed for online run-time verification of properties
expressed in Linear Temporal Logic (LTL). NuRV is a robust real-time verification tool developed for
monitoring systems by evaluating user-defined temporal logic properties against live data streams.
NuRV’s specification language allows the definition of properties using temporal logic operators such
as Globally (G), Eventually (F), Next (X), and Until (U). Properties can express complex temporal
constraints succinctly, providing flexibility to specify safety-critical properties. The tool supports
eficient evaluation by compiling these properties into finite-state automata, enabling real-time verdict
generation on system states.
        </p>
        <p>For example, a property that gauze should not be left in the wound can be expressed as
(¬(  ( ∧ )) (it is never the case that gauze proposition holds until and
including the moment when suturing proposition becomes true.</p>
        <p>For our specific monitoring scenario, ensuring that instruments remain within camera view, we
employed a straightforward safety property, (inCameraView ), i.e., that all four surgical instruments
should always be in the camera view which is a clinically relevant safety requirement.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>In this section, we present a preliminary evaluation of the proof-of-concept intraoperative warning
system. The aim of the evaluation was twofold:
• to determine the reliability of surgical instrument detection, i.e., whether there were false positives
or false negatives regarding the detection of tools in the camera view; and
• to determine the time required by both the visual processing component and the run-time
verification software.</p>
      <p>Together, these two factors determine the practical feasibility of real-time surgical monitoring. If the
visual processing component is not suficiently reliable and/or either the visual processing component
or the run-time verification software requires time greater than the frame rate, the warnings will not be
useful.</p>
      <sec id="sec-4-1">
        <title>4.1. Experimental Setup</title>
        <p>The experimental setup involved real-time monitoring during robot-assisted surgery at the University
Medical Center Utrecht. Due to the need to accommodate the surgical schedule, the experiment was
performed during stomach cancer surgery rather than RAMIE on which the model was fine-tuned. From
the point of view of the monitored property there is no diference in dificulty between the procedures;
however, the reliability of the segmentation model may have been impacted.</p>
        <p>The surgical procedure was performed using the da Vinci Surgical System. The dVSS was directly
connected to a computing system running both the neural model for instrument segmentation and
the NuRV orbit monitoring tool. Intraoperative video was captured from the surgical console using
an HDMI-to-USB video capture card with a resolution of 640x480 pixels and a frame rate of 25 frames
per second. The model segmented the surgical instruments using one CUDA core of a GeForce GTX
1050 Ti Mobile graphics card. The setup also included a Python-based client developed specifically for
interacting with the NuRV monitoring tool and logging the monitoring results. The client managed
data capture, pre-processing, segmentation model execution, and communication with NuRV.1</p>
        <p>The surgical procedure lasted more than two hours. However, the run-time verification experiment
was restricted to a 15 minute period during the procedure. During this time the Python client
continuously logged and evaluated the property (inCameraView ). As the experiment was designed solely
to establish the feasibility of monitoring, no warnings were given to the surgeon.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results</title>
        <p>During the surgical procedure, the property (inCameraView ) was frequently violated, indicating
that instruments at least momentarily exited the camera view. In some cases, violations were due to the
segmentation model failing to identify tools which were actually present in the view. In addition, in a
few cases, instruments were identified as being in the camera view when there were no instruments
visible. We conjecture that this was due to the model being trained with videos of a diferent surgical
procedure (RAMIE). Both false positives (instruments not detected when they are present) and false
1The Python code and experimental data are available at https://github.com/UtrechtUniversity/RAS-verification/tree/main/
experiments/runtime-monitoring-er.
negatives (instruments detected when they are not present) are problematic for monitoring. Too many
false positives are distracting and may cause surgeons to ignore real warnings. False negatives mean
that property violations are not being detected.</p>
        <p>TP
TN
FP
FN
Precision
Recall
F1-score</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and Future Work</title>
      <p>
        We presented the results of a study to investigate the feasibility of intraoperative warning systems
based on run-time monitoring. A survey of expert and novice surgeons confirmed the clinical relevance
of warnings of property violations which could result in adverse events during surgery. Although
preliminary, our proof-of-concept experiment demonstrated the feasibility of coupling advanced neural
segmentation with existing runtime verification tools to improve intraoperative safety. Many challenges
remain. While we expect the time required to detect anatomical structures to be similar to that for
instrument detection (as the neural model inherently predicts all relevant classes simultaneously),
improving reliability of the image segmentation component may be necessary to reduce the level of
2The relatively low precision when detecting the suction &amp; irrigation tool may result in too many “false” warnings, but this
tool is less likely to cause damage to tissues when out of camera view.
false positives to levels that surgeons find acceptable. Another direction for future work is developing
mitigation techniques to handle potential transient “hallucinations” or misclassifications produced
by the neural network when the type of surgery and/or surgical instruments difer from the training
data. We also plan to investigate the suitability of other run-time monitoring tools for intraoperative
monitoring, including, for example, the LoLa family of tools [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and RTAMT [10].
      </p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>We gratefully acknowledge funding by NWO project Run-time verification for robot-assisted surgery,
OCENW.M.21.377.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) did not use generative AI in any shape or form.
[10] T. Yamaguchi, B. Hoxha, D. Ničković, RTAMT – Runtime Robustness Monitors with Application
to CPS and Robotics, International Journal on Software Tools for Technology Transfer 26 (2024)
79–99. doi:10.1007/s10009-023-00720-3.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Watanabe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kuriyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Terayama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Okamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kanamori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Imamura</surname>
          </string-name>
          ,
          <article-title>Robotic-assisted esophagectomy: Current situation and future perspectives, Annals of thoracic and cardiovascular surgery : oficial journal of the Association of Thoracic and Cardiovascular Surgeons of Asia 29 (</article-title>
          <year>2023</year>
          )
          <fpage>168</fpage>
          -
          <lpage>176</lpage>
          . doi:
          <volume>10</volume>
          .5761/atcs.ra.
          <volume>23</volume>
          -
          <fpage>00064</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Watanabe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kuriyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Terayama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Okamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kanamori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Imamura</surname>
          </string-name>
          ,
          <article-title>Robotic-assisted esophagectomy: Current situation and future perspectives, Annals of thoracic and cardiovascular surgery : oficial journal of the Association of Thoracic and Cardiovascular Surgeons of Asia 29 (</article-title>
          <year>2023</year>
          )
          <fpage>168</fpage>
          -
          <lpage>176</lpage>
          . doi:
          <volume>10</volume>
          .5761/atcs.ra.
          <volume>23</volume>
          -
          <fpage>00064</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R. Nagyné</given-names>
            <surname>Elek</surname>
          </string-name>
          , T. Haidegger,
          <article-title>Non-technical skill assessment and mental load evaluation in robot-assisted minimally invasive surgery</article-title>
          ,
          <source>Sensors</source>
          (Basel, Switzerland)
          <volume>21</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .3390/ s21082666.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Gogoladze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Alechina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          , R. C. van
          <string-name>
            <surname>Jaarsveld</surname>
            ,
            <given-names>J. P.</given-names>
          </string-name>
          <string-name>
            <surname>Ruurda</surname>
          </string-name>
          ,
          <article-title>Run-time monitoring for robot-assisted surgery</article-title>
          , in: D.
          <string-name>
            <surname>Porello</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Vinci</surname>
          </string-name>
          , M. Zavatteri (Eds.),
          <source>Short Paper Proceedings of the 6th International Workshop on Artificial Intelligence and Formal Verification</source>
          , Logic, Automata, and Synthesis,
          <string-name>
            <surname>OVERLAY</surname>
          </string-name>
          <year>2024</year>
          , Bolzano, Italy,
          <source>November 28-29</source>
          ,
          <year>2024</year>
          , volume
          <volume>3904</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>127</fpage>
          -
          <lpage>133</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3904</volume>
          /paper16.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T. J. M.</given-names>
            <surname>Jaspers</surname>
          </string-name>
          , R. L. P. D. de Jong, Y. Li,
          <string-name>
            <given-names>C. H. J.</given-names>
            <surname>Kusters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. H. A.</given-names>
            <surname>Bakker</surname>
          </string-name>
          , R. C. van
          <string-name>
            <surname>Jaarsveld</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Kuiper</surname>
            , R. van Hillegersberg,
            <given-names>J. P.</given-names>
          </string-name>
          <string-name>
            <surname>Ruurda</surname>
            ,
            <given-names>W. M.</given-names>
          </string-name>
          <string-name>
            <surname>Brinkman</surname>
            ,
            <given-names>J. P. W.</given-names>
          </string-name>
          <string-name>
            <surname>Pluim</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. H. N. de With</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Breeuwer</surname>
            ,
            <given-names>Y. A.</given-names>
          </string-name>
          <string-name>
            <surname>Khalil</surname>
            ,
            <given-names>F. van der Sommen</given-names>
          </string-name>
          ,
          <article-title>Scaling up self-supervised learning for improved surgical foundation models</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2501.09436. arXiv:
          <volume>2501</volume>
          .
          <fpage>09436</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Si</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Metaformer baselines for vision</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>46</volume>
          (
          <year>2023</year>
          )
          <fpage>896</fpage>
          -
          <lpage>912</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>R. L. P. D.</surname>
            de Jong, Y. al Khalil,
            <given-names>T. J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jaspers</surname>
          </string-name>
          , R. C. van
          <string-name>
            <surname>Jaarsveld</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Kuiper</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>R. van Hillegersberg</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Ruurda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Breeuwer</surname>
          </string-name>
          ,
          <string-name>
            <surname>F. van der Sommen</surname>
          </string-name>
          ,
          <article-title>Benchmarking pretrained attentionbased models for real-time recognition in robot-assisted esophagectomy</article-title>
          ,
          <year>2024</year>
          . URL: https://arxiv. org/abs/2412.03401. arXiv:
          <volume>2412</volume>
          .
          <fpage>03401</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cimatti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tonetta</surname>
          </string-name>
          ,
          <article-title>Assumption-based runtime verification with partial observability and resets</article-title>
          , in: B.
          <string-name>
            <surname>Finkbeiner</surname>
          </string-name>
          , L. Mariani (Eds.),
          <source>Runtime Verification</source>
          , Springer International Publishing, Cham,
          <year>2019</year>
          , pp.
          <fpage>165</fpage>
          -
          <lpage>184</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>B. D'Angelo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sankaranarayanan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Sánchez</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Robinson</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Finkbeiner</surname>
            ,
            <given-names>H. B.</given-names>
          </string-name>
          <string-name>
            <surname>Sipma</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Mehrotra</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Manna</surname>
          </string-name>
          ,
          <article-title>LOLA: runtime monitoring of synchronous systems</article-title>
          ,
          <source>in: 12th International Symposium on Temporal Representation and Reasoning (TIME</source>
          <year>2005</year>
          ), IEEE Computer Society,
          <year>2005</year>
          , pp.
          <fpage>166</fpage>
          -
          <lpage>174</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>