<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>V. Nguyen);
tmtriet@fit.hcmus.edu.vn (M. Tran)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Transparent Tracking of Spermatozoa with YOLOv8</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bao-Tin Nguyen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Van-Loc Nguyen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Minh-Triet Tran</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>John von Neumann Institute</institution>
          ,
          <addr-line>VNU-HCM</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Science</institution>
          ,
          <addr-line>VNUHCM</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Viet Nam National University Ho Chi Minh City</institution>
          ,
          <country country="VN">Vietnam</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>9</fpage>
      <lpage>0009</lpage>
      <abstract>
        <p>Accurate spermatozoa tracking is crucial for understanding fertilization and reproductive health and developing novel sperm-based diagnostics and therapies. This paper explores the application of YOLOv8, a state-of-the-art object detection model, for automated and robust spermatozoa tracking in microscopic videos. Our approach employs transfer learning, fine-tuning the pre-trained YOLOv8 model on the VISEM-Tracking dataset of labeled spermatozoa images. We evaluate the performance of the proposed method on a specific part of the dataset, which does not participate in the training step. By providing a reliable and eficient tool for automated spermatozoa tracking, this work paves the way for further research, advancements, and applications in the field of reproductive medicine.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Spermatozoa tracking in microscopic videos demands both laser-sharp detection and seamless
connection across frames. This challenge divides computer vision algorithms into two main
camps: the methodical two-stage approach and the lightning-fast one-stage approach.</p>
      <sec id="sec-2-1">
        <title>2.1. The Two-Stage Approach</title>
        <p>
          The two-stage approach in object detection involves initially proposing candidate regions in
an image and subsequently classifying and refining those regions. This approach operates
with meticulous precision. Algorithms under this paradigm meticulously examine each frame,
carefully identifying individual sperm. Examples include advanced methods like DeepSORT
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], an extension of SORT [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] integrating deep appearance features for improved accuracy.
Furthermore, there are methods with Deep Convolutional Neural Networks (DCN) [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], and
variants of the Region-based Convolutional Neural Network (R-CNN) family, such as Faster
R-CNN and Mask R-CNN, are noteworthy representatives [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. These methods excel in
accuracy, especially in challenging environments with complex backgrounds or overlapping
sperm. However, its meticulous nature comes at the cost of processing power, limiting its
real-time capabilities.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. The One-Stage Approach with YOLOv8 in Spermatozoa Tracking</title>
        <p>
          In contrast, the one-stage approach in spermatozoa tracking employs a single, high-speed model
to both detect and trace sperm in microscopic videos. YOLO (You Only Look Once), a renowned
object detection system, exemplifies this approach with its real-time eficiency [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In the
context of the Transparent Tracking of Spermatozoa task at MediaEval 2022, various approaches
utilizing YOLO-based models have been successful. For instance, Huynh et al. developed a
simple eficient framework for tracking sperm cells using a YOLOv7 model [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], particularly
focusing on tail-aware sperm detection [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. In the same task, Kosela et al. solved the problem
using YOLOv5 for object detection and StrongSORT with the OSNet tracking algorithm [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. In
our work, we opt for YOLOv8 due to its balanced performance, ofering a favorable trade-of
between speed and accuracy. YOLOv8’s ease of implementation, coupled with the potential for
optimization through fine-tuning and data augmentation, makes it an ideal choice for real-time
applications such as sperm motility analysis in microscopic videos. Notably, its improved
anchor-free detection enhances robustness in challenging scenarios with complex backgrounds
or overlapping sperm.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Approach</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset splitting</title>
        <p>
          By examining the VISEM-tracking dataset, and from the oficial implementation of the dataset
[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], we noticed that this dataset is well-formed for being processed with YOLO models. In the
training dataset provided by task organizers, there are 20 videos with spermatozoa, each with a
30-second length. We divide this into 2 parts: 80% of the dataset is used for training, and the
rest 20% is used for validation.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Subtask 1: Sperm detection and tracking</title>
        <p>As discussed, regarding the compatibility of the dataset format with the YOLO model, we
employed the YOLOv8 small (s) configuration, we conducted the training duration to 100 epochs
to further perform the detection and tracking task.</p>
        <p>Our decision to use YOLOv8 small and extend the training duration was driven by the need
to strike a balance between sensitivity and adaptability. By setting the Confidence Threshold at
0.25 and the NMS IoU Threshold at 0.7, we aimed to achieve several objectives:
• Balanced Sensitivity and Adaptability to Challenging Conditions: We intend our model
to be sensitive enough to detect spermatozoa accurately, even in challenging conditions.
With a lower Confidence Threshold, we want to ensure a higher recall rate and minimize
the false negatives, as well as make the model more robust and reliable, which is crucial
in medical applications.
• Accuracy and Precision: At the same time, we aim to maintain a balance between accuracy
and precision. This ensures that the detected spermatozoa are indeed spermatozoa and
not false alarms.
• Redundancy Removal: The NMS IoU Threshold at 0.7 allowed us to eliminate redundant
or overlapping bounding boxes. This step is crucial for enhancing tracking accuracy, as
it ensures that each spermatozoon is represented by a single bounding box, preventing
multiple detections of the same sperm.</p>
        <p>In summary, our approach with YOLOv8 small and the specified threshold values was designed
to strike a careful balance between sensitivity, precision, adaptability, and tracking accuracy for
the challenging task of spermatozoa detection and tracking in microscopy images, ultimately
contributing to more accurate and eficient sperm analysis in clinical practice.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Subtask 2: Eficient detection and tracking</title>
        <p>While YOLOv8s works well in accurate sperm identification and tracking, its inherent processing
demands may not be readily compatible with the eficiency of system requirements of Task 2.
Recognizing this potential limitation, we opted to prioritize Task 1’s accuracy and robustness,
ensuring a dependable baseline for further improvement.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Analysis</title>
      <sec id="sec-4-1">
        <title>4.1. Experiments on Validation set</title>
        <p>The table below displays our YOLOv8-s model submission using the IoU threshold set at 0.7.</p>
        <p>Submission
YOLOv8s+Conf@.25</p>
        <p>Precision</p>
        <p>The validation set shows a precision of 0.5 and a recall of 0.628, indicating our approach
detects 62.8% of spermatozoa in microscopy images. However, the 50% false positive rate
suggests room for model optimization. For insights into accuracy and robustness across overlap
thresholds, we considered mAP50 and mAP50-95. mAP50 at 0.506 signifies a mean average
precision of 50%, showcasing reasonable detection accuracy. mAP50-95 at 0.191 emphasizes the
need for improved robustness, especially at stricter overlap thresholds.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Submission results on the Test set of the Medico 2023 Challenge</title>
        <p>
          In the evaluation of the sperm tracking task using YOLOv8, the performance metrics reveal
noteworthy insights. The HOTA (Higher Order Tracking Accuracy) [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] and the OWTA
(OpenWorld Tracking Accuracy) [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] scores demonstrate the overall tracking performance, which
takes into account several of metrics of evaluating the tracker, with an overall HOTA score of
0.415 and OWTA score of 0.46. One of the strengths of the model is its precision, as demonstrated
by the high DetPr (0.506) and AssPr (0.691) scores. These scores suggest that the model is very
good at identifying true sperm tracks and avoiding false positives. However, the recall rates, as
measured by DetRe (0.41) and AssRe (0.597), are somewhat lower. This means that the model
may be missing some sperm tracks, particularly during the association phase.
        </p>
        <p>Besides, the localization accuracy (LocA) of the model is also high at 70.7, indicating that it can
accurately estimate the positions of the sperm cells. Moreover, our model performs inference
with an FPS (frame per second) metric of 54.38 and an FLOPS (floating-point operations per
second) score of 8,570,207,776, which indicates that our model is very competitive in terms of
inference time.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion and Outlook</title>
      <p>YOLO provided some techniques for model augmentation like HSV augmentation, Image Flip,
Image FixUp, Segment CopyPaste. These augmentation methods could potentially enhance the
model’s performance. Moreover, utilizing tools like Ray Tune, used for parameter optimization
could also refine the model’s performance. In the extension of our method, we can consider
using these augmentation and hyperparameter tuning methods to improve the result returned
by YOLOv8.</p>
      <p>Even though the dataset is organized to be compatible with YOLO models, diferent methods
can still be eficiently implemented to perform the sperm object detection and tracking task,
resulting in the understanding of fertilization. In future work, we can find some experiments to
perform transparent tracking and detection of spermatozoa with diferent methods, such as
Deep CNN or Faster R-CNN to evaluate and compare the results between models.</p>
      <p>Moreover, the result from this task can also be inferred to be used for the prediction of motility,
the ability of an organism to move independently. And from that, we can observe whether a
spermatozoon can "move forward", or can only move around in circles.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgment</title>
      <p>This research is funded by Viet Nam National University Ho Chi Minh City (VNU-HCM) under
grant number DS2020-42-01.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Ultralytics</surname>
          </string-name>
          ,
          <fpage>YOLOv8</fpage>
          - docs.ultralytics.com, https://docs.ultralytics.com/models/yolov8/,
          <year>2023</year>
          . [Accessed 15-
          <fpage>12</fpage>
          -2023].
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Storås</surname>
          </string-name>
          , T.-L. Huynh, H.
          <string-name>
            <surname>-D. Nguyen</surname>
            , M.-T. Tran,
            <given-names>T.-N.</given-names>
          </string-name>
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Halvorsen</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Hicks</surname>
          </string-name>
          , Medico Multimedia Task at MediaEval 2023:
          <article-title>Transparent Tracking of Spermatozoa</article-title>
          ,
          <source>in: Proceedings of MediaEval 2023 CEUR Workshop</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Hicks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Storås</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Andersen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Witczak</surname>
          </string-name>
          , T. B.
          <string-name>
            <surname>Haugen</surname>
            ,
            <given-names>H. L.</given-names>
          </string-name>
          <string-name>
            <surname>Hammer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Halvorsen</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
          </string-name>
          ,
          <article-title>Visem-tracking, a human spermatozoa tracking dataset</article-title>
          ,
          <source>Scientific Data</source>
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . doi:
          <volume>10</volume>
          .1038/s41597-023-02173-4.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Wojke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bewley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Paulus</surname>
          </string-name>
          ,
          <article-title>Simple online and realtime tracking with a deep association metric</article-title>
          ,
          <source>in: 2017 IEEE international conference on image processing (ICIP)</source>
          , IEEE,
          <year>2017</year>
          , pp.
          <fpage>3645</fpage>
          -
          <lpage>3649</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bewley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ramos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Upcroft</surname>
          </string-name>
          ,
          <article-title>Simple online and realtime tracking</article-title>
          ,
          <source>in: 2016 IEEE international conference on image processing (ICIP)</source>
          , IEEE,
          <year>2016</year>
          , pp.
          <fpage>3464</fpage>
          -
          <lpage>3468</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T. B.</given-names>
            <surname>Haugen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Witczak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Hicks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Björndahl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Andersen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Riegler</surname>
          </string-name>
          ,
          <article-title>Sperm motility assessed by deep convolutional neural networks into who categories</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>13</volume>
          (
          <year>2023</year>
          ).
          <source>doi:10.1038/s41598-023-41871-2.</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Hicks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Andersen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Witczak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Thambawita</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Halvorsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Hammer</surname>
          </string-name>
          , T. B.
          <string-name>
            <surname>Haugen</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Riegler</surname>
          </string-name>
          ,
          <article-title>Machine learning-based analysis of sperm videos and participant data for male fertility prediction</article-title>
          ,
          <source>Scientific Reports</source>
          <volume>9</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .1038/s41598-019-53217-y.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Somasundaram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nirmala</surname>
          </string-name>
          ,
          <article-title>Faster region convolutional neural network and semen tracking algorithm for sperm analysis</article-title>
          ,
          <source>Computer Methods and Programs in Biomedicine</source>
          <volume>200</volume>
          (
          <year>2021</year>
          )
          <article-title>105918</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.cmpb.
          <year>2020</year>
          .
          <volume>105918</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kosela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aszyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jarek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Klimek</surname>
          </string-name>
          , T. Prokop,
          <article-title>Tracking of spermatozoa by yolov5 detection and strongsort with osnet tracker</article-title>
          ,
          <source>in: MediaEval 2022 Workshop</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>T.-L. Huynh</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-H. Nguyen</surname>
            ,
            <given-names>X.-N.</given-names>
          </string-name>
          <string-name>
            <surname>Hoang</surname>
            ,
            <given-names>T. T. P.</given-names>
          </string-name>
          <string-name>
            <surname>Dao</surname>
            ,
            <given-names>T.-P.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
          </string-name>
          , V.-T. Huynh, H.
          <string-name>
            <surname>-D. Nguyen</surname>
            ,
            <given-names>T.- N.</given-names>
          </string-name>
          <string-name>
            <surname>Le</surname>
          </string-name>
          , M.-T. Tran,
          <article-title>Tail-aware sperm analysis for transparent tracking of spermatozoa</article-title>
          , in: MediaEval 2022 Workshop,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>C.-Y. Wang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bochkovskiy</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-Y. M. Liao</surname>
          </string-name>
          ,
          <article-title>Yolov7: Trainable bag-of-freebies sets new state-of-theart for real-time object detectors</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>7464</fpage>
          -
          <lpage>7475</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Luiten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Osep</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dendorfer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Torr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Geiger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Leal-Taixé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Leibe</surname>
          </string-name>
          ,
          <article-title>Hota: A higher order metric for evaluating multi-object tracking</article-title>
          ,
          <source>International Journal of Computer Vision</source>
          <volume>129</volume>
          (
          <year>2020</year>
          )
          <fpage>548</fpage>
          -
          <lpage>578</lpage>
          . URL: http://dx.doi.org/10.1007/s11263-020-01375-2. doi:
          <volume>10</volume>
          .1007/s11263-020-01375-2.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. E.</given-names>
            <surname>Zulfikar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Luiten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ramanan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Leibe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ošep</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Leal-Taixé</surname>
          </string-name>
          , Opening up open-world tracking,
          <year>2021</year>
          . arXiv:arXiv:
          <fpage>2104</fpage>
          .
          <fpage>11221</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Redmon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Divvala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Farhadi</surname>
          </string-name>
          ,
          <article-title>You only look once: Unified, real-time object detection</article-title>
          ,
          <source>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          (
          <year>2016</year>
          ). doi:
          <volume>10</volume>
          .1109/cvpr.
          <year>2016</year>
          .
          <volume>91</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Mohammadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rahimzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Attar</surname>
          </string-name>
          ,
          <article-title>Sperm detection and tracking in phase-contrast microscopy image sequences using deep learning and modified csr-dcf (</article-title>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>J. B. You</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Riordon</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Nosrati</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Sinton</surname>
          </string-name>
          ,
          <article-title>Machine learning for sperm selection</article-title>
          ,
          <source>Nature Reviews Urology</source>
          <volume>18</volume>
          (
          <year>2021</year>
          )
          <fpage>387</fpage>
          -
          <lpage>403</lpage>
          . doi:
          <volume>10</volume>
          .1038/s41585-021-00465-1.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>T.-Y. Lin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Girshick</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dollár</surname>
          </string-name>
          ,
          <article-title>Focal loss for dense object detection</article-title>
          ,
          <source>in: Proceedings of the IEEE international conference on computer vision</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>2980</fpage>
          -
          <lpage>2988</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>