<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>International Journal of Computer Assisted Radiology and Surgery</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Image stitching of sphenoid sinuses from monocular endoscopic views</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>T. Bergen</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>P. Hastreiter</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>C. Münzenmayer</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>M. Buchfelder</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>T. Wittenberg</string-name>
          <email>thomas.wittenberg@iis.fraunhofer.de</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department for Neurosurgery, University Clinics Erlangen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Fraunhofer Institute for Integrated Circuits IIS</institution>
          ,
          <addr-line>Erlangen</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Nasal Septum</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2006</year>
      </pub-date>
      <volume>2</volume>
      <fpage>352</fpage>
      <lpage>354</lpage>
      <abstract>
        <p>For operations of the pituitary glands, the most subtle method is an intervention through the paranasal and especially through the sphenoid sinus. To avoid dangerous interference with adjacent organs and nerves, the surgeon has to orient himself in the very small sphenoid cavity and navigate across the hollow space to break through the sellar floor to the pituitary gland above. Especially in reoperations or anatomical variants such as so-called kissing carotids, transsphenoidal surgery is a challenge even in experienced hands. To support such a surgery, various imaging modalities can be applied such as CT, MRI or endoscopy. While pre-operative MRI or CT-data can be used for intervention planning and navigation support, endoscopy can be applied intra-operatively for the examination of surfaces inside the sphenoid sinus. In this work, we present initial experiments and results from real-time panorama-endoscopy of the sphenoid sinus for navigation and orientation support, based on monocular endoscopic sequences of a skull phantom, yielding partial reconstructions of the walls of the sphenoid sinus.</p>
      </abstract>
      <kwd-group>
        <kwd>pituitary surgery</kwd>
        <kwd>sinus surgery</kwd>
        <kwd>panorama-endoscopy</kwd>
        <kwd>stitching</kwd>
        <kwd>mosaicking</kwd>
        <kwd>real-time</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>The most subtle method for operations of the pituitary glands, such as the
removal of tumors or adenomas, is transsphenoidal surgery. This involves the
difficulty of maneuvering through the paranasal and especially the sphenoid
sinus, a small cavity behind the eyes, to break through the sellar floor and
gain access to the pituitary gland. This is a difficult operation due to the risk
of damaging adjacent nerves and organs, such as the internal carotid artery.
depicts the situation in a CT slice. Various imaging modalities can be applied
to support the surgeon, including e.g. CT, MRI or endoscopy. CT and MRI Optical
are available in the pre-operative planning phase. The standard imaging mo- Sphenoidal Nerve
dality during the operation is the view through an endoscope. One major as- Sinus
pect of difficulty is the limited field of view provided by the endoscope. To Tumor
improve orientation and maneuverability for the surgeon, image stitching
techniques can be applied to provide an augmented field of view. In this pa- Figure 1: Transnasal approach to the
per, we propose a real-time panorama-imaging approach for navigation and tumor in the pituitary gland, depicted in
orientation support, based on monocular endoscopic sequences of a skull an axial CT slice of a head.
phantom, yielding partial reconstructions of the walls of the sphenoid sinus.</p>
      <p>
        These experiments are based on prior experiences, gained from a 3D
reconstruction approach from endoscopic views [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Further work concerning view enhancement in sinus surgery includes
registration techniques for CT/endoscopy registration by Burschka et al. and Mirota et al. [2, 3]. Wise and DelGaudio as
well as Palmer and Kennedy provide review articles of computer-assistance in paranasal sinus surgery [4, 5]. Different
aspects of navigation and registration of pre- and intra-operative imaging techniques are discussed as a means of
facilitating orientation for the surgeon. However, panorama-endoscopy has not yet been considered in the field of sinus
surgery.
      </p>
      <p>Nasal cavity</p>
      <p>Single
endoscopy
image</p>
      <sec id="sec-1-1">
        <title>Image preprocessing</title>
      </sec>
      <sec id="sec-1-2">
        <title>Frame-to-scene registration</title>
      </sec>
      <sec id="sec-1-3">
        <title>Hybrid feature extraction and matching</title>
      </sec>
      <sec id="sec-1-4">
        <title>Panorama</title>
        <p>rendering</p>
        <p>Panorama
image
In this work, we present a system for real-time panorama imaging (“mosaicking”) of monocular endoscopic views with
application to sphenoid sinuses. The approach is based on a system, which we have published earlier [6] for real-time
stitching of the urinary bladder. In this section, the algorithmic components are described. Figure 2 depicts an overview
of the proposed approach. In the following sections, we describe all steps in further detail.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Image Preprocessing</title>
      <p>The video frames are captured from the camera at a rate of about 30 frames per second. Every video frame is
preprocessed to detect the circular mask (aperture) typical for endoscopic recordings. This is achieved by segmenting all
nonblack pixels from the image and fitting a circular disc to the extracted region. Furthermore, we compensate for lense
distortion and inhomogeneous illumination. Endoscopic images usually suffer from a barrel distortion. To reduce this
effect, we apply an undistortion filter, computed on the basis of priorly captured images of a checkerboard pattern. We
use the undistortion filter provided by the OpenCV software library. Inhomogeneous illumination, i.e. a strong
vignetting effect, is caused by the point light source generally used in endoscopy. We compensate for illumination
inhomogeneities by applying a high-pass filtering to the input image: a strongly smoothed image is subtracted before
passing the frame to the feature tracking module.</p>
    </sec>
    <sec id="sec-3">
      <title>Feature tracking and frame registration</title>
      <p>This section is based on a recent publication of ours, describing a hybrid tracking approach for real-time stitching
during cystoscopy [7]. Here, we briefly describe the essential steps and refer the reader to [7] for details. The tracking
module encapsulates both the SURF (Speeded Up Robust Features [8]) and KLT (Kanade-Lucas-Tomasi [9]) tracking
algorithms in a multi-threaded implementation. While SURF generates feature descriptors that allow matching features
from the current video frame to the global set of all prior features, KLT is more suited for matching features between
successive video frames. Consequently, a mosaicking system based on KLT tracking suffers from a drift error, which
increases over time. On the other hand, KLT requires less computational time than SURF, making it very suitable for
real-time applications. In order to exploit the advantages of both approaches, we combine KLT and SURF tracking to
achieve both, a fast processing speed as well as high matching accuracy. Both tracking threads calculate a projective
transformation within a RANSAC (RANdom SAmple Consensus) scheme to align the current video frame to the scene,
i.e. the panorama coordinate space.</p>
    </sec>
    <sec id="sec-4">
      <title>Panorama rendering</title>
      <p>Based on the previous image registration step, all processed frames are rendered as a panorama image using OpenGL.
The two-dimensional projective transformation is converted to a three-dimensional transform, which maps the
respective frame texture to the xy-plane. To reduce visible seams along the edges of a frame, we use a basic alpha blending
approach. The alpha channel of each frame is designed as a center weighted function with α = 1 in the central image
region and linearly decreasing value to the outer image edge. Due to the real-time processing ability, the system is able
to dynamically extent the endoscopic view field during the procedure. This motivates our choice to always use the most
recent video frame as reference frame placed at the center of the screen surrounded by the panorama. For the final
visualization (as depicted in the results section), the panorama is displayed in reference to a coordinate system, defined by
the average projective transform of all images to present a panorama image with small global deformation.
Twenty endoscopic panoramas from monocular image sequences of the sphenoid sinus and the pituitary glands have
been obtained of the skull phantoms in real-time (ref.</p>
      <p>Figure 3). All panoramas were directly computed during slow manual movements of the endoscope tip in a translational
way through the hollows. Fehler! Verweisquelle konnte nicht gefunden werden. depicts typical examples of five
panoramic images obtained from the pituitary glands of the skull phantom. A white circle approximates the size of one
original endoscopic view. Table 1 summarizes further information about the panorama images. Frames Total is the total
number of image frames considered for the panorama generation. Frames In Scene is the number of frames that could
be successfully registered to form the panorama. Other frames are omitted either due to insufficient quality (too few
corresponding feature points) or due to the fact, that the SURF tracking is not executed with full video frame rate but
processes only about every third to fourth frame. In general, this is not of any disadvantage, since it is still sufficient to
provide enough overlap between frames for successful stitching. Scene Features is the number of SURF feature points,</p>
      <p>A
the scene consists of.</p>
      <p>C</p>
      <p>D</p>
      <p>B</p>
      <p>E
Panorama
A
B
C
D
E</p>
      <p>Frames Total
773
541
701
447
532</p>
    </sec>
    <sec id="sec-5">
      <title>Discussion 5</title>
    </sec>
    <sec id="sec-6">
      <title>Summary</title>
      <p>The results show that the proposed approach is applicable to the problem of real-time image stitching of the sphenoid
sinuses. The generated panorama images from five experiments with a skull phantom have been presented, each
consisting of about two to three hundred single video frames. These first experiments show the potential of the approach.
Further experiments will be conducted with clinical endoscopic video sequences obtained during transsphenosoidal surgery
to validate the method with real patient data.</p>
      <p>By applying a real-time image stitching approach to monocular endoscopic images from a skull phantom, an augmented
field of view can be provided to the surgeon. We successfully stitched several image sequences obtained by an
endoscope and generated panorama views in real-time consisting of several hundred single video frames. This technique has
the potential of improving orientation and maneuverability for the surgeon during difficult transsphenosoidal
procedures.
6</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>[1]</mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>