<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Visual support for positioning hearing implants</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>H. Ramm</string-name>
          <email>heiko.ramm@1000shapes.com</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>O.S. Morillo Victoria</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>I. Todt</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>H. Schirmacher</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A. Ernst</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. Zachow</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>H. Lamecker</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>1000shapes GmbH</institution>
          ,
          <addr-line>Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Beuth Hochschule für Technik Berlin</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Otolaryngology, Head and Neck Surgery</institution>
          ,
          <addr-line>Unfallkrankenhaus Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Medical Planning Group, Zuse Institut Berlin</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>116</fpage>
      <lpage>120</lpage>
      <abstract>
        <p>We present a software planning tool that provides intuitive visual feedback for finding suitable positions of hearing implants in the human temporal bone. After an automatic reconstruction of the temporal bone anatomy the tool pre-positions the implant and allows the user to adjust its position interactively with simple 2D dragging and rotation operations on the bone's surface. During this procedure, visual elements like warning labels on the implant or color encoded bone density information on the bone geometry provide guidance for the determination of a suitable fit.</p>
      </abstract>
      <kwd-group>
        <kwd>bone anchored hearing implant</kwd>
        <kwd>surgery planning</kwd>
        <kwd>segmentation</kwd>
        <kwd>visualization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Motivation</title>
      <p>To allocate space for the FMT the surgeon has to mill away bone, a procedure that requires a high degree of experience
and precision. 3-dimensional (3D) computed tomography (CT) data is usually acquired preoperatively [1]. There are
two major criteria that are relevant for the surgeon to judge on the suitability of an implant position: the cortical
thickness at the screw positions and the penetration of vulnerable structures. The cortical thickness is crucial for fixation and
for optimal conduction. The cortical bone should at least cover 3 mm of each screw for an optimal fit. The second
criterion is the penetration of vulnerable structures. If there is not enough space in the mastoid bone for the FMT without
harming the ear canal, the dura, the inner ear or the sigmoid sinus, the implantation cannot be performed. Some
structures, for example the dura, can be penetrated to a certain degree. Here, it is important to know the depth of the
penetration. To our knowledge there is currently no software solution available that provides automatic segmentation and
allows the surgeon to try different 3D implant positions and quantify the suitability of a position w.r.t. the above
requirements. The goal of this work was to develop a prototype of a decision support system for an intuitive positioning of
bone conduction implants based on individual patients’ anatomy.</p>
      <p>Related work: Damann et al. [2] investigated the feasibility of positioning hearing aids in the mastoid bone based on
standard software packages. Their approach requires manual segmentation of the CT data (approx. 45 min.) and does
not provide feedback on important parameters like the distance to vulnerable structures. Waringo et al. [3] propose a
framework that optimizes the position of hearing aids in the bone and computes the milling volume. The surgeon has no
visual information about the bone structure or manual control over the implant position. Salah et al. [4] introduce an
interactive method to perform a virtual mastoidectomy based on semi-automatic segmentation of CT data (approx. 15
min.) with an application in cochlear implant planning. To our knowledge there is currently no method available that
provides automatic segmentation of the temporal bone region. Todd et al. [5] argue that this might be related to the large
number of complex shapes and high variation of structure size within this region.</p>
      <p>Contribution: We present a software prototype for the preoperative assessment of suitable positions of hearing implants
in the temporal bone. Our method performs a fully automatic geometric reconstruction of the anatomical structures that
are relevant for the positioning of the implant, in 2 to 3 minutes. This reconstruction result is then used to automatically
pre-position and interactively adjust the implant geometry on the bone with only a few mouse manipulations. During
this interactive process, visual feedback is presented to the user that provides intuitive guidance for finding a suitable
position. The tool is implemented as an extension to the software ZIBAmira (Zuse Institute Berlin, Berlin, Germany)
and is freely available for research purposes (http://www.1000shapes.com/bonebridgeviewer).
2</p>
    </sec>
    <sec id="sec-2">
      <title>Materials and Methods</title>
      <p>The visualization tool builds upon a three stage process (see Fig. 2). First, the image data is automatically segmented
and geometries of the mastoid bone and the relevant structures at risk are generated. Second, distance fields for fast
look-up are generated to efficiently compute the distance to vulnerable structures and the bone density. Third, CT data,
reconstructed anatomical models and precomputed look-up fields are combined to generate intuitive 3D and 2D
visualizations of a chosen implant position. In the following we will describe segmentation, pre-computation and visualization
in more detail.</p>
    </sec>
    <sec id="sec-3">
      <title>2.1 Automatic Segmentation of the Temporal Bone from CT-Data</title>
      <p>For an automatic segmentation a statistical shape model (SSM) of the temporal bone has been generated from 37 CT
scans following the approach presented in [6]. The database used for training the SSM included scans of adults only
(aged 29 to 73) without anomalies of the temporal bone anatomy. The SSM is represented as a triangular mesh
containing 31,810 triangles. It consists of so called patches, i.e. regions on the surface, describing the outer cortical shell of the
skull and the structures at risk (see Fig. 2(a) and (b)). By adaptation of the SSM to new unsegmented image data, we
reconstruct the individual shape of the temporal bone anatomy. At the same time the predefined patch structure serves as
a local atlas and allows for identification of structures at risk.</p>
      <p>Following the framework presented in [7] the fully automatic segmentation process comprises the following three
phases: (i) 3D pose initialization of the SSM within the image data, (ii) adaption of the SSM to the given image data. and
(iii) an unconstrained but regularized fine adjustment of the SSM to account for an individual anatomical shape that is
not captured by the SSM, yet. The SSM can be extended by each segmented structure to enlarge its shape space. To
cope with age-related variation in size, the (adult) model of the temporal bone is initially scaled according to the
provided patient age employing prior knowledge of growth curves of the skull.</p>
      <p>After automatic segmentation of the temporal bone, the resulting triangular surface is decomposed into a mastoid bone
region (patch) that is considered for automatic pre-positioning of the implant and remaining patches that will be used to
compute distances and possible penetrations of the FMT with structures at risk.</p>
    </sec>
    <sec id="sec-4">
      <title>2.2 Pre-computation of a Bone Density Map and Distances to Structures at Risk</title>
      <p>To select a suitable position for the bone screws (i.e., the implant fixtures) it is indispensable to assess the cortical
density of the mastoid bone, which should be at least 3 mm. Therefore, a bone density map is computed for the mastoid bone
patch of the adjusted SSM. Using the method presented earlier in [8], we densely sample the intensity values
represented by Hounsfield Units (HUs) at the inside of the mastoid bone in a 5 mm margin. At each surface location an averaged
value is mapped to the surface and provides an estimate of the density of cortical bone in this region.
For a selected implant position it is important to assure that no vulnerable structure is penetrated, neither by one of the
screws nor by the cylindrical body of the FMT itself. A risk structure distance map is computed to efficiently query (1)
the shortest (signed) Euclidean distance to the closest point of a structure at risk and (2) the patch id, or name of the
vulnerable structures, corresponding to this distance. The distance map is generated by the vector-city vector distance
transform (VCVDT) [9] introduced by Satherley and Jones. As an extension to the VCVDT, we do not only propagate
vector components to compute the Euclidean distance (see [9] for details), but also the index of the corresponding
surface patch. The resulting bone density map and the risk structure distance map (Euclidean distance and patch field) are
stored for later use.</p>
    </sec>
    <sec id="sec-5">
      <title>2.3 3D Visualization of Implant and Bone</title>
      <p>The 3D visualization is the core component of the implant planning tool. It provides an interactive environment for a
targeted search for valid screw positions, i.e. with sufficient cortical bone, and an assessment of implant position and
alignment w.r.t. surgical constraints, e.g. penetration of structure at risk. To identify valid screw positions, the
previously sampled bone density map is employed and visualized color coded. We map the averaged HUs on the mastoid surface
utilizing a bone-like colormap that displays areas of low HU-values as dark semi-transparent regions, whereas dense
cortical bone is displayed in an opaque beige color (cf. Fig. 3(b)). The choice of this colormap follows the natural
intuition, where very thin cortical bone appears semi-transparent.</p>
      <p>Besides the static display of the bone density map for each selected implant position the following parameters are
evaluated and displayed as labels: the distance (or penetration depth) to structures at risk for the cylindrical FMT body and
the screws, as well as the cortical coverage of the screws. Note, that each parameter is only displayed if a critical value
has been reached, e.g. the cortical thickness at a screw is smaller than 3 mm. The cortical coverage of each screw is
sampled from the original CT data in real time while the screw is positioned. At each screw we use one line profile that
follows the centerline of the screw. We densely sample the Hounsfield Units (HU) from the image data onto that profile.
The profiles are then traversed to find a largest connected component, i.e. neighboring points, with a HU value above
250. If the cortical thickness falls below 3 mm a warning label is displayed hovering above the corresponding screw
(see Fig. 3 (b) and (c)).
The third key aspect of the 3D visualization is the display of the penetration depth for the two screws and the cylindrical
body of the FMT. To achieve an efficient distance computation we sample the distance map for each vertex of the three
implant structures, i.e. cylinder and two screws. The signed distance map provides the smallest Euclidean distance to
the structure at risk (note that negative values correspond to positions outside the skull bone, therefore, indicate
penetrations). A look-up of the patch id allows for an efficient query of the corresponding structure. Again, a (yellow) warning
label is displayed close to the respective part of the implant if it is reaching a critical distance to a vulnerable structure.
If a penetration occurs, a red label is displayed including the penetration depth (see Fig. 3 (c)).
3</p>
    </sec>
    <sec id="sec-6">
      <title>Results</title>
      <p>The framework described in the previous section has been implemented as an extension (BoneBridgeViewer) to the
software ZIBAmira that already provides a DICOM import option, as well as 3D and 2D visualization. On start of the
BonebridgeViewer application a simple user interface is presented that allows for import of 3D DICOM data. After
verifying patient information the user selects the desired laterality (left or right ear) before starting the automatic
segmentation process. Including the pre-computation of the distance map this process takes approx. 4 to 5 minutes. After
successful anatomical reconstruction a quad-view or single-view mode is presented. In an interactive 3D viewer the user can
manipulate the implant on the mastoid surface by only a few mouse interactions, e.g., dragging the cylinder to move the
implant on the surface, dragging the wings to rotate the implant, or dragging the wing while pressing the CTRL key to
lift the wings in order to simulate washers that might be used during implantation. At all time the user can verify the 3D
position in the 2D slice viewers. The intersection lines of implant and CT data are visualized to allow for good
assessment of the implant fit. Once finished, typically when no warning labels or only warnings within acceptable bounds
(penetration) are displayed on the implant, screenshots can be generated for documentation. The BoneBridgeViewer is
currently undergoing an evaluation in a clinical research study, where usability and manual effort will be assessed. First
tests indicate manual manipulation times of less than 1 minute for easy cases, or up to 5 to 10 minutes for problematic
cases (e.g. children with a very narrow sinodural angle).</p>
      <p>In a first leave-one-out study we evaluated the automatic segmentation framework on the 37 training datasets. We
compared the automatic reconstruction results to ground truth data of manually segmented mastoid regions and structures at
risk by means of a symmetric surface distance. Automatic segmentation results achieved an average mean distance of
0.68 mm (stddev 0.93 mm) and an RMS distance of 1.18 mm to the ground truth data.
4</p>
    </sec>
    <sec id="sec-7">
      <title>Conclusion and Future Work</title>
      <p>We presented the BoneBridgeViewer, a software prototype that provides intuitive visual support for the positioning of
hearing implants based on patient specific anatomical models derived from medical image data. To our knowledge this
is the first tool that combines fully automatic segmentation of the temporal bone and interactive, visually guided
implant positioning. Although the BoneBridgeViewer provides a full 3D visualization environment, valid implant
positions are restricted to the bony surface and therefore interactions are reduced to simple 2D dragging and rotation
operations. During this interaction all relevant implantation parameters are displayed as easy-to-grasp visual elements in a 3D
environment. Assuming, that the manual effort won’t exceed 10 minutes, the BoneBridgeViewer adds only a small
amount of manual effort for the surgeon (if any, because the CT scan has to be examined anyway). Additionally, various
alternative implant positions can quickly be explored and assessed. In its current state the BoneBridgeViewer allows for
exploration of different implantation scenarios. For a future clinical application it is necessary to transfer a found
implant position to surgery. This could easily be achieved by measuring distances to temporal bone landmarks that are well
know to the surgeon and can easily be identified during surgery.</p>
      <p>Until now the employed SSM of the temporal bone region does not contain anomalies, e.g. a missing ear canal. Future
SSMs will include such anomalies to provide automatic segmentation capabilities for a wider range of patients. The
results of an ongoing clinical evaluation (involving four surgeons) will be used to further improve the usability of the
application and to evaluate its benefit compared to the standard preoperative procedure. Future features might also include
automatic implant positioning capabilities based on an optimization w.r.t. relevant parameters, like cortical coverage of
the screws or the distance to structures at risk. By simply exchanging the implant geometry, our framework could easily
be adapted to similar applications with different implant designs.
6
[1]
[2]</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Vagli</surname>
            ,
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>F</given-names>
            <surname>Turini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F</given-names>
            <surname>Cerri</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E</given-names>
            <surname>Neri</surname>
          </string-name>
          .
          <article-title>Temporal Bone</article-title>
          . In Image Processing in Radiology,
          <volume>137</volume>
          -
          <fpage>149</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Dammann</surname>
            ,
            <given-names>F</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>A</given-names>
            <surname>Bode</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E</given-names>
            <surname>Schwaderer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Schaich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Heuschmid</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M M</given-names>
            <surname>Maassen.</surname>
          </string-name>
          <article-title>Computer-aided Surgical Planning for Implantation of Hearing Aids Based on CT Data in a VR Environment</article-title>
          .
          <source>Radiographics: a Review Publication of the Radiological Society of North America, Inc</source>
          <volume>21</volume>
          , no.
          <issue>1</issue>
          (
          <year>2001</year>
          ):
          <fpage>183</fpage>
          -
          <lpage>191</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>CURAC</surname>
          </string-name>
          (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Salah</surname>
            ,
            <given-names>Z</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Kastner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F</given-names>
            <surname>Dammann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E</given-names>
            <surname>Schwaderer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M M</given-names>
            <surname>Maassen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D</given-names>
            <surname>Bartz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W</given-names>
            <surname>Straßer</surname>
          </string-name>
          .
          <article-title>Preoperative Planning of a Complete Mastoidectomy: Semiautomatic Segmentation and Evaluation</article-title>
          .
          <source>International Journal of Computer Assisted Radiology and Surgery</source>
          <volume>1</volume>
          , no.
          <issue>4</issue>
          (
          <issue>November</issue>
          23,
          <year>2006</year>
          ):
          <fpage>213</fpage>
          -
          <lpage>222</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Todd</surname>
            ,
            <given-names>C</given-names>
          </string-name>
          , M Kirillov,
          <string-name>
            <given-names>M</given-names>
            <surname>Tarabichi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F</given-names>
            <surname>Naghdy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G</given-names>
            <surname>Naghdy</surname>
          </string-name>
          .
          <article-title>An Analysis of Medical Image Processing Methods for Segmentation of the Inner Ear</article-title>
          . In IADIS International Conference Computer Graphics, Visualization,
          <source>Computer Vision and Image Processing</source>
          ,
          <fpage>213</fpage>
          -
          <lpage>218</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Lamecker</surname>
            ,
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Seebass</surname>
          </string-name>
          ,
          <string-name>
            <surname>H-C Hege</surname>
            , and
            <given-names>P</given-names>
          </string-name>
          <string-name>
            <surname>Deuflhard</surname>
          </string-name>
          .
          <article-title>A 3D Statistical Shape Model of the Pelvic Bone for Segmentation</article-title>
          .
          <source>Proceedings of SPIE 5370</source>
          (
          <year>2004</year>
          ):
          <fpage>1341</fpage>
          -
          <lpage>1351</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Seim</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>D</given-names>
            <surname>Kainmueller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Heller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H</given-names>
            <surname>Lamecker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S</given-names>
            <surname>Zachow</surname>
          </string-name>
          ,
          <string-name>
            <surname>H-C Hege</surname>
          </string-name>
          <article-title>Automatic Segmentation of the Pelvic Bones from CT Data Based on a Statistical Shape Model Eurographics Workshop on Visual Computing for Biomedicine (</article-title>
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Seim</surname>
            ,
            <given-names>H</given-names>
          </string-name>
          , H Lamecker, and
          <string-name>
            <given-names>S</given-names>
            <surname>Zachow</surname>
          </string-name>
          .
          <article-title>Segmentation of Bony Structures with Ligament Attachment Sites</article-title>
          .
          <source>In Bildverarbeitung Für Die Medizin (BVM)</source>
          ,
          <fpage>207</fpage>
          -
          <lpage>211</lpage>
          . Springer,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Satherley</surname>
            ,
            <given-names>R</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>M W</given-names>
            <surname>Jones. Vector-City Vector</surname>
          </string-name>
          Distance Transform.
          <source>Computer Vision and Image Understanding</source>
          <volume>82</volume>
          , no.
          <issue>3</issue>
          (
          <year>June 2001</year>
          ):
          <fpage>238</fpage>
          -
          <lpage>254</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>