<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Bone Age Classi cation Using the Discriminative Generalized Hough Transform</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Markus Brunk</string-name>
          <email>markus@brunk-net.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Heike Ruppertshofen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sarah Schmidt</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Peter Beyerlein</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hauke Schramm</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dept. of Engineering, Technical University of Applied Sciences Wildau</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Applied Computer Science, University of Applied Sciences Kiel</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Institute of Electronics, Signal Processing and Communication Technology, Otto-von-Guericke-University Magdeburg</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>284</fpage>
      <lpage>288</lpage>
      <abstract>
        <p>We present an approach for automatic bone age classification from hand x-ray images using the Discriminative Generalized Hough Transform (DGHT). To this end, a region, characteristic for the bone age (e.g. an epiphyseal plate), is localized and subsequently classified to determine the corresponding age. Both steps are realized using the DGHT, whereat the difference of the approaches lies within the employed models. The localization model is able to localize the target region over a broad age range and therefore focuses on the common features of all ages. The model for the classification, in contrast, focuses on the age discriminating features. The classification model consists of several submodels, one for each age class, where each submodel contains information about its age characteristics as well as discriminating features. In a first test the new method was applied to classify images into the two classes 11-12 and 14-15 years and achieved of 95% correct classifications.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The determination of bone age in the clinic is necessary to diagnose early
development disorders of children, as well as in legal cases to determine, if an
adolescent has reached the age of majority.</p>
      <p>
        The two most common approaches for bone age classification are the
manual methods from Greulich &amp; Pyle (GP) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and from Tanner &amp; Whitehouse
(TW) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. In the first method a physician compares all hand bones in an x-ray
image to reference images from an atlas to determine the correct age. In the
latter method a subset of hand bones are classified independently and these
results are then weighted and combined to determine the age. Since both methods
are performed manually, they are time-consuming, user-dependent and require
a substantial amount of expert knowledge.
      </p>
      <p>
        A semi-automatic method is given by computer-assisted skeletal age scores
(CACAS) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which automates the TW approach. However, the classification
still has to be carried out manually. A fully automatic method is given by the
BoneXpert-System [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which determines the age according to the method of
GP or TW using active appearance models and regression. Fischer et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]
developed another fully automatic method using content-based image retrieval.
In this method the regions of interest are compared to an image database and
the similarity is computed using cross-correlation. The age of the best matches
are combined for the final result.
      </p>
      <p>
        In this work, we apply the discriminative generalized Hough transform
(DGHT) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to automatically locate and classify regions of interest for bone age
determination. For this purpose, two models are learned in a fully automatic
manner; one model is used for the localization of the age characteristic region,
the other for its classification to the corresponding age group.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Materials and Methods</title>
      <p>
        The DGHT [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] is a method which has been successfully applied to object
localization in medical images. The method combines the Generalized Hough Transform
(GHT) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] with the Discriminative Model Combination (DMC) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which is used
to assign individual weights to the model points of the GHT models. The
employed point models, representing the target shape, are thereby generated from a
number of images by extracting the edge points from a region around the target
point. The localization results are very robust and the training and evaluation
runs fully automatic [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Here, the DGHT is used o solve two tasks: (i)
localization of an age characteristic region and (ii) classification to determine the
patient’s bone age.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Localization</title>
        <p>The characteristic region is defined as finger base joint of the middle finger.
To obtain a general localization model, capable of robustly detecting the finger
base joints in a wide range of ages, we used image data from various ages for
initial model generation and training. Due to the strong similarity of the four
finger joints over the different ages, we cannot generate a model, which localizes
only the middle finger without any confusion with the other joints. However,
by fitting a mean distance model of the joints to the peaks visible in the Hough
space the position of the middle finger can be robustly determined.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Classi cation</title>
        <p>In the second step, a region of interest around the localized point is extracted and
used for classification. For this purpose, a new kind of GHT models is trained,
containing a combination of age-specific submodels (Fig. 1). An initial model is
generated in the same way as described in Sec. 2.4 using training data from all
considered age classes. Thereafter this model is duplicated for each age class and
combined into a single classification model using class specific offsets. Each class
of the training data is assigned the same offset as the corresponding submodel.
In the training step, each submodel is adapted to the corresponding class using
the DMC weighting procedure; thereby class specific model points get a higher
weight than common points; points belonging to rivaling classes might also be
assigned negative weights.</p>
        <p>Using the trained model in the GHT, it extracts the age specific
characteristics of an image by voting for these characteristics in the resulting Hough space.
The most distinct region of the Hough space, also containing the highest vote, is
the region with the highest number of matching characteristics and corresponds
to the submodel of the correct class.
2.3</p>
      </sec>
      <sec id="sec-2-3">
        <title>Material and Experimental Setup</title>
        <p>For localization we used 158 left hand male and female x-ray images of the ages
11–15. The skeletal age was manually classified by a medical expert using the
GP–method and is used as ground-truth. Each image has an isotropic resolution
of 0:1 mm and a size of approx. 2000 2000 pixel. For classification 69 images of
only female patients are used to account for gender-specific differences. In a first
study we try to separate two classes: (1) 11–12 and (2) 14–15 years. Example
images of the two classes are given in Fig. 2.
2.4</p>
      </sec>
      <sec id="sec-2-4">
        <title>Discriminative Generalized Hough Transform</title>
        <p>To train a localization model, we used 75 images, from which 15 are taken for
model generation. For validation of the localization functionality 83 images in
the age range from 11 to 15 are used. Each image is down-sampled twice and
the Hough space is quantized by a factor of two. Choosing these settings leads
to a better performance without degrading classification rate.</p>
        <p>For the classification we cut out the region of the localized middle finger
and automatically assigned the image extract to one of the two considered age
classes. We used 8 images per class for model generation, 14 images per class
for training and performed the test on the remaining 41 images. These images
are also downsampled twice. Since the classification task does not require a high
precision, the Hough space was quantized by a factor of 10.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <p>In the localization experiments, we found the joint in 95% of the 83 test images.
A localization is considered to be successful, if it is sufficiently exact to cut out
the whole joint for subsequent classification.</p>
      <p>In Fig. 1a the trained classification model with given offsets is illustrated.
In the center an example x-ray image of a finger base joint is displayed with
highlighted reference point. The two submodels for the age classes 11–12 and
14–15 are shown in the right top and bottom, respectively. Figures 1b and
1c show the two submodels in enlarged view with underlaid image extracts for
better illustration. From these images it becomes obvious that the submodel
for 11–12 years focuses on the epiphyseal plate, which has very strong positive
weights in this region. The epiphyseal plate is less adhered at this age than at the
age of 14–15 as can be seen in the underlying example images. In this test 95%
of the test images were classified correctly; only two images were misclassified.
(a) 11–12 years
(b) 14–15 years
(c) 11–12 years
(d) 14–15 years</p>
      <p>Brunk et al.</p>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>The localization and classification results achieved with the described technique
are promising. Both tests yielded a result of 95% of right localizations and
classifications on unknown images.</p>
      <p>In the localization experiment the 5% failures are due to confusions with the
thumb joint and rotations of the hand. These mislocalizations will most likely be
eliminated in future tests by improving the mean distance model used to identify
the middle finger as described in Sec. 2.4.</p>
      <p>In the classification test we observed two misclassifications in the class 11–12.
These two images are shown in Fig. 3(a) and 3(b). Figure 3(c) and 3(d) show
example images of correct classifications. Comparing these images it becomes
obvious that in the range of the epiphyseal plate the bone is much more adhered
in the misclassified images then would be expected for this age. From this, one
can conclude that either the ground-truth is wrong or one joint is not sufficient
for age classification since the ground-truth was obtained from the whole image
with the GP method using all bones for classification. Potentially the untested
joints or the carpal bone might appear more like the ground-truth class.</p>
      <p>In the future, we will combine the classification results of all joints and the
carpal bone to obtain a better and more detailed estimate for groups with a
difference in age of less then one year. The results of each classification should
then be combined in a suitable weighted manner, e.g. following the TW method.
Furthermore, additional age specific submodels will be included into the
classification model, allowing for an age classification with a larger number of classes.
Further experiments will be conducted with larger number of images, also using
public databases for better comparison of classification performance.
Acknowledgement. We thank Dr. Kayser, University Medical Center
Schleswig-Holstein, Germany for providing the x-ray images used in this study.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Greulich</surname>
            <given-names>WW</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pyle</surname>
            <given-names>SI</given-names>
          </string-name>
          .
          <article-title>Radiographic atlas of skeletal development of the hand and wrist</article-title>
          . 2nd ed. Stanford University Press;
          <year>1976</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Tanner</surname>
            <given-names>JM</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Healy</surname>
            <given-names>MJR</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldstein</surname>
            <given-names>H</given-names>
          </string-name>
          , et al.
          <article-title>Assessment of skeletal maturity and prediction of adult height (TW3 method)</article-title>
          .
          <source>3rd ed. WB Saunders</source>
          ;
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Thodberg</surname>
            <given-names>HH</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kreiborg</surname>
            <given-names>S</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Juul</surname>
            <given-names>A</given-names>
          </string-name>
          , et al.
          <article-title>The boneXpert method for automated determination of skeletal maturity</article-title>
          .
          <source>IEEE Trans Med Imaging</source>
          .
          <year>2009</year>
          ;
          <volume>28</volume>
          (
          <issue>1</issue>
          ):
          <fpage>52</fpage>
          -
          <lpage>66</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Fischer</surname>
            <given-names>B</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brosig</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Welter</surname>
            <given-names>P</given-names>
          </string-name>
          , et al.
          <article-title>Content-based image retrieval applied to bone age assessment</article-title>
          .
          <source>In: Proc SPIE</source>
          . vol.
          <volume>7624</volume>
          ;
          <year>2010</year>
          . p.
          <fpage>35</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ruppertshofen</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lorenz</surname>
            <given-names>C</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beyerlein</surname>
            <given-names>P</given-names>
          </string-name>
          , et al.
          <article-title>Fully automatic model creation for object localization utilizing the generalized Hough transform</article-title>
          .
          <source>Proc BVM</source>
          .
          <year>2010</year>
          ; p.
          <fpage>281</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ballard</surname>
            <given-names>DH</given-names>
          </string-name>
          .
          <article-title>Generalizing the Hough transform to detect arbitrary shapes</article-title>
          .
          <source>Patt Recogn Lett</source>
          .
          <year>1981</year>
          ;
          <volume>13</volume>
          (
          <issue>2</issue>
          ):
          <fpage>111</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Beyerlein</surname>
            <given-names>P</given-names>
          </string-name>
          .
          <article-title>Discriminative model combination</article-title>
          .
          <source>In: Proc ICASSP; 1998</source>
          . p.
          <fpage>481</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>