<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>ReVeS Participation - Tree Species Classi cation using Random Forests and Botanical Features?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Guillaume Cerutti</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Violaine Antoine</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Laure Tougne</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julien Mille</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lionel Valet</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Didier Coquin</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antoine Vacavant</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Clermont Universite , Universite d'Auvergne, ISIT</institution>
          ,
          <addr-line>F-63001, Clermont-Ferrand</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LISTIC</institution>
          ,
          <addr-line>Domaine Universitaire, F-74944, Annecy le Vieux</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universite Lyon 1</institution>
          ,
          <addr-line>LIRIS, UMR5205, F-69622</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Universite Lyon 2</institution>
          ,
          <addr-line>LIRIS, UMR5205, F-69676</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Universite de Lyon</institution>
          ,
          <addr-line>CNRS</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper summarizes the participation of the ReVeS project to the ImageCLEF 2012 Plant Identi cation task. Aiming to develop a system for tree leaf identi cation on mobile devices, our method is designed to cope with the challenges of complex natural images and to enable a didactic interaction with the user. The approach relies on a two step model-driven segmentation and on the evaluation of high-level characteristics that make a semantic interpretation possible, as well as more generic shape features. All these descriptors are combined in a random forest classi cation algorithm, and their signi cance evaluated. Our team ranks 4th overall, 3rd on natural images, which constitutes a very satisfying performance with respect to the project's objectives.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        The ability to recognize a plant species and to understand its speci cities has
now become a task accessible mostly to specialists. Most ora books promise an
arduous time to the willing neophyte, who does not possess the compulsory
theoretical background. Mobile systems however o er the opportunity to introduce
such knowledge in an interactive way, at the level of the user. Mobile guides
for plant species identi cation have already seen light [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] with great success on
white background images. The goal of the ReVeS project is to build a system to
help users to recognize a tree in a natural environment, from the photograph of
a leaf, in an educational and interactive way.
      </p>
      <p>
        With this objective, we participated to the ImageCLEF Plant Identi cation
task [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for the second time, treating almost all of the 126 species in the database,
making a strong distinction between the 24 species with compound leaves, and
the 100 species with simple leaves we considered. The task consisted in
associating, after a training phase, each one of the 3150 images in the Test database to
an ordered list of species. Our work focused mainly of the application of methods
dedicated to the case of photographs of one leaf in a natural environment.
? This work has been supported by the French National Agency for Research with the
reference ANR-10-CORD-005 (REVES project).
      </p>
    </sec>
    <sec id="sec-2">
      <title>Model Based Leaf Segmentation</title>
      <p>
        Retrieving the leaf contour is the rst and crucial step for the understanding
of the image. It is a really challenging issue in unsupervised, complex, natural
images [
        <xref ref-type="bibr" rid="ref20">21,20</xref>
        ] where it is necessary to incorporate as much knowledge as possible
to ease the task [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Including prior knowledge on the expected shape of the
object we look for is a good way to reduce the risk of mistakes. But in the
context of a mobile application, it would be regrettable not to take advantage of
a human user to guide an automatic process that would otherwise be very prone
to erratic behaviour.
2.1
      </p>
      <p>Leaf Color Model
In potentially imperfect images, and with the idea of approximating the shape
of the leaf, color seems to be the most relevant information to rely on. A valid
a priori color model for all leaves is impossible given the variety induced by
season, species and lighting, but to extract a color model from every image, we
rst need to have a rough idea of the leaf's location.</p>
      <p>
        What is quite easily solvable on white background image where a simple
thresholding the grey-level image is enough to locate the leaf, becomes much
more complicated with natural images. This is where we require the assistance
of the user to draw a region inside the leaf, region that has to contain at least
three components in the case of a compound leaf. We also rotated and cropped
some photographs so that they clearly contain only one leaf of interest, with its
apex pointing approximately to the top of the image, which corresponds to our
frame of work for a mobile application. This is the only human intervention in
the recognition process, and it is performed for photograph images only.
Based on this rst, initial region, we try to estimate a model of the color of
the leaf, and to compute the distance of each pixel to this model. This is done
by using an evidence-based combination of the dissimilarity to a global color
model computed by linear regression on the initial region (Figure 2(b)) and the
dissimilarity to a local adaptive model built by adapting an expected mean color
while exploring the image (Figure 2(c)). The use of belief function theory [
        <xref ref-type="bibr" rid="ref18 ref6">6,18</xref>
        ]
is here more e cient than simpler combinations (mean, minimum) to take the
speci cities of two relevant yet distinct sources of information 2.
We rely on explicit shape models to drive the segmentation of the leaves in
images. It is however very complicated if not impossible to cover the diversity of
leaf types, from needles to bipinnate leaves. This is why we had to introduce two
di erent models, one for simple or palmately lobed leaves and one for compound
leaves (pinnate, digitate and bipinnate indi erently) We thus considered only
Angiosperm species, leaving Ginkgo biloba and Juniperus oxycedrus aside.
Parametric Active Polygon In the case of simple or palmately lobed leaves
we still use a polygonal leaf model [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to produce both a rough segmentation of
the leaf and an estimation of its global shape. The number of lobes is estimated
during the evolution, which is based only on the previously computed color
dissimilarity map.
      </p>
      <p>A Model For Compound Leaves A leaf is supposed to be compound when
the initial region (after either colouring or thresholding) has at least three
connected components. To evaluate the number of lea ets nF and their organization,
we designed a deformable template for compound leaves that will evolve based
on the dissimilarity map.</p>
      <p>(a)
(b)
(c)
(d)
(e)</p>
      <p>This model represents lea ets by a xed, excessive number of pairs of circles,
and relies again on two points B and T and on a set of parameters to be optimized
by minimizing an energy based on the total dissimilarity:
{ kF , the curvature of the axis
{ rF , the radius of the circles
{ dF , the distance of the circles to the axis
{ pF (i), the relative position of each pair of circles</p>
      <p>Given the possibility of overlap between lea ets, and subsequently model
circles, we chose not to optimize the number of lea ets during the evolution,
but to try to estimate it a posteriori with a a global view. Using the radius and
position parameters, we rst locate groups of connected circles, and based on
their sizes and the gaps between them, we estimate how many lea ets actually
compose each of them. This is a hard problem, and to reduce the risk of error,
we make to estimation the retrieve a minimal and maximal number of lea ets
and their location in the image. This process is illustrated in Figure 4. The
numbers of lea ets and the parameters of the model are used as descriptors for
the compound leaf structure.</p>
      <p>To additionally represent the shape of the lea ets, we evaluate a simple
polygonal model on one of the located lea ets. To minimize the risk of the model
over owing in the neighbouring lea ets and losing any accuracy, we chose the
lea et which stands out the most in the maximal lea et estimation, considering
the distance to the previous and next pair of lea ets.</p>
      <p>
        Constrained Active Contour
We use the polygonal approximation both as an initialization and a shape
constraint to retrieve the actual contour of the leaf, using an active contour
algorithm for exact segmentation [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. In the case of plain background images, the
shape constraint is suppressed so that it doesn't prevent the contour to t the
actual leaf margin. Figure 5 shows the interest of including a shape constraint
on complicated images.
To represent the discriminating properties encountered on the leaf, we chose
to seek for the information investigated by botanists to identify species. The
descriptors we use are then high-level morphological features designed to capture
these speci cities.
The represent the global shape of the leaf, we use directly the parameters
obtained after the evolution of the global models. For simple and palmately lobed
leaves, they consist of:
{ model width w
{ model center position p
{ model apex angle A
{ model base angle B
{ number of lobes nL
{ length of each pair of lobes lL(i)
{ angle of each pair of lobes L(i)
      </p>
      <p>For compound leaves, we keep parameters from the compound leaf model,
and those from the polygonal model extracted on one lea et:
{ minimal number of lea ets nF min
{ maximal number of lea ets nF max
{ position of the rst lea et hF
{ average gap between lea ets gF
{ distance to the axis dF
{ lea et size sF
{ lea et model width wF
{ lea et model center position pF
{ leaf model apex angle AF
{ leaf model base angle BF</p>
      <p>Contour Interpretation
In order to capture more local shapes generally considered in the process of leaf
identi cation, we rely on the axes of the polygonal model to partition the nal
contour into base, apex, lobe tip and right or left margin areas and know where
to look at to estimate the determining features characterizing the leaf margin,
basal shape and apical shape.</p>
      <p>
        Curvature Scale Space Transform A rich representation of the contour
used for shape matching is the Curvature Scale Space [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] that has already been
used in the context of leaf image retrieval [
        <xref ref-type="bibr" rid="ref12 ref4">12,4</xref>
        ]. It consists simply in piling
up curvature measures estimated on a normalized contour, producing a visually
intuitive, scale invariant description of a contour. Figure 6 illustrates the interest
of this transform for leaf margin analysis.
CSS Image Description As such, the CSS image presents already a lot of
visual information that may well represent the properties of the margin.
Considering it as a describable visual object, we decided to extract texture information
on this very image. We computed Haralick descriptors on a grey level version of
the image to get a generic source of information on the margin, that does not
take the risk of making too many assumptions.
      </p>
      <p>
        Detecting and Characterizing Teeth On the other hand, we also wanted to
locate and describe explicitly the teeth and pits on the leaf's margin. Considering
that such structures clearly stand out in the CSS representation as maxima and
minima of curvature, we followed an approach close to the detection of dominant
points [
        <xref ref-type="bibr" rid="ref14 ref19">19,14</xref>
        ] to retrieve at each scale the salient features on the contour. For
each one of them, the last scale at which a point is detected informs us on its
size, while its sharpness can be estimated as the mean curvature of the point at
all the scales it is detected.
      </p>
      <p>Such points are searched only in the contour areas corresponding to the
margin, so that the largest elements (apex, lobe tips, petiole) do not absorb the
interesting smaller teeth. This detection process produces an interpretation of the
contour where the base and the apex are precisely located, and with a sequence
of convex and concave parts, characterized by their scale S and curvature K, as
depicts Figure 7</p>
      <p>To produce descriptors suitable for classi cation, we chose to compute the
average values and standard deviations for both concave ( ) and convex (+)
structures, giving a set of 8 margin descriptors:
{ mean scale of teeth S+ { mean curvature of teeth K+
{ standard deviation S+ { standard deviation K+
{ mean scale of pits S { mean curvature of pits K
{ standard deviation S { standard deviation K</p>
      <p>Additionally, we measured what percentage w+ of the total margin length
was part of a convex element, what percentage w was part of a concave one,
and what percentage w0 was part of none. These descriptors sum up some of the
speci cities botanists would look at to characterize a leaf margin, yet in a very
condensed and e cient way.
3.3</p>
      <p>Basal and Apical Shape Estimation
To account for the shape of the leaf int the basal and apical areas, we transposed
the approach used with the global shape, by designing a simple, parametric,
exible model to adjust to the contour. It is attached to the contour point
detected as the base or apex and composed of two parametric Bzier curves
that try to minimize the distance of their points to the contour. Figure 8 shows
examples of the evolution of these models.</p>
      <p>
        The parameters used to build the necessary control points and optimized
during the evolution, are also the descriptors we use to represent those shapes:
{ , the global angle of the model
{ o, the orientation angle relatively to the leaf axis
{ ( t1)l;r, for each side, the tangent angle at the tip point
{ ( t2)l;r, for each side, the tangent angle at the end point
{ ( t1)l;r, for each side, the distance of the 1st control point from the tip
{ ( t2)l;r, for each side, the distance of the 2nd control point from the end
In addition to these dedicated descriptors, we also consider more generic shape
features that could be apply basically to any kind of object. The moments are
a popular choice for their invariance properties and have already been used in
the context of leaf images [
        <xref ref-type="bibr" rid="ref12">21,12</xref>
        ]. We computed the central moments on the
segmented leaves and use them in combination with the other descriptors as a
complementary shape representation, less directly pertinent but possibly more
stable.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Feature Selection and Classi cation</title>
      <p>Segmentation methods previously described enable us to build an ensemble of
descriptors which de nes the properties of the leaves. First, knowing this
ensemble, we chose an appropriate classi er and we selected features that are strongly
relevant for the classi cation, removing useless and noisy descriptors. We studied
then the behavior of the selected classi er when its parameters vary. Plant
identi cation was nally carried out using the optimized parameters of the classi er.
4.1</p>
      <p>Particularity of the database
The extracted descriptors correspond to numerical properties of the images. It
forms a vectorial data set that can be study in order to select a suitable classi er.
Let us rst denote that the database may include noise, specially in the case of
photographs, since the background images may contain colors quite similar to
the leaves. Furthermore, the number of images per class is heterogeneous: some
classes contain more than 200 objects whereas some are represented by less than
5 objects. This property can a ect the e ciency of some classi ers. Finally, the
di erence between leaves for the same species are most of the time high. For
example the shape or the color of a leaf can be di erent following its age or
depending on the season the picture has been taken. As a result, there exists
a large variation of the similarity within species. Conversely, a low variation
between classes may happen, as some species have similar leaves.
4.2</p>
      <p>
        A random forest classi cation
Random forest, introduced by Breiman [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], is a recent technique which consists
in ensemble of decision trees using for the nal prediction a majority vote. To
build a decision tree in a random forest, a bootstrap sample of the data is
used and at each node a set of random variable is selected to split on. This
random sampling strategy makes increase the error of each tree and reduces the
correlation between trees. As a consequence, the ensemble achieves both low
variance and low bias [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        Random forest has shown high performances in many domains such as
bioinformatics [
        <xref ref-type="bibr" rid="ref10 ref15 ref7">10,15,7</xref>
        ], ecology [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] or computer vision [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. This method is robust
to noise and generalize correctly the class models, even when the database has
few examples per class. We decided then to use this classi er for the task of
plant identi cation.
4.3
      </p>
      <p>
        Methodology
For the next experiments, we used as decision trees the CART algorithm with the
gini impurity measure [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. We chose to built 200 trees and we set the parameter
mtry, i.e. the number of randomly selected variables at each split, to pp (where
p is the number of attributes), as it is the default value commonly used in the
literature. In order to measure the accuracy of the random forest, a 3-fold cross
validation is employed.
      </p>
      <p>The dataset Pl@ntLeaves II includes three types of leaves: the rst one
corresponds to compound leaves, the second and third ones to simple leaves with
and without lobes. For each type of leaves, the number of attributes di ers, as
well as the importance of the attributes in the classi cation process. Thus, we
decided to built three random forests.
4.4</p>
      <p>Feature selection
For simple leaves, we opted for the four following strategies:
{ Strategy 1 consists in selecting the polygonal leaf model inducing the global
shape model, the basal and apical estimations, as well as the attributes
extracted from the CSS representation.
{ Strategy 2 includes the polygonal leaf model, the basal estimation, the apical
estimation and the contour characterization with the Haralick descriptors.
{ Strategy 3 is a mix of the strategies 1 and 2.
{ Strategy 4 includes the strategy 3 and the central moments.
73
72
71
70
69
68
67
66
65
64
63
66
64
62
60
58
56
54</p>
      <p>For compound leaves, the previous models (i.e. the polygonal model, the basal
and apical model, etc.) characterize a small region of the image corresponding to
one lea et. As the resolution is lower than for simple leaves, the attributes are
less e cient. We decided then to automatically erase useless or noisy attributes.</p>
      <p>
        Such task can be performed using the variable importance measure described
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. This measure consists in observing the variation of the accuracy when one
variable of the dataset is randomly permuted. Figure 10(a) shows the results
when the attributes from all the models are taken. We can observe the existence
of a quite large amount of useless variables, since they are represented by low
measures. We decided then to remove all the variables that have an importance
close to 0.
      </p>
      <p>As a result, we considered 6 strategies which always includes the compound
model, the polygonal model and the basal and apical shape estimations:
{ Strategy 1 is composed of the Haralick descriptors computed from the
contour detection.
{ Strategy 2 corresponds to the strategy 1 with the add of the attributes
extracted from the CSS representation.
{ Strategy 3 includes the strategy 2 and the central moments.
{ Finally, the three last strategies consist in the three rst strategies
suppressing the noisy variables.
53
52
51
50
49
48
47
46
45
44
1
2
3
4
5</p>
      <p>6
(b)</p>
      <p>Results are presented in gure 10(b). We can observe that deleting noisy
variables leads most of the time towards better performances. We eventually
chose the last strategy, ending up with 41 attributes.</p>
      <p>nal parameters for the random forest
In order to achieve good performances, a random forest needs to ne-tune its
parameters. One of the most important to adjust is mtry, the number of variables
selected randomly at a split. Figure 11 presents the accuracy varying with mtry
for simple leaves with a unique lobe. When mtry = 1 the strength (i.e. the
accuracy) of each tree is low and the correlation between trees is minimal. As a
consequence, the ensemble performance is low. Conversely, if mtry corresponds
to the number of attributes, the decision trees of a random forest are built
without randomness. Although the strength of each tree is high, the correlation
between trees is quite important and induces low performances. Thus, we set
mtry = 20 in order to have a good trade between strength and correlation.
However, due to a lack of time when carrying out the experiments concerning
this parameter, we chose for the contest to set mtry = p74 9.</p>
      <p>The same reasoning has been achieved for the two other types of leaves,
resulting in a selection of 16 attributes for simple leaves with several lobes and
15 attributes for compound leaves.</p>
      <p>A second important parameter to ne-tune is the number of trees to build
for a forest. This number must be high enough to decrease the variance, leading
to stable and accurate results. However, the time execution increases with the
number of trees. We decided then to set this parameter to 5000 for the three
forests.
70
yca
rca
cu69
68
67
66</p>
      <p>mtry40
10
20
30
50
60
70
Three random forests are trained with the previous parameters and the whole
dataset. As we only considered Angiosperm species, untreated leaves from the
test set are assigned to Ginkgo biloba with a 100 percent con dence. For each leaf
resting, a class probability is computed knowing the prediction of each tree of the
forest. Results are presented table 1. Surprisingly, it shows better performances
with pseudoscan images than scan images. This behavior has been observed by a
quite important number of team. Thus, we suppose it is mostly due to the species
presence in each type of images, as some class are more di cult to predict than
others.</p>
      <p>In total, 33 runs were submitted by 11 groups. We achieved 4th place for
pseudoscan images and the 5th place for scan and natural images. As a team, we
rank third team for photographs. This last result is a good performance, since we
concentrated our e ort on photographs and we used a poor human supervision.
6</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusions</title>
      <p>The results we obtained are satisfactory enough to validate our general approach.
The investigation of morphological features that model explicitly the criteria
used by botanists to identify trees proves to be a convincing way to treat the
problem of plant identi cation. The role of segmentation still seems prevailing,
and the performance on photographs, though being better than the average,
underlines its importance.</p>
      <p>A rst implementation of our identi cation system on mobile devices has been
engaged, with promising results, but not on as many species, leaving notably
aside compound leaves. With the number of potential classes increasing, it will
become a necessity to reduce the scope of the search, by making advantage
of the GPS system that now exists in every smartphone. Fusing geographical
information together with image based descriptors is the biggest challenge of
our futur work. And knowing in advance which species are likely to be found in
the geographical area where the user stands would be a decisive step towards a
truly reliable identi cation.</p>
      <p>The other prospect made possible by the use of high level descriptors will be
the possibility of putting words on the extracted information. Explaining what
the system understands and rendering it in a comprehensible form is a way to
teach a non-specialist user what to look for, and could be used to implicate
the user in the decision process by his intuitive corrections. All in all, such an
explicative and interactive application would constitute a way not only to help
recognize a plant, but to teach and transmit a rare knowledge.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>P.</given-names>
            <surname>Belhumeur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Kress</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ling</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ramamoorthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sheorey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>White</surname>
          </string-name>
          , and
          <string-name>
            <surname>L. Zhang.</surname>
          </string-name>
          <article-title>Searching the world's herbaria: A system for visual identication of plant species</article-title>
          .
          <source>In European Conference on Computer Vision</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          .
          <article-title>Classi cation and regression trees</article-title>
          .
          <source>Chapman &amp; Hall/CRC</source>
          ,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          .
          <article-title>Random forests</article-title>
          .
          <source>Machine learning</source>
          ,
          <volume>45</volume>
          (
          <issue>1</issue>
          ):5{
          <fpage>32</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Carlos</given-names>
            <surname>Caballero</surname>
          </string-name>
          and
          <string-name>
            <given-names>M. Carmen</given-names>
            <surname>Aranda</surname>
          </string-name>
          .
          <article-title>Plant species identi cation using leaf image retrieval</article-title>
          .
          <source>In Proceedings of the ACM International Conference on Image and Video Retrieval</source>
          ,
          <string-name>
            <surname>CIVR</surname>
          </string-name>
          <year>2010</year>
          , pages
          <fpage>327</fpage>
          {
          <fpage>334</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>G.</given-names>
            <surname>Cerutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tougne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mille</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vacavant</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Coquin</surname>
          </string-name>
          .
          <article-title>Guiding active contours for tree leaf segmentation and identi cation</article-title>
          .
          <source>In CLEF (Notebook Papers/Labs/Workshop)</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Arthur</surname>
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dempster</surname>
          </string-name>
          .
          <article-title>A generalization of bayesian inference</article-title>
          .
          <source>Journal of the Royal Statistical Society</source>
          ,
          <volume>30</volume>
          :
          <fpage>205</fpage>
          {
          <fpage>247</fpage>
          ,
          <year>1968</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>R.</surname>
          </string-name>
          <article-title>D az-Uriarte and</article-title>
          <string-name>
            <surname>S. De Andres</surname>
          </string-name>
          .
          <article-title>Gene selection and classi cation of microarray data using random forest</article-title>
          .
          <source>BMC bioinformatics</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ):
          <fpage>3</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. H. Goeau,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          , I. Yahiaoui,
          <string-name>
            <given-names>N.</given-names>
            <surname>Boujemaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Barthelemy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Molino</surname>
          </string-name>
          .
          <article-title>The imageclef 2012 plant identi cation task</article-title>
          .
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>B.</given-names>
            <surname>Goldstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hubbard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cutler</surname>
          </string-name>
          , and
          <string-name>
            <surname>L. Barcellos.</surname>
          </string-name>
          <article-title>An application of Random Forests to a genome-wide association dataset: Methodological considerations &amp; new ndings</article-title>
          .
          <source>BMC genetics</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <fpage>49</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>A</surname>
            .-
            <given-names>G.</given-names>
            Manh, G.
          </string-name>
          <string-name>
            <surname>Rabatel</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Assemat</surname>
            , and
            <given-names>M.-J.</given-names>
          </string-name>
          <string-name>
            <surname>Aldon</surname>
          </string-name>
          .
          <article-title>Weed leaf image segmentation by deformable templates</article-title>
          .
          <source>Journal of agricultural engineering research</source>
          ,
          <volume>80</volume>
          (
          <issue>2</issue>
          ):
          <volume>139</volume>
          {
          <fpage>146</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>F.</given-names>
            <surname>Mokhtarian</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Abbasi</surname>
          </string-name>
          .
          <article-title>Matching shapes with self-intersections: Application to leaf classi cation</article-title>
          .
          <source>IEEE Transactions on Image Processing</source>
          ,
          <volume>13</volume>
          (
          <issue>5</issue>
          ):
          <volume>653</volume>
          {
          <fpage>661</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>F.</given-names>
            <surname>Mokhtarian</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.K.</given-names>
            <surname>Mackworth</surname>
          </string-name>
          .
          <article-title>A theory of multiscale, curvature-based shape representation for planar curves</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>14</volume>
          :
          <fpage>789</fpage>
          {
          <fpage>805</fpage>
          ,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Pei</surname>
          </string-name>
          and
          <string-name>
            <given-names>C. N.</given-names>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>The detection of dominant points on digital curves by scale-space ltering</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>25</volume>
          (
          <issue>11</issue>
          ):
          <volume>1307</volume>
          {
          <fpage>1314</fpage>
          ,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>J.</given-names>
            <surname>Pooja</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Jonathan</surname>
          </string-name>
          .
          <article-title>Automatic structure classi cation of small proteins using random forest</article-title>
          .
          <source>BMC Bioinformatics</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <fpage>364</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>A.</given-names>
            <surname>Prasad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Iverson</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Liaw</surname>
          </string-name>
          .
          <article-title>Newer classi cation and regression tree techniques: bagging and random forests for ecological prediction</article-title>
          .
          <source>Ecosystems</source>
          ,
          <volume>9</volume>
          (
          <issue>2</issue>
          ):
          <volume>181</volume>
          {
          <fpage>199</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <given-names>F.</given-names>
            <surname>Schro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Criminisi</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          .
          <article-title>Object class segmentation using random forests</article-title>
          .
          <source>In Proceedings of the 19th British Machine Vision Conference (BMVC 08)</source>
          , Lake, UK,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>Glenn</given-names>
            <surname>Shafer</surname>
          </string-name>
          .
          <source>A Mathematical Theory of Evidence</source>
          . Princeton University Press,
          <year>1976</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>C. H. Teh</surname>
            and
            <given-names>R. T.</given-names>
          </string-name>
          <string-name>
            <surname>Chin</surname>
          </string-name>
          .
          <article-title>On the detection of dominant points on digital curves</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>11</volume>
          ,
          <year>1989</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>C.-H. Teng</surname>
            , Y.-T. Kuo, and
            <given-names>Y.-S.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          .
          <article-title>Leaf segmentation, its 3d position estimation and leaf classi cation from a few images with very close viewpoints</article-title>
          .
          <source>In Proceedings of the 6th International Conference on Image Analysis and Recognition, ICIAR '09</source>
          , pages
          <fpage>937</fpage>
          {
          <fpage>946</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <given-names>X.F.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.S.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.X.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Heutte</surname>
          </string-name>
          .
          <article-title>Classi cation of plant leaf images with complicated background</article-title>
          .
          <source>Applied Mathematics and Computation</source>
          ,
          <volume>205</volume>
          (
          <issue>2</issue>
          ):
          <volume>916</volume>
          {
          <fpage>926</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>