<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Late Information Fusion for Multi-modality Plant Species Identi cation?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Guillaume Cerutti</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Laure Tougne</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Celine Sacca</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thierry Joliveau</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pierre-Olivier Mazagol</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Didier Coquin</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Antoine Vacavant</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Clermont Universite , Universite d'Auvergne, ISIT</institution>
          ,
          <addr-line>F-63001, Clermont-Ferrand</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LISTIC</institution>
          ,
          <addr-line>Domaine Universitaire, F-74944, Annecy le Vieux</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universite Lyon 2</institution>
          ,
          <addr-line>LIRIS, UMR5205, F-69676</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Universite de Lyon</institution>
          ,
          <addr-line>CNRS</addr-line>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Universite de Saint Etienne</institution>
          ,
          <addr-line>EVS, UMR5600, F-42000, Saint Etienne</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This article presents the participation of the ReVeS project to the ImageCLEF 2013 Plant Identi cation challenge. Our primary target being tree leaves, some extra e ort had to be done this year to process images containing other plant organs. The proposed method tries to bene t from the presence of multiple sources of information for a same individual through the introduction of a late fusion system based on the decisions of classi ers for the di erent modalities. It also presents a way to incorporate the geographical information in the determination of the species by estimating their plausibility at the considered location. While maintaining its performance on leaf images (ranking 3rd on natural images and 4th on plain backgrounds) our team performed honorably on the brand new modalities with a 6th position.</p>
      </abstract>
      <kwd-group>
        <kwd>plant identi cation</kwd>
        <kwd>con dence fusion</kwd>
        <kwd>biogeographical information</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Considering the identi cation of a plant species, the study of its leaves is a
common path, and undoubtedly the most tting for an automatic system, given the
relative visual stability of the planar, slowly growing objects that are leaves.
However, they are not the organs on which was built the classi cation of the Plantae
kingdom, which considers the temporal proximity of the species as a baseline
grouping criterion. Flowers and fruits, representative of the plant's intrinsic
reproduction properties are the most evident witnesses of this proximity, and have
been used since the earliest classi cation works [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] as key features de ning the
species. Nevertheless, they constitute much touchier objects for automatic
recognition, given their 3-dimensional complex structure, their high variability and
short life cycles.
      </p>
      <p>
        The main objective of the ReVeS project is to build a tree species identi
cation system based on an explicit botanical description of leaf images. Unlike for
the previous editions, the challenge laid by this year's identi cation task [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] was
quite distant from this initial perspective, and required us to address di erent
objects that are not suitable for the same analytic shape processing. The
literature still provides numerous examples of methods dealing with ower [
        <xref ref-type="bibr" rid="ref13 ref16">13,16</xref>
        ],
bark [
        <xref ref-type="bibr" rid="ref19 ref6 ref9">19,9,6</xref>
        ] and even fruit [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] images, using generally statistical descriptions
? This work has been supported by the French National Agency for Research with the
reference ANR-10-CORD-005 (REVES project).
and rarely explicit modelling. Considering that most of the 5092 test images
could be grouped by individual, producing sets of images of di erent organs
for the same plant, our work focused especially on a strategy to combine those
various sources of information, and on a way to make use of the GPS location
provided to enhance the performance of the classi cation.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Image Processing and Feature Extraction</title>
      <p>
        This year's database approximately doubled in size compared to last year, and
it was separated into 2 big categories :
{ Sheet As Background images, containing only leaves, and thus
combining the categories Scan and Pseudoscan of previous editions [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], with the
di erence of containing some images that have previously fallen in the
Photograph category, photographs of leaves on a plain but sometimes dark
background.
{ Natural Background images, being photographs of plant organs (or
entire plants) in their natural environment. Only one type of content (Leaf,
Flower, Fruit, Stem or Entire) is given for each image, and no distinction
is made between images containing only one whole organ, part of an organ,
several organs more or less in focus, or even di erent types of organs.
      </p>
      <p>Among the 20985 images forming the Train database, we could nd 250
plant species, but a distinction can be made between tree species (127 generally
ligneous, taller than 1.2 m plant species) that were the ones considered in the
previous editions and the ones represented in the SheetAsBackground category,
and the 123 rather herbaceous remaining species. This distinction is important
for the identi cation, the tree species presenting for example bark in the Stem
images, that may be analyzed in a speci c way, more accurately than a generic
stem image.</p>
      <p>The Test database contained 5092 images with the same information on the
content, but with no indication whatsoever whether the species was a tree or
not, which forced us to make assumptions and selections for the good running
of our methods.
2.1</p>
      <p>Manual Image Formatting
Before any actual processing, we had to perform some transformations on the
Natural Backgroung images to convert them into suitable entry points for our
algorithm. As mentioned previously, our objective is to develop a mobile
application for identi cation, which allows a certain part of interaction with the
user, by giving guidelines for the picture taking or directly asking for help in the
image processing phase though the tactile interface. The procedure we applied
on the images simply simulates what we would have asked of the supposed user
of the application.</p>
      <p>Most of the Leaf, Fruit and Flower images contained more than one example
of the desired object. A rst selection was done by placing all the images for
which no object big and in focus enough could be found in the Entire category,
and keeping the others. We subsequently cut those images to a region of interest
containing only one object, and rotated them (90 or 180 rotations only) so that
the resulting images correspond to what input a user asked to take a photograph
of one, vertically oriented, organ would have given. A typical example of this
process can be found on the Figure 1 (b).</p>
      <p>In addition to this rst manipulation, we drew on top of each of the obtained
images a rough mark inside the object that is used as an initialization of our
segmentation algorithms. Note that in the case of leaves where the methods are
the most developed, a distinction is made between simple and compound leaves
: in the rst case, one single mark is required whereas we require three distinct
marks in three of the top lea ets for second type of leaves. An example of such
quick colouring, as it would be done with the nger on a mobile device, is shown
in Figure 1 (c).</p>
      <p>For Stem images, where no segmentation process was considered, we simply
rotated and cut the images to obtain a vertical view containing only bark pixels.
The images clearly corresponding to non-tree species were discarded to the Entire
category, as well as the non-tree leaves and fruits. Unfortunately, we did not have
the time to do anything with the images of this extended category.
2.2</p>
      <p>
        Leaf Understanding
Leaf images were processed with almost the same method as last year [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] with
the idea of introducing some modelling of the leaf. This is a purpose that has
proven to work quite e ciently [
        <xref ref-type="bibr" rid="ref1 ref20">1,20</xref>
        ]. The goal here is to represent faithfully
the botanical properties of the leaf, in an attempt to model the analytic
determination process used by botanists. For the recall, a leaf shape model is used
as both a pre-segmentation and a description of the global shape [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and after
the segmentation by a constrained active contour, descriptors accounting for the
margin [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], base and apex shapes are computed.
      </p>
      <p>
        The main di erences concern the processing of compound leaves in which we
try to describe all the lea ets at the same time. This is achieved by estimating
the number and location of the lea ets, and jointly modelling their global shape,
using highly variable deformable templates [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and then segmenting each lea et
independently with multiple region-based active contours. Figure 2 illustrates
the running of this new process.
      </p>
      <p>Another improvement lies in the representation of the margin that now uses a
string-like structure to keep track of the spatial repartition of teeth along the leaf
contour. When our former descriptors used aggregative methods (computation
of histograms, or averaging of curvature-scale measures) this new description is
more discriminant as it can represent di erent levels of dentition or di erent
orderings of teeth.
2.3</p>
      <p>Fruit, Bark and Flower Image Analysis
The methods we used on the other organs were not nearly as dedicated and
thought through as the ones elaborated for leaves, as they are not our main
concern. We made the decision to rely on statistical shape descriptors and on
vocabulary learning to produce color, texture and interest point based features.
The only particularity came from the fact that in the case of Flower and Fruit
images, a segmentation had to be performed rst for these descriptors to be
extracted correctly.</p>
      <p>
        Fruits and owers may be easier to isolate in an image given how their
colors stand out, but they still require some advanced processing [
        <xref ref-type="bibr" rid="ref16 ref18">16,18</xref>
        ] or some
modelling [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] to be segmented. The segmentation algorithm we used is derived
from the one implemented for leaves. However, given the impossibility to model
e ciently the shapes of fruits and owers in the images, it relies only on the
color dissimilarity part of the method. Starting from the user-made colouring,
the average expected color is propagated throughout the image, leading to a
map of the dissimilarity of image pixels to a local color model for the organ [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
This dissimilarity map is then simply thresholded to produce a segmentation of
the desired object. Some results of this algorithm are shown in Figure 3.
      </p>
      <sec id="sec-2-1">
        <title>Bagging features</title>
        <p>
          To represent the color information, we used a bag-of-colors-like feature based
on an organ-speci c color vocabulary, much like what can be found in other
works considering owers [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. This vocabulary was learned in a rst phase by
clustering the L*a*b* colors encountered in the considered (segmented) regions
of all images into a prede ned number of reference colors. Then the pixels for a
given image are associated with one word of the color vocabulary, and the
occurrences of each word are counted to form a histogram, later normalized by the
size of the region. This generates xed-sized, comparable, and even averageable
features, that are displayed in Figure 4. Such color features were extracted only
for Fruit and Flower images, with respective vocabulary sizes of 100 and 200
color words.
        </p>
        <p>
          The same idea was adopted for texture description keeping for each pixel in
the region, instead of a 3-channel L*a*b* vector, a vector of 30 values
corresponding to the responses obtained by ltering the image with the commonly
employed Gabor wavelets [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] of 6 di erent orientations and 5 di erent sizes.
Then again a vocabulary is learned based on these features and bags of textures
used to describe the image. This was performed on Stem and Fruit images, with
a vocabulary size of 500 for each.
        </p>
        <p>Finally, we computed the classical bag of visual words by repeating the same
procedure again, this time on SURF descriptors obtained on detected keypoints
(limited to 500 per image, for computational reasons). The only di erence is that
the histograms are not normalized by the number of keypoints extracted on the
image, otherwise it is in every respect similar. These bag of words were used on
Stem images only, with a vocabulary of 1000 visual words.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Shape descriptors</title>
        <p>To capture the shape properties of the segmented organs, we decided to rely
on well established, robust shape descriptors, commonly used in the literature
for object recognition purposes, namely :
{ Centered moments (8 values)
{ Eccentricity (1 value)
{ Hu moments (7 values)
{ Zernike moments (9 values)</p>
        <p>These values were computed on the binary segmentation images resulting from
the algorithm described above. Therefore, they were only extracted on Fruit and
Flower images.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Species Classi cation</title>
      <p>
        We chose to view the problem of species identi cation as a classi cation
problem rather than an image retrieval one, and subsequently trained classi ers to
perform the recognition. Unfortunately, the time dedicated to this task was too
short to reproduce the random forest classi cation that gave us the best results
last year [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and we had to rely on a more naive process. Furthermore, the
content of the database somehow pushed us to reconsider the decision making, with
the obvious interest of combining the di erent views and organs of the same
plant individual to improve the accuracy.
3.1
      </p>
      <p>Naive Distance-based Classi cation
For each set of the aforementioned descriptors, and for each considered species,
we were able to model the distribution of the parameters by gaussians, evaluating
the means and standard deviations from all the examples. This was also what
we did for the histogram-like features that are maybe not the most suitable for
this type of approach. The only di erence concerns the strings describing leaf
margins, non-vectorial structures that had to be addressed di erently, simply by
estimating an average string per species.</p>
      <p>The set of considered species, on which the training was performed and that
are likely to appear in the classi cation results, was also di erent depending on
the organ :
{ Leaf (L) : 126 tree species, separated in simple-leaved and compound-leaved
{ Fruit (F ) : 44 tree species with enough examples
{ Flower (f ) : 139 species, essentially herbaceous
{ Bark (B) : 78 species, only ligneous
{ Entire plant (p) : No training, by lack of time</p>
      <p>To classify a new image, we simply evaluate the distances do(x; S) of the
descriptors extracted on it to the gaussian models of the retained species (So)
for the corresponding organ o. The metric we used for histogram-like features is
a simple L2 distance, for the string structures, a normalized Levenshtein, and
for the vector features a so-called ellipsoid distance, designed to take variability
into account, without distorting the parameter space as would the Mahalanobis
distance. That distance represents the distance of a point x to the surface of
the ellipsoid de ned by the center and the axes of the covariance matrix .
Using the Mahalanobis distance kx kM = (x ) 1(x )T , the metric we
compute can be written as :
kx
kE = kx
k2 max 1
kx
1
kM
; 0
(1)</p>
      <p>Each one of the ND sets of descriptors extracted on the image thus produces
a distance, that is then normalized by the average distance to the correct class
(learned for each feature during the training phase). Once those terms are on
the same scale, they are just summed, and the set of all considered species are
ordered according to this last measure. This way, for each image, we produce a
ranked list of species along with a distance value measuring their similarity.</p>
      <p>Individual-based Information Fusion
This ranking could be enough to give a suitable answer per image, but it is
much more interesting to try and make use of the di erent modalities available
for a same individual plant. The images corresponding to a same value of the
IndividualPlantID are actually di erent views of the same plant, taken at the
exact same location. Consequently, the answer to each of these images needs to
be the same, and every element that has to be taken into account to produce it
with as much accuracy as possible.</p>
      <p>To perform an e cient fusion of these various data sources, we used a weighted
sum of con dence measures obtained from the distances to species issuing from
the di erent classi ers. The fusion process is illustrated in Figure 5. An
individual may be associated to a certain number of images (up to 24) of di erent
organs. Each one having been processed independently by the according
classier, we retrieve for each of the no images of the organ o the distances do (1::no)
to the di erent species considered in the context of this organ.</p>
      <p>The distances are rst converted into con dence measures. Each distance
value being a sum of ND;o terms expected to be equal to 1 for the correct
species, we de ne the con dence score associated to the distance as following :
Co(x; S) =
8
&gt;&lt; exp
&gt;: 0
do(x; S) 2</p>
      <p>ND;o
if S 2 (So)
if S 2= (So)
(2)</p>
      <p>
        In the end, even if distances were computed for only a subset of all the
considered species, we now have for each image a set of con dences for each
and every species. The value is 0 for those that were not learned but they are
still present in the possible answer list. These con dence values all lie between
0 and 1 and the normalization of the distance by nD;o makes sure that they
constitute comparable measures even for di erent organs. However, the results
we obtained on the di erent modalities tended to show that some are more
reliable than others. With this in mind, the con dence measures are weighted
by an organ-wise coe cient wo derived from the classi cation rates we could
observe on the training base, in a common attempt of giving more importance
to reliable sources of information [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>Finally, we used an additive process on these weighted con dences, in order
not to penalize too much the species that were absent for a given organ. The
sum of weighted con dences is then normalized again by the sum of weights
to remain between 0 and 1. The nal con dence score for a species S is then
computed as :</p>
      <p>C(S) =</p>
      <p>Po2fL;f;F;Bg (Pin=o1 wo Co(xi; S))</p>
      <p>Po2fL;f;F;Bg (Pin=o1 wo)
(3)</p>
      <p>With these similarity measures, we can produce a new ranking of the species,
this time all the 250 of them, that will be the same for all the images of the
considered individual. The information available on each of the image is then
taken into account, with a degree of trust depending on the modality, and the
ordering that would have been obtained on one image only is then re ned and
hopefully improved by the content of the other associated images. Of course
in the case when there is only one image per individual, the order remains
unchanged, but when di erent organs are available, or di erent leaves of the
same plant, the species appearing regularly as close will be put up front.
3.3</p>
      <p>Introducing Biogeographical Knowledge
Geographical and environmental factors play a big part in the development (and
survival) of plants, and not all species are likely to be found under every
conditions. Knowing where a plant was observed, and what species may grow in that
particular location might constitute a great help for identi cation. With this
perspective, it was a core objective of our project to come up with methods to
bene t from the location (easily accessible through GPS on mobile devices, for
example) to enhance the identi cation by knowing in advance the plausibility of
species.</p>
      <p>
        In this respect, a large-scale work of digital cartography was undertaken
(focusing on the French metropolitan territory) to map some environmental
parameters that are related to the presence of species, with a resolution of 90 meters.
Indeed, variables such as temperature, altitude, precipitation, humidity,
exposition have a key in uence on the repartition of species. Along with these raw
explanatory variable, we also considered more global, integrative values
elaborated by botanists and phytosociologists :
{ Altitudinal zonation, or vegetation levels as de ned in phytogeography [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ],
9 zones for di erent altitudes, temperatures and orientations (modelled using
CNRS vegetation maps at scale 1:200,000)
{ GRECO (great ecological regions) from the French forest institute [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], 10
regions of homogeneous geological and environmental characteristics
identied by letters
      </p>
      <p>The images in the PlantView database were all referenced in terms of
longitude and latitude, and most of them issued from the french territory, making
them suitable to experiment this geographical approach. Unfortunately, there
was a big incertitude about the precision of this location, whether it was the
place where the image was actually acquired or simply uploaded. In the case
of scanned leaves, is it the location of the tree or the scanner that is supplied?
Therefore, trying to include local parameters like the raw measurements hardly
made any sense, and we chose to take only levels and regions into account. The
maps for those two variables can be seen in Figure 6.</p>
      <p>
        To derive the plausibility of nding a species at a given location, we did
not use the training base but included external botanical knowledge about the
species. Based on the repartition knowledge and maps one can nd in oras
[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], we associated for each region and each zone a presence value of the species.
However, this task was only performed for native tree species, leaving almost half
of the challenge scope empty of information. The presence levels were simply 0
or 1 for the vegetation zones but 0, 1, 2, 3, 4 and 5 (in increasing order of
plausibility) for the ecological regions.
      </p>
      <p>Then, for given values of the region index g and of the vegetation level l, it
is possible to compute a numerical value of plausibility for each species given its
theoretical repartition. Knowing the presence values of the species in the region
p(S; g) and the level p(S; l) we simply derive an additional plausibility score :
P (S; g; l) =
8 1 if p(S; g) = 5
&gt;
&gt;&gt;&gt; 0:95 if p(S; g) = 4
&gt;
&gt;
1 &lt;&gt;&gt; 0:9 if p(S; g) = 3
2 &gt; 0:8 if p(S; g) = 2
&gt;
&gt;&gt;&gt; 0:7 if p(S; g) = 1
&gt;
&gt;
&gt;: 0:5 if p(S; g) = 0
+
1 ( 1 if p(S; l) = 1
2
0:5 if p(S; l) = 0
(4)</p>
      <p>For visualisation purposes, we computed this plausibility for di erent species
all over the French territory, for which the values of vegetation levels and
ecological regions were mapped. The result can be seen as a repartition map of the
species, in which the places where the species is most plausible appear in darker
green, on those where it will certainly not be found in lighter green. Examples
of such maps are shown in Figure 7.</p>
      <p>For the species on which we had no prior repartition knowledge, we ensured
the geography would not have a big in uence on classi cation by setting the
P (S; l; g) to 1 for all values of l and g. This of course is not an optimal solution,
and actually favours those species, since they will at least be as plausible as the
other species, whatever the location.</p>
      <p>The way we insert the geographical knowledge into the classi cation process
is fairly simple. Given the coordinates (lon; lat) of the observation, we can extract
from the mapped data the values g(lon; lat) and l(lon; lat) of the corresponding
GRECO and level. Then for each species S the con dence score used to rank it
simply becomes :</p>
      <p>CG (S; lon; lat) = P (S; g(lon; lat); l(lon; lat)) C(S)
(5)</p>
      <p>The introduction of this geographical plausibility leaves unchanged the
condence score for the species that are perfectly adequate with the observation's
location. However, it will penalize the species which are not, making them drop
in the rankings and conversely letting the more plausible species make their way
up to the rst answers. It is a way to make a rearrangement of the top answers
without actually overshadowing the information extracted from the image. It
is in any case coherent with the idea that geography should not contradict the
evidence found in the content of the image, but simply provide a lighting on the
results of its analysis to make a better nal decision.
4
4.1</p>
    </sec>
    <sec id="sec-4">
      <title>Results and Comments</title>
      <p>Training Results
The classi cations of the di erent plant organs do not perform all the same,
the most e ective one naturally being the classi cation of leaves on plain
backgrounds. As a matter of fact, leaves are the organ for which most work has been
dedicated, and the methods used to process them were the most extensively
tested. It is then only natural that the experimental results makes them appear
clearly as the best performing modality.</p>
      <p>The Figure 8 sums up the experimental classi cation rates obtained for the 4
modalities we addressed. Those scores were measured on the Train database, in a
cross validation process (validation rates are displayed, with the rates obtained
on the training data left in lighter color). If the Leaf category (126 classes, 4
descriptor sets) clearly bene ts from the e orts put in the segmentation and the
botany-inspired description (nearly 85% of presence of the correct species in the
top 5 answers, up to 70% in the case of natural images), the other modalities
are not as far behind as expected.</p>
      <p>The Fruit category (44 classes, 3 descriptor sets) is the second best performing
(more than 50% at 5 answers) but this performance has to be moderated by the
fact the the number of classes is relatively small. On the other hand the Stem
category (78 classes, 2 descriptor sets) performs honorably (more than 40% at
5 answers) given the number of species and the rawness of the description used.
Same goes for the Flower modality (139 classes, 2 descriptors) for which it might
be interesting to add that most of the accuracy (37% at 5 answers) actually
comes from the simple color feature we used. All in all, these results underline
the limits of the descriptions (hastily) implemented, and in the case of fruits
and owers the di culties of the segmentation, given the low contribution of the
shape description. It is also crucial to point out the greater abstraction power
of the dedicated description created for leaves compared to the generic features
used on the other organs ; this is visible in the blatant gap between the curves
obtained on the training and validation sets.</p>
      <p>A very interesting point is to measure the contribution of geography in the
performance of the classi cation. To do so, we considered only leaf images from
tree species that are native in France, and on which we had the necessary
theoretical geogaphical information. The Figure 9 shows the comparison between
a classi cation where the geographical plausibility of the species is taken into
account and one where it is not. More than the actual rates we obtain, it is
clearly appearing that geographical knowledge leads to a better classi cation,
and simply improves the ranking of the correct species. The Figure 10 shows
this improvement by measuring the di erence between the rank of the correct
species in the classi cation without geographical information and with it, a
difference that goes in the right direction for a major share of examples for which
the ranking could still be improved. Even with unprecise location (obviously for
scanned images) the region and level prove to give enough information to re ne
the decision in a more accurate way.</p>
      <p>(a)
We submitted two distinct runs this year, both relying on the same visual
analysis process but di ering in the fusion phase. They both make use of the
IndidualPlantID eld, but the second run (LirisReVeS run2) additionally used the
longitude and latitude coordinates to compute the geographical plausibility of
the species and use it to rearrange the results of the image processing phase.</p>
      <p>Both runs were then mixed (visual + textual) runs, with an automatic
method on the SheetAsBackground category and manual initalizing on the
NaturalBackground category. To be exhaustive, the images remaining untreated, and
not belonging to a processed individual were submitted as specimens of Ginkgo
biloba, L. with an absolute con dence of 1.</p>
      <p>As expected, the SheetAsBackground category, containing almost only leaves,
is the one on which our leaf-focused algorithms gave the most satisfying results.
As shown in Figure 11, we reach a score of 41% bringing us in 4th place among
the 12 teams that have submitted runs.</p>
      <p>Concerning the NaturalBackground images, the results are with no surprise
lower, but our ranking on the Stem images (3rd place, yet with a score of 16%)
was unexpectedly quite interesting. Our low performance on fruits and owers
(respectively 8% and 10%) is a sign, if needed, that the representation of such
organs requires more dedicated descriptors, and would probably bene t from the
introduction of some kind of model to focus on the discriminant criteria from a
botanical point of view. And nally, our 3rd place on the Leaf images corresponds
to our previous rankings, with again the same limitations (only broad-leaved tree
species considered, some images out of our processing scope, learning performed
on plain background images) that might explain the rather low raw score value
of nearly 17 %.
Fig. 12. Results on the NaturalBackground images : 3rd on Leaf images (a), 6th on
Fruit images (b), 3rd on Stem images (c) and 7th on Flower images (d)</p>
      <p>Concerning geography, the run LirisReVeS run2 gave slightly better results
than the run LirisReVeS run1 from which the location information was completely
left out. But the di erence (almost 1% in every category) does not appear as
clearly as it did on native species, and it is hard to draw any conclusion. There
were however too many manifest aws in the context of this task to make this
a valid experiment for us. First, we had theoretical phytogeographical data at
hand for hardly half of the 250 species, which made the plausibility values we
computed only half-relevant. More importantly, because of the doubt about the
precise location of the observation, we were unable to use the local geographical
parameters that might be more accurate. The fact that many images come from
cities, where it is more di cult to assess whether the plant grew in its natural
environment, also made the problem a bit less clear. It is nevertheless an
encouraging step towards the use of geographical information to help the identi cation,
a direction that will in any case be further investigated.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>This year's task constituted a real new challenge for our leaf-oriented project,
and it came as no surprise that the generic, mechanical, methods we had the
time to implement on fruits, barks and owers were largely outperformed by the
now mature leaf analysis process put into practice in the Folia applicationx. It
was however a good chance to experiment with a late fusion mechanism which
seemed to bear fruit. It was also the opportunity for a rst testing of the novel
approach proposed to combine the image information with geographical
knowledge. The results of this last contribution are engaging, and we believe that with
an adequate scope of species and a greater certainty on location it may constitute
a real asset for eld plant identi cation.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>We would like to thank Tristan Coulange for his contribution in the hard task
of developing of competitive ower segmentation methods, as well as Marine
Aghadjanian, Nicolas Charrel, Nadya Dahir and Maxime Melinon for their
precious help in processing and analyzing thousands of fruit, bark and ower images.
x</p>
      <p>: https://itunes.apple.com/app/folia/id547650203</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>A.</given-names>
            <surname>Arora</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bagmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mishra</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhattacharya</surname>
          </string-name>
          .
          <article-title>A plant identication system using shape and morphological features on segmented lea ets</article-title>
          .
          <source>In CLEF (Notebook Papers/Labs/Workshop)</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>G.</given-names>
            <surname>Cerutti</surname>
          </string-name>
          , L. Antoine,
          <string-name>
            <given-names>Violaine</given-names>
            <surname>Tougne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Mille</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Julien</given-names>
            <surname>Valet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Coquin</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Vacavant</surname>
          </string-name>
          .
          <article-title>Reves participation - tree species classi cation using random forests and botanical features</article-title>
          .
          <source>In CLEF (Online Working Notes/Labs/Workshop)</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>G.</given-names>
            <surname>Cerutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tougne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Coquin</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Vacavant</surname>
          </string-name>
          .
          <article-title>Curvature-scale-based contour understanding for leaf margin shape recognition and species identi cation</article-title>
          .
          <source>In VISAPP 2013 - Proceedings of the International Conference on Computer Vision Theory and Applications</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>G.</given-names>
            <surname>Cerutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tougne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mille</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vacavant</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Coquin</surname>
          </string-name>
          .
          <article-title>A model-based approach for compound leaves understanding and identi cation</article-title>
          .
          <source>In IEEE International Conference on Image Processing</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>G.</given-names>
            <surname>Cerutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tougne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vacavant</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Coquin</surname>
          </string-name>
          .
          <article-title>A parametric active polygon for leaf segmentation and shape estimation</article-title>
          .
          <source>In 7th International Symposium on Visual Computing</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>S.</given-names>
            <surname>Fiel</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Sablatnig</surname>
          </string-name>
          .
          <article-title>Automated identication of tree species from images of the bark, leaves and needles</article-title>
          .
          <source>In Proceedings of the 16th Computer Vision Winter Workshop</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7. H. Goeau,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Boujemaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Barthelemy</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Molino</surname>
          </string-name>
          .
          <article-title>The imageclef 2013 plant identi cation task</article-title>
          .
          <source>In CLEF 2013 Working Notes</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. H. Goeau,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Boujemaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Barthelemy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Molino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Birnbaum</surname>
          </string-name>
          , E. Mouysset, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Picard</surname>
          </string-name>
          .
          <article-title>The clef 2012 plant images classi cation task</article-title>
          .
          <source>In CLEF 2012 Working Notes</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Z.-K. Huang</surname>
            ,
            <given-names>C.-H.</given-names>
          </string-name>
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>J.-X.</given-names>
          </string-name>
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>Z.-H.</given-names>
          </string-name>
          <string-name>
            <surname>Quan</surname>
            , and
            <given-names>S.-B.</given-names>
          </string-name>
          <string-name>
            <surname>Guo</surname>
          </string-name>
          .
          <article-title>Bark classi cation based on gabor lter features using rbpnn neural network</article-title>
          .
          <source>In International Conference on Neural Information Processing</source>
          , pages
          <volume>80</volume>
          {
          <fpage>87</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10. Inventaire Forestier National. Les
          <string-name>
            <surname>Sylvoecoregions (SER) de</surname>
          </string-name>
          <article-title>France metropolitaine</article-title>
          .
          <source>IGN</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>P.</given-names>
            <surname>Julve</surname>
          </string-name>
          . Carte phytogeographique de la France. Cahiers de Geographie Physique,
          <volume>13</volume>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. C. v. Linne.
          <article-title>Species plantarum : exhibentes plantas rite cognitas, ad genera relatas, cum di erentiis speci cis, nominibus trivialibus, synonymis selectis, locis natalibus, secundum systema sexuale digestas</article-title>
          .
          <source>Holmiae : Impensis Laurentii Salvii</source>
          ,
          <volume>1753</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>M.-E. Nilsback</surname>
            and
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Zisserman</surname>
          </string-name>
          .
          <article-title>Delving into the whorl of ower segmentation</article-title>
          .
          <source>In British Machine Vision Conference</source>
          , volume
          <volume>1</volume>
          , pages
          <fpage>570</fpage>
          {
          <fpage>579</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>J. Rameau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Mansion</surname>
            , G. Dume,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Timbal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lecointe</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dupont</surname>
          </string-name>
          , and R. Keller.
          <article-title>Flore forestiere francaise: Guide ecologique illustre</article-title>
          .
          <source>Institut pour le Developpement Forestier</source>
          ,
          <year>1989</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>A.</given-names>
            <surname>Rocha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Hauagge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wainer</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Goldenstein</surname>
          </string-name>
          .
          <article-title>Automatic fruit and vegetable classi cation from images</article-title>
          .
          <source>Computers and Electronics in Agriculture</source>
          ,
          <volume>70</volume>
          (
          <issue>1</issue>
          ):
          <volume>96</volume>
          {
          <fpage>104</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16. T. Saitoh and
          <string-name>
            <given-names>T.</given-names>
            <surname>Kaneko</surname>
          </string-name>
          .
          <article-title>Automatic recognition of blooming owers</article-title>
          .
          <source>In International Conference on Pattern Recognition</source>
          , volume
          <volume>1</volume>
          , pages
          <fpage>27</fpage>
          {
          <fpage>30</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          .
          <article-title>Con dence fusion</article-title>
          .
          <source>In Proceedings of ROSE'2004 (2004 IEEE Workshop on Robotic Sensing)</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18. L.
          <string-name>
            <surname>Suta</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Scuturuici</surname>
          </string-name>
          , V.
          <string-name>
            <surname>-M. Scuturici</surname>
            , and
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Miguet</surname>
          </string-name>
          . Pathos:
          <article-title>Part-based tree hierarchy for object segmentation</article-title>
          .
          <source>In International Conference on Computer Analysis of Images and Patterns (CAIP)</source>
          , aug
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19. Y.
          <string-name>
            <surname>-Y. Wan</surname>
            ,
            <given-names>J.-X.</given-names>
          </string-name>
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>D.-S.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            , and
            <given-names>Z.-R.</given-names>
          </string-name>
          <string-name>
            <surname>Chi</surname>
          </string-name>
          .
          <article-title>Bark texture feature extraction based on statistical texture analysis</article-title>
          .
          <source>In International Symposium on Intelligent Multimedia, Video and Speech Processing</source>
          , pages
          <volume>482</volume>
          {
          <fpage>485</fpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Yanikoglu</surname>
          </string-name>
          , E. Aptoula, and
          <string-name>
            <given-names>C.</given-names>
            <surname>Tirkaz</surname>
          </string-name>
          .
          <article-title>Sabanci-okan system at imageclef 2012: Combining features and classi ers for plant identi cation</article-title>
          .
          <source>In CLEF (Online Working Notes/Labs/Workshop)</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>