<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The ImageCLEF 2012 Plant identi cation Task</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Herve Goeau</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pierre Bonnet</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexis Joly</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Itheri Yahiaoui</string-name>
          <email>itheri.yahiaoui@univ-reims.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniel Barthelemy</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nozha Boujemaa</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jean-Francois Molino</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>INRIA</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>IMEDIA</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ZENITH teams</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>name.surname@inria.fr</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>http://www-rocq.inria.fr/imedia/</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>http://www-sop.inria.fr/teams/zenith/</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CIRAD</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>UMR AMAP</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>pierre.bonnet@cirad.fr</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>http://amap.cirad.fr</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CIRAD</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>BIOS Direction</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>UMR AMAP</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>daniel.barthelemy@cirad.fr</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>http://amap.cirad.fr/fr/index.php</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>UMR AMAP</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>France</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>jean-francois.molino@ird.fr</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>http://amap.cirad.fr</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Laboratoire CReSTIC, Universite de Reims</institution>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The ImageCLEF's plant identi cation task provides a testbed for the system-oriented evaluation of plant identi cation, more precisely on the 126 tree species identi cation based on leaf images. Three types of image content are considered: Scan, Scan-like (leaf photographs with a white uniform background), and Photograph (unconstrained leaf with natural background). The main originality of this data is that it was speci cally built through a citizen sciences initiative conducted by Tela Botanica, a French social network of amateur and expert botanists. This makes the task closer to the conditions of a real-world application. This overview presents more precisely the resources and assessments of task, summarizes the retrieval approaches employed by the participating groups, and provides an analysis of the main evaluation results. With a total of eleven groups from eight countries and with a total of 30 runs submitted, involving distinct and original methods, this second year pilot task con rms Image Retrieval community interest for biodiversity and botany, and highlights further challenging studies in plant identi cation.</p>
      </abstract>
      <kwd-group>
        <kwd>ImageCLEF</kwd>
        <kwd>plant</kwd>
        <kwd>leaves</kwd>
        <kwd>images</kwd>
        <kwd>collection</kwd>
        <kwd>identi cation</kwd>
        <kwd>classi cation</kwd>
        <kwd>evaluation</kwd>
        <kwd>benchmark</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Convergence of multidisciplinary research is more and more considered as the
next big thing to answer profound challenges of humanity related to health,
biodiversity or sustainable energy. The integration of life sciences and computer
sciences has a major role to play towards managing and analyzing cross-disciplinary
scienti c data at a global scale. More speci cally, building accurate knowledge
of the identity, geographic distribution and uses of plants is essential if
agricultural development is to be successful and biodiversity is to be conserved.
Unfortunately, such basic information is often only partially available for
professional stakeholders, teachers, scientists and citizens, and often incomplete for
ecosystems that possess the highest plant diversity. A noticeable consequence,
expressed as the taxonomic gap, is that identifying plant species is usually
impossible for the general public, and often a di cult task for professionals, such
as farmers or wood exploiters and even for the botanists themselves. The only
way to overcome this problem is to speed up the collection and integration of
raw observation data, while simultaneously providing to potential users an easy
and e cient access to this botanical knowledge. In this context, content-based
visual identi cation of plant's images is considered as one of the most promising
solution to help bridging the taxonomic gap. Evaluating recent advances of the
IR community on this challenging task is therefore an important issue.
This paper presents the plant identi cation task that was organized for the
second year running within ImageCLEF 20126 dedicated to the system-oriented
evaluation of visual based plant identi cation. The task was again focused on
tree species identi cation based on leaf images, but with more species (126
instead of 70) which is an important step towards covering the entire ora of a
given region. The task was more related to to a retrieval task instead of a pure
classi cation task in order to consider a ranked list of retrieved species rather
than a single brute determination. Visual content was being the main available
information but with additional information including contextual meta-data
(author, date, locality name and geotag, names at di erent taxonomic ranks) and
some EXIF data. Three types of image content were considered: leaf scans, leaf
photographs with a white uniform background (referred as scan-like pictures)
and unconstrained leaf's photographs acquired on trees with natural background.
The main originality of this data is that it was speci cally built through a
citizen sciences initiative conducted by Telabotanica7, a French social network of
amateur and expert botanists. This makes the task closer to the conditions of
a real-world application: (i) leaves of the same species are coming from distinct
trees living in distinct areas (ii) pictures and scans are taken by di erent users
that might not used the same protocol to collect the leaves and/or acquire the
images (iii) pictures and scans are taken at di erent periods in the year.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Task resources</title>
      <p>Building e ective computer vision and machine learning techniques is not the
only side of the taxonomic gap problem. Speeding-up the collection of raw
observation data is clearly another crucial one. The most promising approach in
that way is to build real-world collaborative systems allowing any user to enrich
the global visual botanical knowledge [15]. To build the evaluation data of
ImageCLEF plant identi cation task, we therefore set up a citizen science project
around the identi cation of common woody species covering the Metropolitan
French territory. This was done in collaboration with TelaBotanica social
network and with researchers specialized in computational botany.</p>
      <sec id="sec-2-1">
        <title>6 http://www.imageclef.org/2012</title>
      </sec>
      <sec id="sec-2-2">
        <title>7 http://www.tela-botanica.org/</title>
        <p>Technically, images and associated tags were collected through a crowd-sourcing
web application [15] and were all validated by expert botanists. Several cycles
of such collaborative data collection and taxonomical validation occurred.</p>
        <p>Scans of leaves were rst collected over the two summers 2009 and 2010
thanks to the work of active contributors from TelaBotanica social network. The
idea of collecting only scans during these two seasons periods was to initialize
training data with limited noisy background and to focus on plant variability
rather than mixed plant and view conditions variability. This allowed to collect
2228 scans over 55 species. A public version of the web application8 was then
opened in October 2010 and additional data were collected up to March 2011.
The new collected images were either scans, or photographs with uniform
background (referred as scan-like photos), or unconstrained photographs with natural
background. They involved besides 15 new species from the previous set of 55
species. Since April 2011 botanists from TelaBotanica contribute regularly every
month on more and more tree species (174 on leaves at the time of writing), on
more localities all over France and neighbouring countries, with more growing
stages with spring leaves, mature summer leaves, dry autumn leaves and
evergreen leaves. This nonstop data collecting over the months introduce slowly
visual and morphological variabilities.</p>
        <p>The Pl@ntLeaves dataset used within ImageCLEF2012 nally contained 11572
images: 6630 scans, 2726 scan-like photos and 2216 photographs (see Figure 1).</p>
        <sec id="sec-2-2-1">
          <title>Scan</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>Scan-like</title>
        </sec>
        <sec id="sec-2-2-3">
          <title>Photograph Fig. 1. The 3 image types illustrated with the same species Celtis australis L.</title>
          <p>{ Date date and time of shot or scan
{ Type (acquisition type: Scan, Scan-like or Photograph)
{ Content type (see gure3):</p>
          <p>Leaf (tagged by default for all images, supposing that a picture is focused
one single leaf),</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>8 it is closed now, but a newer a application can be found at http://identify.plantnet</title>
        <p>project.org/en/base/plantscan</p>
        <p>Picked leaf most of the time mature leaf - not dry - on the oor,
Upper side and Lower side involving generally 2 images from one leaf,
Branch for some pictures containing enumerable leaves,</p>
        <p>
          Leafage.
{ Taxon full taxon name according the botanical database[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ](Regnum, Class,
        </p>
        <p>Subclass, Superorder, Order, Family, Genus, Species)
{ VernacularNames English commun name
{ Author name of the author of the picture
{ Organization name of the organization of the author
{ Locality locality name (a district or a country division or a region)
{ GPSLocality GPS coordinates of the observation
These meta-data are stored in independent xml les, one for each image. Figure
2 displays an example image with its associated xml data.</p>
        <p>
          Additional but partial meta-data information can be found in the image's EXIF,
and might include the camera or the scanner model, the image resolution and
dimension, the optical parameters, the white balance, the light measures, etc.
The main originality of Pl@ntLeaves compared to previous leaf datasets, such as
the Swedish dataset [22], the ICL dataset [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] or the Smithsonian one [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], is that
it was built in a collaborative manner through a citizen sciences initiative. This
makes it closer to the conditions of a real-world application: (i) leaves of the
same species are coming from distinct trees living in distinct areas (ii) pictures
and scans are taken by di erent users that might not used the same protocol to
collect the leaves and/or acquire the images (iii) pictures and scans are taken at
di erent periods in the year. Intra-species visual variability and view conditions
variability are therefore more stressed-out which makes the identi cation more
        </p>
        <sec id="sec-2-3-1">
          <title>Scan</title>
        </sec>
        <sec id="sec-2-3-2">
          <title>Upper side Lower side</title>
        </sec>
        <sec id="sec-2-3-3">
          <title>Scan-like</title>
        </sec>
        <sec id="sec-2-3-4">
          <title>Branch</title>
        </sec>
        <sec id="sec-2-3-5">
          <title>Picked leaf Photograph Branch</title>
        </sec>
        <sec id="sec-2-3-6">
          <title>Leafage</title>
          <p>realistic but more complex. Figures 4 to 11 provide illustrations of the
intraspecies visual variability over several criteria including growing stages, colors,
global shape, margin appearance, number and relative positions of lea ets,
compound leaf structures and lobe variations. On the other side, Figure 12 illustrates
the light re ection and shadows variations of Scan-like photos. It shows that this
acquisition protocol is actually very di erent than pure scans. Both share the
property of a limited noisy background but Scan-like photos are more complex
due to the lighting conditions variability ( ash, sunny weather, etc.) and the
unatness of leaves. Finally, the variability of unconstrained photographs acquired
on the tree and with natural background is de nitely a much more challenging
issue as illustrated in Figure 13.</p>
        </sec>
        <sec id="sec-2-3-7">
          <title>Winter 2011 Fig. 4. Four leaves from distinct plants of one same species (Platanus x hispanica Wild) collected at di erent growing stages.</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Task description</title>
      <p>Training and Test data
The task was evaluated as a supervised classi cation problem with tree species
used as class labels. A part of Pl@ntLeaves dataset was provided as training
data whereas the remaining part was used later as test data. The two subsets
were created with the respect to the individual plants: pictures of leaves
belonging to the same individual tree cannot be split across training and test data.
This prevents identifying the species of a given tree thanks to its own leaves
and that makes the task more realistic. In a real world application, it is indeed
much unlikely that a user tries to identify a tree that is already present in the
training data. The training subset was built rst by integrating the whole
ImageCLEF2011 plant task dataset. New pictures were then added to the training
and test data, by selecting randomly new individual plants, in a such way to
reach so far as possible a ratio around 2 individual plants in the training data
for 1 individual plant in the test data. Detailed statistics of the composition of
the training and test data are provided in 1.
The goal of the task was to associate the correct tree species to each test image.
Each participant was allowed to submit up to 3 runs built from di erent methods.
Compared to last year the task was more related to plant species retrieval than
a pure classi cation problem. This year evaluation metric was slightly modi ed
in order to consider a ranked list of retrieved species rather than a single brute
determination. Each test image was attributed with a score between 1 and 0
equals to the inverse of the rank of the correct species. An average normalized
score is then computed on all test images. A simple mean on all test images would
indeed introduce some bias with regard to a real world identi cation system.
Indeed, we remind that the Pl@ntLeaves dataset was built in a collaborative
manner. So that few contributors might have provided much more pictures than
many other contributors who provided few. Since we want to evaluate the ability
of a system to provide correct answers to all users, we rather measure the mean of
the average classi cation rate per author. Furthermore, some authors sometimes
provided many pictures of the same individual plant (to enrich training data
with less e orts). Since we want to evaluate the ability of a system to provide
the correct answer based on a single plant observation, we also decided to average
the classi cation rate on each individual plant. Finally, our primary metric was
de ned as the following average classi cation score S:</p>
      <p>S =
1 XU 1 XPu 1 NXu;p su;p;n
U u=1 Pu p=1 Nu;p n=1
(1)
U : number of users (who have at least one image in the test data)
Pu : number of individual plants observed by the u-th user
Nu;p : number of pictures taken from the p-th plant observed by the u-th user
su;p;n : score between 1 and 0 equals to the inverse of the rank of the correct
species for the n-th picture taken from the p-th plant observed by the u-th user</p>
      <p>It is important to notice that while making the task more realistic, the
normalized classi cation score also makes it more di cult. Indeed, it works as if a
bias was introduced between the statistics of the training data and the one of the
test data. It highlights the fact that bias-robust machine learning and computer
vision methods should be preferred to train such real-world collaborative data.
Finally, to isolate and evaluate the impact of the image acquisition type (Scan,
Scan-like, Photogragh), a normalized classi cation score S was computed for each
type separately. Participants were therefore allowed to train distinct classi ers,
use di erent training subsets or use distinct methods for each data type.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Participants and techniques</title>
      <p>A total of 11 groups submitted 30 runs, which con rms the successful
participation rate of the previous year task on a new topic. Participants are mainly
academics, specialized in computer vision and multimedia information retrieval,
coming from all around the world: Australia (1), Brazil (1), China (1), France (3),
Germany (2), India (1), Italy (1) and Turkey (1). We list below the participants
and give a brief overview of the techniques they used to run the plant identi
cation task. We remind here that ImageCLEF benchmark is a system-oriented
evaluation and not a formal evaluation of the underlying methods. Readers
interested by the scienti c and technical details of any of these methods should
refer to the ImageCLEF 2012 working note of each participant.</p>
      <p>ArTe-Lab (1 run) [14] In their preliminary evaluations, these participants
experiment several features and classi cation methods. They were guided by the
wish to design a mobile application and thus aimed to have a good compromise
between accuracy and computing time and cost. They retained the Pyramid of
Histograms of Orientation Gradients (PHOG) and a variant of the HAAR
descriptor evaluated as the most stable and robust features in their preliminary
tests. They chose a multi-class probability estimation based on Support Vector
Machine (SVM) classi ers with a one-vs-one, or pairwise comparisons, strategy.
They choose to train a single SVM model for all 3 categories of images, without
distinguish Scan, Scan-like and Photograph.</p>
      <p>
        Brainsignals (1 run) [16] The aim of these participants was to focus less
on possible computer-vision related techniques (especially without costly image
preprocessing and feature extraction), and the much as possible to deal with
machine learning techniques. First, they binarized the images using the Otsu
algorithm, in order to extract two features: global shape features (from lateral
projections), and local features (histograms of content types of small rectangular
patches of bits, with and without sub-sampling). A Random Forest Classi er[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
was then used with 100 trees trained with several cross-validation steps on the
whole training data without distinguish Scan, Scan-like and Photograph
categories.
      </p>
      <p>BTU DBIS (4 runs) [10] This team used a rather non-typical approach not
relying on classi cation techniques. Logical combinations of low-level features are
expressed in a query language, the Commuting Quantum Query Language[21],
and used to assess a document's similarity to a species. The principle is to
combine similarity predicates as found in information retrieval and relational
predicates common in databases. They tested 4 distinct methods without
distinguishing the 3 categories Scan, Scan-like and Photograph. Their preliminary
evaluations suggested to use MPEG7 Color Structure Descriptor eventually
combined with GPS information. The rst two runs explored two Query By Example
based approaches, with or without GPS information. As a third approach, they
combined the second approach with a Top-k Clustering (or "K-methoid")
techniques which exploits relationships within the top-k results. The last run used
a pure clustering image approach with GPS information, including the whole
train and test datasets.</p>
      <p>
        IFSC USP (3 runs) [11] Mainly focused on automatic shape boundary
analysis, these participants adapted their approach according the image content and
obtained the best results for the Photograph with a semi-automatic approach.
Full-automatic approaches were used for Scan and Scan-like, but also on
Photograph for one run. They started by an automatic leaf contour extraction with
Otsu's method on Scan and Scan-like, while they used a k-means clustering
algorithm in the RGB colorspace for Photograph. The semi-automatic approach
on Photograph needs that a user marks leaf and background on regions
automatically detected by a Mean Shift algorithm. Then, a merging process labelled
gradually all content of the picture with these 2 parts. Contour were described
with complex network[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], volumetric fractal dimension [?] and geometric
parameters. Gabor lters, and local binary patterns were used too. They included for
each run the GPS information. Classi cation was performed by using Linear
Discriminant Analysis. The rst run used all samples from the 3 images
categories for the training stage, which paid o comparing to the second run where
only Scan and Scan-like were used. Only Gabor and GPS features were used for
the automatic run on Photograph because contour extraction were su ciently
reliable.
      </p>
      <p>
        Inria IMEDIA (3 runs) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] These participants used distinct approaches. For
Scan and Scan-like, they applied a late fusion of an approach based on shape
boundary features and a large-scale matching approach. The second run used an
advanced shape context descriptor combining boundary shape information and
local features within a matching approach. These two runs used a top-K decision
rule as classi er, while the third used a multi-class SVM technique on contour
based descriptors with a one-vs-one schema and a linear kernel. For Photograph,
local features around constrained Harris points in order to reduce the impact
of the background, but automatic segmentation with a rejection criterion was
attempted in order to extract shape features when possible. Two rst runs used
matching and top-K classi er. Last run used a method were Harris points were
associated to bounding boxes detected as images' interesting zones; each box
proposed a classi cation based on a multi-class SVM and a voting scheme
unifying their responses. All preliminary evaluations used cross validation without
splitting images from a same individual plants.
      </p>
      <p>
        LIRIS ReVes (3 runs) [12] The method used by these participants is clearly
oriented mobile application and is designed to cope with the challenges of
complex natural images and to enable a didactic interaction with the user. The
system rst performs a two-step segmentation of the image, fully automatic on
plain background images, and only guided by a rough coloring on photographs.
It relies on the evaluation of a leaf model representing the global morphology of
the leaf of interest, which is then used to guide an active contour towards the
actual leaf margin. High-level geometric descriptors, inspired by the criteria used
by botanists, are extracted on the obtained contour, along with more generic
shape features, making a semantic interpretation possible. All these descriptors
are then combined in a Random Forest classi cation algorithm[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
LSIS DYNI (3 runs) [19] These participant used a modern computer vision
framework involving feature extraction coupled with Spatial Pyramidal
Matching for local analysis and large-scale supervised classi cation based on linear
SVM with the 1-vs-all multi-class strategy. For all submitted runs, they do not
consider the three image categories. They notably obtained good results on Scan
and Scan-like without any segmentation and/or speci c pre-processing, and they
obtained the 3 best results for Photograph among all full-automatic approaches
with a signi cant di erence. The di erences between the runs concerns the choice
of features and their combination (in late fusion schema). They used a
Multiscale and Color extension of the Local Phase Quantization[17], dense Multiscale
(Color) Improved Local Binary Patterns with Sparse Coding of patches[24], and
dense SIFT feature with Sparse Coding too. More they combined the
descriptors, more they obtained better performances.
      </p>
      <p>Sabanci-Okan (2 runs) [25] Mainly focused by automatic shape boundary
analysis, these participants adapted their approach according the image
content and obtained the best results for Scan and Scan-like. They used a balanced
combination of 2 kinds of classi cations, a Local Features Matcher and SVM
classi ers. Only Scan and Scan-like images were used as a training dataset. They
used two distinct segmentation methods: Otsu's method for Scan, Scan-like,
and two methods based on watershed transform for Photograph (full or
semiautomatic). The Local Feature Matcher exploited local features (shape context,
HOG, local curvature) from salient points on the boundary of the leaf. The SVM
classi ers used texture and optionally shape features if test image were from the
category Scan, or Scan-like or interactively segmented Photograph. For
automatically segmented Photograph test images, the leaf boundary extractions were not
su ciently reliable for using shape information, and thus they used texture
features for re-training the classi ers. These participants are among the rare ones
to mention that they used cross validation without splitting images from a same
individual plants in order to avoid over tting problems.</p>
      <p>
        The Who IITK (4 runs) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] One originality is that these participants
rede ned 3 categories taking into account morphological properties of leaves:
Simple leaf in Scan or Scan-like, Compound leaf in Scan or Scan-like and
Photograph. For the "Simple" category, they developed an automatic segmentation
method based on Otsu working on two steps in order to remove shadows, noisy
background and the petiole of the leaf since it can a ect the shape. For the
"Compound" category, they used a combination of polynomial curve tting and
minimum-bounding-ellipse scores in order to extract the contour of the "most"
representative lea et. For Photograph, they developed a GrabCut [20] based
interactive segmentation technique. Shapes were then described with complex
network [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and geometric parameters, and extracted specialized features on margin
and venation. In order to fully automatise the process, they designed a
Simple/Compound leaf classi er for test images. Note that on some "complex"
photos where contour was too di cult to extract, like bipinnate compound leaves,
they chose to use SURF with a bag-of-word technique. Random Forest Classi
cation [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] for each category was nally used for the species prediction. The run
giving the best result used in addition the Smote algorithm [13] designed to deal
with unbalanced classi cation problems.
      </p>
      <p>ZhaoHFUT (3 runs) [26] These participants investigated a linear Spatial
Pyramidal Matching using Sparse coding (ScSPM) since this kind of approach
achieve very good performance in object recognition. Indeed, SPM extends
popular bag-of-features approach by taking account the spatial order of local
appearances. They used the same structure in [24] and they tested thought the 3
runs the spatial encoding of dense SIFT, dense ip-SIFT descriptors (a SIFT
extension considering leaf symmetry), and a late fusion of the two approaches.
The sparse encoded features are pooled among several scales as spatial pyramid
histogram representations, and these last representations are used as input of
SVM classi ers tuned by cross-validations. The species predictions are formed
by normalizing the degrees of attribution in di erent SVMs.</p>
      <p>Table 2 attempts to summarize the methods used at di erent stages (feature,
classi cation, subset selection,...) in order to highlight the main choices of
participants. This table should be used in next section on result analysis, in order to
see if there is some common techniques which tend to lead to good performances.
w G</p>
      <p>e
to iv</p>
      <p>su tc
tsu i-au
t a</p>
      <p>r
O m O e
: e : t
s S s in
n : n :
a o a o
c t c t
lS o lS o
l h l h
i o
c h
S
p</p>
      <p>p
t
if n
h io
S r
n te</p>
      <p>i
s ea r r
n c u
ea M tr to</p>
      <p>e o n
-M iv b co
tsu .ok trca tsu a+ ev</p>
      <p>u i
O t e O s t
: u t : t c
s A In s O a
n : : n : d
ca to to ca to ed
llS oh oh llS oh iu</p>
      <p>A</p>
      <p>I
S</p>
      <p>O</p>
      <p>A P P A P G</p>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <p>Global analysis
Best runs: Although scores are more uniform across the three image types,
there is still no run this year achieving the best performances on the three
types. There is even no run in the top-3 of each image type. Sabanci-Okan runs
achieved very good results on both Scan (1st) and Scan-like (2nd). TheWho
runs achieve good results across the 3 categories but do not obtain the best
performances on any of them. IFSC USP run1 obtained the best performance
on Photograph category using semi-supervised segmentation but gets modest
performances on Scan. Similarly, INRIA run 2 obtained the best performance
on Scan-like but gets mitigated performances on Scan and Photograph. Finally,
the 3rd run of LSIS DYNI is interesting even if it is positioned only at the 9th
rank when looking at the average score across all categories: it actually obtained
the best fully automatic performances on Photograph category which is
actually the most challenging task. A normalized recognition rate of 32% on that
category makes an automatic real-world application more realistic (whereas last
year's automatic runs on photographs did not exceed 20% and with less species).
Impact of the number of species: A rst general remark is that the scores are
globally lower than the one obtained during the 2011 campaign whereas the new
evaluation metric of the 2012 campaign is more advantageous for participants
(only the top-1 specie contributed to the score within the old metric whereas the
new one allows retrieving the right species at larger ranks). That means that the
di culty of the task increased, and that the technological progress achieved by
the participants did not compensate the increased di culty. The main change
between the two campaigns being the number of species covered by the task
(from 70 to 126 species), it is very likely that this growth explains the
overall lower performances. More classes actually involve more confusion. From a
botanical point of view, the 2012 dataset actually contains more pairs of
morphologically similar species. The species Betula pendula Roth was for instance
quite well recognized in many runs during the 2011 campaign but is this year
often confused with other species, probably Betula pubescens Ehrh. or Betula
utilis var. jacquemontii sharing several morphological attributes.
Impact of the image type: Another global remark is that the scores of the
2012 campaigns are more uniform across the di erent image types than the
year before. So that it is not true anymore that the performances are degrading
with the complexity of the acquisition image type. This might be due to several
reasons: (i) many species in the training set are relatively to other types more
populated with Scan-like images than last year, (ii) several participants did use
semi-supervised approaches to extract leaf contours in the Photograph category,
(iii) participants may have work more on improving their methods for Scan-like
and Photograph (since scans were identi ed more easily in the 2011 campaign).
About full automatic approaches and using semi-supervised
segmentation: Every runs were performed with full automatic methods on Scan and
Scan-like, with or without automatic image segmentation, revelling that all teams
conceive that the problem is solvable in a long term. But with a max of 0:58,
and average and a median around 0; 32 for Scan, there is still an important room
of improvement especially when all tree species from a given ora (like around
the 500 to 600 tree species in France) will be considered.</p>
      <p>Concerning Photograph, results demonstrated that a quite accurate leaf
contour extraction in all training dataset performed with semi-supervised
segmentation approaches enable to tend towards equivalent performances for Scan and
Scan-like like for the IFSC-USP or TheWho teams. However it is not the case
for Sabanci-Okan semi-automatic run, maybe because they performed
semisupervised segmentation on test-images only while learning models on Scan and
Scan-like. In the other side, IFSC-USP team obtained lower results on Scan and
Scan-like than on Photograph which maybe highlights that accurate leaf contour
is required. Moreover, it is di cult to conceive an approach needing tens or even
hundreds thousands of interactive leaf segmentations of a dataset on a entire
ora.</p>
      <p>
        About matching and shape boundary approaches: The most frequently
used class of methods is shape boundary analysis (16 runs among 30 used),
exclusively or not) which is not surprising since state-of-the-art methods addressing
leaf-based identi cation in the literature are mostly based on leaf segmentation
and shape boundary features [
        <xref ref-type="bibr" rid="ref5 ref7 ref9">7, 18, 9, 23, 5</xref>
        ]. Last year, one important remark was
that more generic image retrieval approaches like matching could provide
better results especially on Scan. But, rather than opposing these two approaches,
several participants combined them with success. This suggests that modelling
leaves as part-based rigid and textured objects might be complement to shape
boundary approaches that do not characterize well margin details, ribs or limb
      </p>
      <p>Impact of the training subsets and cross-validations: Considering test
images of Scan and Scan-like categories, we can say it was preferable to exclude
Photograph images from training dataset and to not distinguish Scan and
Scanlike during preliminary evaluations. But for Photograph test images, this rule
seems not to be con rmed if we look the best performances obtained
automatically by LSIS DYNI who used all images in the training dataset. Moreover, one
can note that Sabanci-Okan and Inria teams seem to have take bene t from
the respect of individual-plant during their cross-validation. Indeed, they
mentionned that they do not split images from one same individual plant, in order
to avoid over tting problems, because images from a one same plants (and thus
one same event) are very similar. This approach could be one key of the success
of these runs independently to the relevance of the methods.</p>
      <p>About using metadata: Using geo-tags to help the identi cation has been
tried again this year by some participants (IFSC USP, THEWHO and DBIS).
Unfortunately, this year again, the results show that adding GPS information
in the identi cation process is likely to degrade the performances. Best runs of
TheWho and DBIS are actually the ones that did not use any additional
metadata complementary to their baseline visual-based technique. The main reason is
probably that the geographic spread of the data is limited (French Mediterranean
area). So that most species of the dataset might be identically and uniformly
distributed in the covered area. Geo-tags would be for sure more useful at a global
scale (continent, countries). But at a local scale, the geographical distribution
of plants is much more complex. It usually depends on localized environmental
factors such as sun exposition or water proximity that would require much more
data to be modelled.</p>
      <p>We have to note that some potentially relevant information were not or very few
explored. The content tags (Leaf, Picked leaf, Leafage) for instance were only
used by Inria Team. Date could have been used in order to identify evergreen
plants and improve performances on these species. None of teams explored either
the hierarchical taxonomy structure which could be a source of improvement.
But some participants like The Who and ReVes LIRIS teams, created their own
metadata about the leaf organisation (Simple or Compound ) in order to treat
di erently these two categories of leaves, which seems to have paid o for
improving performances.
5.2</p>
      <p>Performances per species
This section aims to analyse, considering the results of the various approaches,
how and which part of the task were di cult. We analyse which species, images
and kind of leaves according to morphological attributes, were easy or very hard
to identify.</p>
      <p>We attempt here to evaluate which species are more di cult to identify than
others. We focus only on Scan in order to limit view conditions bias. We
computed some basic statistics (mean, median, max) on the score by species over
the runs of all participants. Figure 17 shows these statistics by species with a
decreasing order of the max score values. We can note that few species enable to
obtain good performances over all approaches. For around one third there is a
least one approach which performed perfectly on all images of theirs associated
species. Most of them were species yet contained in the previous year dataset,
and thus most of the time with more individual-plants and images. Figure 18
shows in addition how are related max scores to the number of individual plant
by species and it indicates clearly how more a species has some individual plants
(as existing species previous year) more the identi cation of the species is
reachable by most of the methods. For species with a small number of individual
plants it is more di cult to understand precisely the score variations. They can
be due to morphological variations but also to di erent view conditions (we will
see it next subsections).</p>
      <p>Coming back to gure 17, for the other two-thirds of species, none of the tested
methods perform perfectly. However, by comparing max and mean scores, we
can say that there is each time on method performing distinctly better than the
other. It is encourageing for all teams, beacuse it means there is for all these
species a room of improvement.
5.3</p>
      <p>Performances per image
To qualitatively assess which kind of images causes good results and which one
makes all methods failed, we sorted all test pictures by the rate of succesfull runs
in which they were correctly identi ed (where the correct species was retrieved at
rst rank). Figure 19 attempts to represent that for each image category and with
a decreasing order. First we can note that that Scan and Scan-like have more or
less the same distribution relatively to their number of images, con rming that
these two categories have a same level of di culty in spite of the great variety of
the methods experimented through the run. Moreover, the gure con rms that
Photograph is de nitely the most di cult category: a handful of photographs
has a rate of successful runs over 0.5 while at the same time around one third
of Scan and Scan-like images have it. 260 images were not correctly identi ed
by any run with a majority of unconstrained photos (63 Scan, 27 Scan-like, 168
Photograph.</p>
      <p>Figure20 displays 4 very well identi ed images with at least a rate of 0.8
successful run). They all are very standard leaf images similar to the one found
in books illustrations.</p>
      <p>Concerning Photograph, the three images with the higher rate of successful
runs are showed in rst line in gure21. One explication of these relative good
rates can be that the associated species contains after 3 years numerous images
and individual plants now, but also that these pictures were well framed, with a
close-up on the whole leaf. In comparison photos from the same species in second
line of gure21 have a very low, even 0, rate of successful runs: we can note that</p>
      <p>Fig. 19. Distribution of the rates of successful runs for each test-image, re-arranged
by the three image categories and with a decreasing order.
these pictures are not well framed, too small, too dark, or with a part of the leaf
out of the frame. . .</p>
      <p>However, we have to be careful with the "easy" or "di cult" images ccording
to the methods used in the runs. We notice that numerous of the "easiest"
y
s
a
E
t
l
u
c
i</p>
      <p>D
.
p
S</p>
      <p>Cercis siliquastrum L.</p>
      <p>Corylus avellana L.</p>
      <p>Fig. 21. The 3 photos with the highest rate of successful runs, and 3 photos wih very
low ones from the same species.
images can be considered "too much easy" because on some species we can have
images from a same "event" in the training and the test dataset. Indeed, we
remind that we split the Pl@ntLeaves II dataset with the respect of individual
plants, i.e. pictures of leaves belonging to the same individual tree cannot be
split across training and test data. This prevents identifying the species of a
given tree thanks to its own leaves and that makes the task more realistic. In
a real world application, it is indeed much unlikely that a user tries to identify
a tree that is already present in the training data. However, some individual
plants from a same species are closely related by one same "event" because they
were observed the same day, by the same contributor, and images were collected
with the same device, with similar pose, framing and conditions. Moreover, the
individual plants from a same "event" are observed at the same place and can
be parent tree and/or present the same morphological attributes in function of
the environment. In other words, these individual plants can be considered as
one same individual-plant, and Since we not split the dataset with the respect of
the notion of "event", it still have some bias between test and training datasets,
as as it can be observed in gure 22.
test image 10921
indiv. 209
12/10/11
train image 5483
indiv. 42
12/10/11
train image 6358
indiv. 103
12/10/11</p>
      <p>This notion of "event", instead of "individual-plant", should be considered
in a next plant identi cation task.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusions</title>
      <p>This paper presented the overview and the results of ImageCLEF 2012 plant
identi cation testbed following the pilot one in 2011. The number of
participants increased from 8 to 11 groups showing an increasing interest in applying
multimedia search technologies to environmental challenges (in 8 distinct
countries). The main change between the two campaigns was the number of species
covered by the task which was increased from 70 to 126 species. The
technological progress achieved by the participants did unfortunately not compensate
this increased di culty since the overall performances degrade. This shows that
scaling state-of-the-art plant recognition technologies to a real-world application
with thousands of species might still be a di cult task. On the other side, a
positive point was that the performances obtained on unconstrained pictures
of leaves with non-uniform background were better than last year. The use of
semi-supervised segmentation techniques has notably been a key element towards
boosting performances showing that interactivity might play an important role
towards successful applications, typically on mobile phones. But, the best
automatic method also provided very honourable results showing that the door is
not closed for a fully automatic identi cation. Interestingly this method was a
generic visual classi cation technique without any speci city related to plants.
Regarding the other image types, the best method on scans used a combination
of many leaf boundary shape and texture features trained with a SVM
classier. The best run on pseudo-scans (photographs with uniform backgrounds) was
achieved by another technology purely based on shape-context features that are
already known to provide good performances for leaf recognition in the
literature. Finally, using GPS data in the recognition process was still not bene cial
probably because the geographic spread of the data is limited. A possible
evolution for a new plant identi cation task in 2013 is to extend the task to multiple
organs including owers, trunks and fruits.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This work was funded by the Agropolis fundation through the project Pl@ntNet
(http://www.plantnet-project.org/) and the EU through the CHORUS+
Coordination action (http://avmediasearch.eu/). Thanks to all participants. Thanks
to Violette Roche, Jennifer Carre and all contributors from Tela Botanica.
Thanks to Vera Bakic and Souheil Selmi from Inria and Julien Barbe from Amap
for their help for checking the datasets and developing the crowd-sourcing
applications.
10. Bottcher, T., Schmidt, C., Zellhofer, D., Schmitt, I.: BTU DBIS' Plant
Identication Runs at ImageCLEF 2012. In: Working notes of CLEF 2012 conference
(2012)
11. Casanova, D., ao Batista Florindo, J., Goncalves, W.N., Bruno, O.M.: IFSC/USP
at ImageCLEF 2012: Plant identi cation task. In: Working notes of CLEF 2012
conference (2012)
12. Cerutti, G., Antoine, V., Tougne, L., Mille, J., Valet, L., Coquin, D., Vacavant,</p>
      <sec id="sec-7-1">
        <title>A.: ReVeS Participation - Tree Species Classi cation using Random Forests and</title>
      </sec>
      <sec id="sec-7-2">
        <title>Botanical Features. In: Working notes of CLEF 2012 conference (2012)</title>
        <p>13. Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: Smote: Synthetic
minority over-sampling technique. Journal of Arti cial Intelligence Research 16,
321{357 (2002)
14. Gallo, I., Zamberletti, A., Albertini, S.: Fast Tree Leaf Image Retrieval using a
Probabilistic Multi-class Support Vector. In: Working notes of CLEF 2012
conference (2012)
15. Goeau, H., Joly, A., Selmi, S., Bonnet, P., Mouysset, E., Joyeux, L.: Visual-based
plant species identi cation from crowdsourced data. In: Proceedings of ACM
Multimedia 2011 (2011)
16. Grozea, C.: Brainsignals Submission to Plant Identi cation Task at ImageCLEF
2012. In: Working notes of CLEF 2012 conference (2012)
17. Heikkila, J., Ojansivu, V., Rahtu, E.: Improved blur insensitivity for decorrelated
local phase quantization. In: Proceedings of the 2010 20th International
Conference on Pattern Recognition. pp. 818{821. ICPR '10, IEEE Computer Society,
Washington, DC, USA (2010), http://dx.doi.org/10.1109/ICPR.2010.206
18. Neto, J.C., Meyer, G.E., Jones, D.D., Samal, A.K.: Plant species identi cation
using elliptic fourier leaf shape analysis. Computers and Electronics in Agriculture
50(2), 121 { 134 (2006)
19. Paris, S., Halkias, X., Glotin, H.: Participation of LSIS/DYNI to ImageCLEF 2012
plant images classi cation task. In: Working notes of CLEF 2012 conference (2012)
20. Rother, C., Kolmogorov, V., Blake, A.: grabcut": interactive foreground extraction
using iterated graph cuts. ACM Trans. Graph pp. 309{314 (2004)
21. Schmitt, I.: Qql: A db&amp;ir query language. The VLDB Journal 17(1), 39{56 (Jan
2008), http://dx.doi.org/10.1007/s00778-007-0070-1
22. Soderkvist, O.J.O.: Computer Vision Classi cation of Leaves from Swedish Trees.</p>
        <p>Master's thesis, Linkoping University, SE-581 83 Linkoping, Sweden (September
2001), liTH-ISY-EX-3132
23. Yahiaoui, I., Herve, N., Boujemaa, N.: Shape-based image retrieval in botanical
collections. In: Advances in Multimedia Information Processing - PCM 2006, vol.
4261, pp. 357{364 (2006)
24. Yang, J., Yu, K., Gong, Y., Huang, T.: Linear spatial pyramid matching using
sparse coding for image classi cation. In: in IEEE Conference on Computer Vision
and Pattern Recognition(CVPR (2009)
25. Yanikoglu, B., Aptoula, E., Tirkaz, C.: Sabanci-Okan at ImageCLEF 2012 Plant</p>
      </sec>
      <sec id="sec-7-3">
        <title>Identi cation Competition: Combining Features and Classi ers (2012)</title>
        <p>26. Zheng, P., Zhao, Z.Q., Glotin, H.: ZhaoHFUT at ImageCLEF 2012 Plant Identi
cation Task. In: Working notes of CLEF 2012 conference (2012)</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>1. The icl plant leaf image dataset</article-title>
          , http://www.intelengine.cn/English/dataset
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Tropicos</surname>
          </string-name>
          (
          <year>August 2012</year>
          ), http://www.tropicos.org
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Belhumeur</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feiner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jacobs</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kress</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ramamoorthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.B.</given-names>
            ,
            <surname>Dixit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Ling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Mahajan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Russell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Shirdhonkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Sunkavalli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>White</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          :
          <article-title>First steps toward an electronic eld guide for plants</article-title>
          .
          <source>Taxon</source>
          <volume>55</volume>
          ,
          <issue>597</issue>
          {
          <fpage>610</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Arora</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bagmar</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhattacharya</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A Plant Identication System using Shape and Morphological Features on Segmented Lea ets</article-title>
          .
          <source>In: Working notes of CLEF 2012 conference (</source>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Backes</surname>
            ,
            <given-names>A.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Casanova</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruno</surname>
            ,
            <given-names>O.M.:</given-names>
          </string-name>
          <article-title>A complex network-based approach for boundary shape analysis</article-title>
          .
          <source>Pattern Recognition</source>
          <volume>42</volume>
          (
          <issue>1</issue>
          ),
          <volume>54</volume>
          {
          <fpage>67</fpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Bakic</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yahiaoui</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mouine</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ouertani-Litayem</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ouertani</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>VerroustBlondet</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Goeau, H.,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Inria IMEDIA2's participation at ImageCLEF 2012 plant identi cation task</article-title>
          .
          <source>In: Working notes of CLEF 2012 conference (</source>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Belhumeur</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feiner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jacobs</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kress</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ling</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramamoorthi</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sheorey</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>White</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , L.:
          <article-title>Searching the world4s herbaria: A system for visual identi cation of plant species</article-title>
          .
          <source>In: ECCV</source>
          , pp.
          <volume>116</volume>
          {
          <issue>129</issue>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Random forests</article-title>
          .
          <source>Mach. Learn</source>
          .
          <volume>45</volume>
          (
          <issue>1</issue>
          ),
          <volume>5</volume>
          {32 (Oct
          <year>2001</year>
          ), http://dx.doi.org/10.1023/A:1010933404324
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Bruno</surname>
            ,
            <given-names>O.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>de Oliveira</surname>
            <given-names>Plotze</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Falvo</surname>
          </string-name>
          , M.,
          <string-name>
            <surname>de Castro</surname>
          </string-name>
          , M.:
          <article-title>Fractal dimension applied to plant identi cation</article-title>
          .
          <source>Information Sciences</source>
          <volume>178</volume>
          (
          <issue>12</issue>
          ),
          <volume>2722</volume>
          {
          <fpage>2733</fpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>