<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Sabanci-Okan System at ImageClef 2011: Plant identi cation task</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Berrin Yanikoglu</string-name>
          <email>berrin@sabanciuniv.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Erchan Aptoula</string-name>
          <email>erchan.aptoula@okan.edu.tr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Caglar Tirkaz</string-name>
          <email>caglart@sabanciuniv.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Okan University</institution>
          ,
          <addr-line>Istanbul, Turkey, 34959</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sabanci University</institution>
          ,
          <addr-line>Istanbul, Turkey 34956</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2011</year>
      </pub-date>
      <abstract>
        <p>We describe our participation in the plant identi cation task of ImageClef 2011. Our approach employs a variety of texture, shape as well as color descriptors. Due to the morphometric properties of plants, mathematical morphology has been advocated as the main methodology for texture characterization, supported by a multitude of contour-based shape and color features. We submitted a single run, where the focus has been almost exclusively on scan and scan-like images, due primarily to lack of time. Moreover, special care has been taken to obtain a fully automatic system, operating only on image data. While our photo results are low, we consider our submission successful, since besides being our rst attempt, our accuracy is the highest when considering the average of the scan and scan-like results, upon which we had concentrated our e orts.</p>
      </abstract>
      <kwd-group>
        <kwd>Plant identi cation</kwd>
        <kwd>mathematical morphology</kwd>
        <kwd>morphological covariance</kwd>
        <kwd>Fourier descriptors</kwd>
        <kwd>Support Vector machines</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The plant identi cation task in ImageCLEF 2011 consisted of labelling images
of plants that were captured by di erent means (scans, scan-like photos called
pseudo-scans and unrestricted photos). The details of the recognition task are
described in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. A content-based image retrieval (CBIR) system for plants would
be very useful for plant enthusiasts or botanists who would like to learn more
about a plant they encounter. The goal of the competition was to benchmark
state-of-the-art in this open problem where there are very few systems for
identifying unconstrained whole or partial plant images [
        <xref ref-type="bibr" rid="ref13 ref9">9, 13</xref>
        ]. The existing research
in this area is concentrated on isolated leaf identi cation [
        <xref ref-type="bibr" rid="ref10 ref12 ref14 ref15 ref3 ref4">4, 3, 10, 12, 15, 14</xref>
        ].
      </p>
      <p>Content-based plant identi cation problem faces many challenges such as
color, illumination, size variations that are also common in other CBIR problems,
as well as some speci c problems such as the variations in the composition of the
leaves that makes the plant shape variable. In addition, one can see that color is
less identifying in the plant retrieval problem compared to many other retrieval
problems, since most plants have green tones as their main color with subtle
di erences. In the rare cases that color is discriminative for a certain plant, that
is when the plant has an unusual color, then it may be the case that the leaves
of that plant may also have other colors, due to individual plant or seasonal
variations (e.g. Gingko, Eurasian smoketree), as shown in Fig. 1. Another issue
with color in plant identi cation is due to the challenges posed by the color of
the owers: a owering plant should be matched despite di erences in ower
colors.</p>
      <p>
        While shape is quite discriminative in identifying isolated leafs, it is not
as useful in identifying full plant images, since the global shape of a plant is
a ected from its leaf composition [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In that regard, isolated leaf identi cation
can be said to be a simpler problem compared to the unconstrained images of
full or partial plants. One method to address this problem could be to extract
an individual leaf image by segmenting the overall plant image. Texture on the
other hand seems to be a more robust and useful feature category for plant
identi cation, and is widely used in plant identi cation.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Overall architecture</title>
      <p>Upon examination of the training images that were categorized according to the
capture type, we observed that the scanned and scan-like categories were similar
in di culty and both seemed signi cantly easier compared to the photo category
that included a larger variation in scale. Due to shortage of time, we decided
to concentrate our e orts on the scan categories, while the photo category was
tackled during the last week { which was insu cient for such a di cult problem.</p>
      <p>The nal system is designed as two separate sub-systems, one for scan and
scan-like images and another one for photos. Since the meta-data included the
acquisition type, an input image is automatically sent to the correct subsystem.
The acquisition method was the only meta-data used in the overall system.</p>
      <p>
        Based on our previous work on plant identi cation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], we had some
experience with the usefulness of di erent feature groups. For handling photographs,
which was the problem addressed in our previous work, we found that global
shape descriptors and many of the descriptors considered in the present work,
would not be useful if the photo consisted of an overlapping set of leaves, rather
than a single leaf. On the other hand for scan and scan-like categories, all three
main feature categories are useful: color, texture and shape.
      </p>
      <p>After experimenting with a large number of descriptors, we selected a
115dimensional feature vector for the scan/scan-like sub-system and its 91-dimensional
subset for the photo category. The features used in our system are explained in
Section 3. For training the system, we trained a classi er combination using
Support Vector Machines (SVMs), as explained in Section 4.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Feature extraction</title>
      <p>As we are dealing with objects characterized mainly by their morphometric
properties, whenever possible we attributed special preference to using morphological
solutions. Since mathematical morphology, a nonlinear image processing
framework, excels at shape based image analysis as well as at exploiting the spatial
relationships of pixels.</p>
      <p>
        An additional motivation in this regard has been to test our recently
conceived morphological texture descriptors [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] in the context of a real-world
application. Although a rich variety of content descriptors has been investigated, we
present in this section primarily those that were included in the nal system.
3.1
      </p>
      <sec id="sec-3-1">
        <title>Texture Features</title>
        <p>As far as scan and scan-like images are concerned, one can easily remark that
we are dealing with relatively low scale variations, that can be easily countered
with some form of normalization, while plant alignment is also not a major
issue. Consequently, scale invariance set aside, from a texture description point
of view, we require descriptors possessing a) a high discriminatory potential, b)
illumination invariance, and unless we apply some form of angle normalization,
then also c) rotation invariance.</p>
        <p>When it comes to photos however, global texture characterization methods
are bound to fail, since besides requiring all kinds of invariances, the background
varies extremely in terms of complexity, thus presenting a considerable challenge.
Hence, in order to apply any global morphological texture operators, a successful
segmentation isolating the plant is necessary.</p>
        <p>
          A set of novel morphological grayscale texture descriptors, possessing the
aforementioned qualities has been recently introduced [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], leading to the
highest classi cation scores among grayscale approaches, with a variety of texture
benchmark collections, including Outex, CUReT and ALOT. They have been
formulated as extensions for morphological covariance, that equip it with rotation
and illumination invariance. Among them, we focused particularly on circular
covariance histograms (CCH) and rotation invariant points (RIT). In summary,
these two features achieve rotation invariance straightforwardly by replacing the
point pairs of standard morphological covariance, with a circular structuring
element (SE), possessing its center. Although any isotropic SE would su ce,
this particular shape has the advantage of preserving the principle of covariance,
consisting of comparing pixels at various distances.
        </p>
        <p>As to illumination invariance, they take advantage of the complete lattice
foundation of mathematical morphology. More precisely, morphological
operators operate on pixel extrema, and not on their linear combinations. In other
words, even if in a set of pixels the overall intensity levels change, as long as
the relative order of pixels with respect to their intensity remains the same, the
morphological operator under consideration, be it erosion, dilation or a
combination thereof, will be una ected, and will still pick as extremum the same pixel,
albeit with a modi ed intensity value. That is why, conversely to granulometries
and covariance, where a Lebesgue measure is used in order to quantify the
morphological series, CCH and RIT rely on using directly the characteristic scale of
each pixel. While the entire input image is described by means of its histogram
of characteristic scales.</p>
        <p>As to the di erence of RIT from CCH, it is computed similarly, with the
exception of rst decomposing the circular SE into anti-diametrical point triplets.
Thus there is an additional step of computing a label image, by means of a pixel
based fusion. In particular, a rotation invariant measure is used to this end
(e: g: minimum, maximum), upon all the intermediately ltered images by point
triplets of various orientations.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Color Features</title>
        <p>
          Since the previously chosen texture descriptors have not yet been extended to
color data, it was decided to employ the parallel color texture description
strategy, where color is described independently from texture. Among the investigated
methods we can mention multi-resolution histograms based on morphological
scale-spaces [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], in both polar and perceptual color spaces, non-uniformly
subquantized saturation weighted LSH histograms, as well as color invariants and
color moments. Yet, following an experimental evaluation of these color
descriptors, only color moments have been included in the nal system.
        </p>
        <p>
          To explain, a color image corresponds to a function I de ning RGB triplets
for image positions (x; y) : I : (x; y) 7! (R(x; y); G(x; y); B(x; y)). By regarding
RGB triplets as data points coming from a distribution, it is possible to de ne
moments. Mindru et al. [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] have de ned generalized color moments Mpaqbc:
Mpaqbc =
        </p>
        <p>Z Z
xpyq[IR(x; y)]a[IG(x; y)]b[IB(x; y)]cdxdy
(1)
Mpaqbc is referred to as a generalized color moment of order p+q and degree a+b+c.
This descriptor uses all generalized color moments up to the second degree and
the rst order, which leads to nine possible combinations for the degree: Mp1q00,
Mp0q10, Mp0q01, Mp2q00, Mp1q10, Mp0q20, Mp0q11, Mp0q02 and Mp1q01. These are combined
with three possible combinations for the order, M0a0bc, M1a0bc and M0a1bc, which
makes a 27-dimensional feature vector, possessing additionally shift-invariance,
if the average is subtracted from all input channels.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Shape Features</title>
        <p>Undoubtedly, shape plays a major role in plant identi cation and a plethora of
shape descriptors, usually categorized as region- and contour-based, are available.
In our case we employed a variety of shape descriptors from both categories.
Fourier descriptors We used the Fourier descriptors that are widely used to
describe shape boundaries, as the main shape feature in our system. The Fourier
Transform coe cients of a discrete signal f (t) of length N is de ned as:
Ck =</p>
        <p>N</p>
        <p>t=0
1 N 1</p>
        <p>X f (t)e j2 tk=N
k = 0; 1:::; N
1
(2)
In our case, f (t) is the 8-directional chaincode of the plant, N is the number of
points in the chaincode, and Ck is the k-th Fourier coe cient.</p>
        <p>The coe cients computed on the chaincode is invariant to translation since
the chaincode is invariant to translation. Rotation invariance is achieved by using
only the magnitude of the coe cients and ignoring the phase information. Scale
invariance is achieved by dividing all the coe cients by the magnitude of the
DC component. We used the rst 50 coe cients to obtain a xed-length feature
and to eliminate the noise in the leaf contour.</p>
        <p>
          Width length/volume factor: These two descriptors are slight variations of
the leaf width factor (LWF) introduced by Hossain and Amin [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. Speci cally,
given an isolated leaf image (f ), their method consists in dividing it into n strips,
perpendicular to its major axis (Fig. 2). For the nal n-dimensional feature, they
compute the length of each strip (li), divided by the length of the entire leaf (l):
LW Fn = fli=lg1 i n
(3)
        </p>
        <p>We derived two new features from this. The Width length factor normalizes the
lengths of each strip by the maximum width of the leaf. This is necessary as we
normalize the length of each leaf into a xed size during preprocessing, leaving
the width as variable. The second derived feature is obtained by integrating into
LWF the grayscale variations of each strip (fi), thus obtaining Width volume
factor (WVF). Speci cally, we employ the ratio of volumes (i: e: sum of pixel
values) instead of lengths:</p>
        <p>W V Fn = fVol(fi)= Vol(f )g1 i n
(4)</p>
        <p>Convexity: This mono-dimensional feature aims to describe the overall contour
smoothness of its binary input, which is assumed to be consisting of a single
connected component. To explain, after isolating and binarizing the plant image,
we compute its convex hull (CH) and then trivially derive its convexity:
Convexity(f ) =</p>
        <sec id="sec-3-3-1">
          <title>Area(CH(f )) Area(f )</title>
        </sec>
        <sec id="sec-3-3-2">
          <title>Area(f )</title>
          <p>BKTn(f ) =</p>
          <p>Vol(TP2;v (S(f ))</p>
          <p>Vol(S(f ))
v
where T denotes the morphological operator (either an opening or a closing
), Vol the sum of pixel values of f and P2;v a pair of points separated by a
vector v. Moreover, it should be noted that since we use a horizontal pair of
points (i: e: translation invariant w.r.t the contour pro le), the resulting feature
is thus rotation invariant.</p>
          <p>gi(f ) = f
"(f )
Next, we calculate the Euclidean distances from the aforementioned center to
each of the border pixels, which leads to a discrete series S(f ); in case of rotation
of the input image, S(f ) is only shifted horizontally. Thus the nal feature is
obtained by means of 4 simple statistical measures on S:</p>
          <p>BSS(f ) = fmax(S(f )); min(S(f )); med(S(f )); var(S(f ))g
using its maximum (max), minimum (min), median (med) and variance (var).
And since they are horizontally translation invariant w.r.t S(f ), they lead to a
simple yet e ective and rotation invariant description.</p>
          <p>Basic shape statistics: This descriptor (BSS) on the other hand, operates on
the contour pro le of its binary input. Speci cally, we start by computing the
center of mass of a given binary plant image (f ), assumed to be consisting of a
single connected component. Then we obtain its morphological internal gradient,
computed by means of 3 3 square SE:
Border Covariance: Similarly to basic shape statistics, border covariance
(BK) also operates on the contour pro le S(f ) of its binary input, under the
same assumptions. This time however, instead of computing simple statistical
measures, we aim to capture contour regularity. To this end we employ
morphological covariance, along with a horizontal pair of points. In other words, we
treat the contour pro le as a mono-dimensional texture.</p>
          <p>We modi ed the standard morphological covariance operator so as to employ
openings and closings instead of erosions, in order to capture respectively both
bright details on dark background, as well as dark details on bright background:
(5)
(6)
(7)
(8)</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Classi cation</title>
      <sec id="sec-4-1">
        <title>Scan and Scan-like Images</title>
        <p>Mainly due to the collaboration of two universities, we trained two separate
classi ers using two di erent sets of features. For the rst classi er (Classi er1),
we used a 67-dimensional shape feature consisting of the 50 Fourier
descriptors, the width length factor, eccentricity and solidity. For the second classi er
(Classi er2), we used a 115-dimensional feature consisting of all the contour,
texture and shape features described in Section 3, excluding the Fourier descriptors.
For both classi ers, we used an SVM using the radial basis function kernel.</p>
        <p>The outputs of these classi ers are distances of the test instance to each
plant class. We then trained a third classi er to learn how to combine these two
classi ers at score level. Hence, the feature vector used in training the combiner
is of length 2 K, where K is the number of classes in the problem.</p>
        <p>
          In the nal stage of classi cation the most probable 5 classes are selected and
a multi-class SVM is trained speci cally for those classes for disambiguation.
This was done because we found it bene cial to train classi ers that would learn
to distinguish similar classes (e.g. di erent kinds of maples which are very similar
amongs themselves, compared to the other plants). While the original idea was
to train one such classi er according to the number of lobes in the leaves, the
di culty in assessing this information and the remaining complexity of this task
led us to train a new classi er on the y, using only the training instances
from the 5 most probable classes and all of the 182 (=115+67) features. We
use the outcome of this stage as the nal classi cation decision. Cross-validation
accuracies obtained for each classi er using the training data set are summarized
in Table 1.
As far as photos are concerned, due to time constraints, neither their feature
extraction nor their classi cation received the attention they deserved. Since
shape features were not used, we trained only a single SVM classi er, using a
91-dimensional feature vector. For this classi er, default parameters (cost = 25,
rst degree polynomial kernel) of the Weka SVM software [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] and the Sequential
Minimal Optimization (SMO) algorithm were used.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Experimental design</title>
      <p>
        In this section we describe the implementational choices that have been made,
as well as the experiments that have been carried out, while designing &amp;
optimizing our plant identi cation system. The majority of experimentations has
been realized using Weka (v.3.6.4) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In all stages of our experiments and in the
results given in throughout this paper, we used cross-validation on the training
data and measured the average accuracy obtained across images, rather than
the average user-based accuracy used in the competition.
      </p>
      <p>For practical reasons, we chose to rst optimize our descriptor combination,
then the preprocessing scheme, followed nally by the classi cation step. Given
their visual similarity, we handled scan and scan-like images identically, while
treating photos separately. Special care has been taken to obtain a fully
automatic system.</p>
      <p>For the sake of simplicity, and in order to minimize the number of
variables, initial feature selection experiments have been carried out with a nearest
neighbour (1NN) classi er along with the 2-distance, while subsequently we
switched to Support Vector Machines (SVM). Whenever necessary, conversion
to grayscale has been realized using the weighted combination of RGB channels,
as 0:299 R + 0:587 G + 0:114 B, while binarization has been achieved using
Otsu's threshold.
5.1</p>
      <sec id="sec-5-1">
        <title>Feature Selection</title>
        <p>Preliminary experimentation with various features has been realized by handling
shape, color and texture descriptors separately, in an e ort to determine the most
suitable among them for the problem under consideration. After this step, we
experimented with their combination and parameterization.</p>
        <p>Scan and scan-like images: At this stage a relatively simple preprocessing
step was realized, consisting of rst extracting the bounding rectangle of the
plant, followed by scale normalization resulting in a xed height of 600 pixels.
This was applied to all 71 classes containing a total of 3066 scan and scan-like
images. Then a series of cross-validation experiments took place, in an e ort to
determine the most suitable descriptors for distinguishing among these classes.
The results are shown in Table 2, along with their arguments.</p>
        <p>The accuracy scores have been obtained by dividing the available data
randomly into train (1444 samples) and test (1622 samples) sets, using the
aforementioned classi cation settings. Interestingly, one can observe that texture exhibits
the highest discriminatory potential, followed by color and shape.</p>
        <p>In addition to measuring the individual discriminatory potential of each
feature, we experimented with many of their combinations. The resulting scores are
given in Table 3, where we present the classi cation accuracies obtained with
various combinations of feature sets.
Photographs: Conversely to scan and scan-like images, the main challenge
presented by photos lies in isolating the plant from its often very complicated
background. Due mainly to lack of time, we hardly had any chance of
constructing an optimized feature set for this image category, as done for the other images.
Instead, we transferred almost directly our descriptor choices for scan and
scanlike data, with no additional preprocessing whatsoever. Nonetheless, one of the
very few experiments with photos that has been carried out, consisted of simply
testing the combination of the features given in Table 2, with the end of adapting
them to the new content.</p>
        <p>In particular, we joined the scan and scan-like images with photos, thus
obtaining a total of 3996 samples, and divided them equally and randomly into
training and test sets. The classi cation accuracies are provided in Table 4.
Considering the background complexity of photos, contour-based shape descriptors
su ered a signi cant performance loss; which was expected, since they rely
heavily on correct border extraction. Border covariance in particular has been unable
to contribute any longer, so it has been removed from the set of descriptors used
to characterize photos. Consequently the length of the feature vector used with
photos is 115 24 = 91. In summary, the addition of photographs has decreased
the overall classi cation performance considerably, with shape descriptors being
a ected the most.
5.2</p>
      </sec>
      <sec id="sec-5-2">
        <title>Preprocessing</title>
        <p>
          Having determined a set of features for describing the plant collection, we focused
on optimizing the preprocessing stage, in order to further improve performance.
Besides the already applied scale normalization, we considered illumination
normalization as well, through histogram equalization as proposed in Ref. [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
However this did not contribute in any substantial way, probably due to the fact
that our primary descriptors are already illumination invariant. Moreover, the
removal of the leaf petiole has also been tackled, with the end of obtaining a
more accurate plant border for contour-based shape descriptors, but
unfortunately its e ect has been negligible, probably due to the partial success of the
removal procedure.
        </p>
        <p>Furthermore, although most of our operators are rotation invariant, WVF
and color moments are not. That is why, we chose to apply an additional step,
that would align the plant vertically along its major axis. However, the small
gains were hindered by mistakes in the angle estimate; thus this normalization
did not improve our classi cation rates. Consequently, as far as scan and
scanlike images are concerned their preprocessing consists of extracting the bounding
rectangle of a given plant image, followed by its scale normalization to a xed
height of 600 pixels.</p>
        <p>Photographs on the other hand, given their content variation, should
benet from a background removal step. Due to time constraints, we experimented
with relatively simple and automatic hue based background removal; but it did
not contribute signi cantly to classi cation performance, and therefore were not
included in the nal system's operation. In short, photographs were not
preprocessed in any way.
5.3</p>
      </sec>
      <sec id="sec-5-3">
        <title>Classi er Optimization</title>
        <p>Having determined the descriptors and preprocessing operations for each
subproblem (scan and photo), we optimized the parameters of the classi ers. For
the two base classi ers used in scan/scan-like categories, the cost and spread
parameters (C and ) of the SVM are learned using 5-fold cross validation and
grid search. As for the combiner, we rst divided the training set into half. Then,
we trained the two classi ers with their optimum parameters on the rst half
and used the second half to produce distances. In the next step, we found the
optimum parameters for the combiner using 5-fold cross validation and grid search
on the parameters using the produced distances. Furthermore, as we analyzed
the errors of the system on training data, we decided to add a nal classi er
which is trained with only the closely scoring (top-5 classes) classes' instances,
as explained in Section 4. For the single classi er used in photos, we used the
default SVM parameters due to shortness of time.
6</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Results and discussion</title>
      <p>According to the o cial results (Table 5) our run achieved 6th place in overall
classi cation, 2nd place with scan type images, 6th place with scan-like images
and 16th place with photographs. However, we did achieve the best score when
considering the average of the scan and scan-like results, upon which we had
concentrated our e orts.</p>
      <p>Nevertheless, although the placements came as no surprise, the scores on the
other hand have been lower than our expectations, especially when compared
with the values obtained during our cross-validation tests with the training data
(Tables 3 and 4). As a matter of fact, there is a very signi cant drop of
approximately 40% overall. This important di erence may be due to several factors, one
of which is over tting of the classi ers. Our descriptor optimization stage may
have very well led to excessively powerful features capable of distinguishing the
training data quite e ectively, yet when faced with the test dataset, containing
distinct plants of the same genus, the same features failed to generalize their
performance. Furthermore, the scoring function is no longer the average
accuracy across images, but instead the average accuracy obtained per user; and of
course there is always the possibility of implementation errors.</p>
      <p>In conclusion, given that this is our rst participation, we consider our
attempt satisfactory and even successful, in the sense that we accomplished our
main goal; which consisted of identifying e ectively the plants in scan and
scanlike images. All the same, our score is far from being perfect. Future work in
this category will include testing non-morphological descriptors, in an attempt
to harness the advantages of both non-linear and linear image analysis
methodologies. As far as photographs are concerned, our main focus will be on the
preprocessing stage, with the end goal of isolating the plant e ectively from its
background, so as to be able to employ the same optimized features as with the
other image types.</p>
      <p>Yanikoglu et al.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>E.</given-names>
            <surname>Aptoula</surname>
          </string-name>
          .
          <article-title>Extending morphological covariance</article-title>
          .
          <source>Pattern Recognition</source>
          , May
          <year>2011</year>
          . Submitted.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>E.</given-names>
            <surname>Aptoula</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Lefevre</surname>
          </string-name>
          .
          <article-title>Morphological description of color images for contentbased retrieval</article-title>
          .
          <source>IEEE Transactions on Image Processing</source>
          ,
          <volume>18</volume>
          (
          <issue>11</issue>
          ):
          <volume>2505</volume>
          {
          <fpage>2517</fpage>
          ,
          <string-name>
            <surname>November</surname>
          </string-name>
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Backes</surname>
          </string-name>
          and
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Bruno</surname>
          </string-name>
          .
          <article-title>Shape classi cation using complex network and multi-scale fractal dimension</article-title>
          .
          <source>Pattern Recognition Letters</source>
          ,
          <volume>31</volume>
          (
          <issue>1</issue>
          ):
          <volume>44</volume>
          {
          <fpage>51</fpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Bruno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. O.</given-names>
            <surname>Plotze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Falvo</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Castro</surname>
          </string-name>
          .
          <article-title>Fractal dimension applied to plant identi cation</article-title>
          .
          <source>Information Sciences</source>
          ,
          <volume>178</volume>
          (
          <issue>12</issue>
          ):
          <volume>2722</volume>
          {
          <fpage>2733</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>G. D.</given-names>
            <surname>Finlayson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hordley</surname>
          </string-name>
          , G. Schaefer, and
          <string-name>
            <given-names>G. Y.</given-names>
            <surname>Tian</surname>
          </string-name>
          .
          <article-title>Illuminant and device invariant colour using histogram equalisation</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>38</volume>
          (
          <issue>2</issue>
          ):
          <volume>179</volume>
          {
          <fpage>190</fpage>
          ,
          <year>February 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. H. Goeau,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonnet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Boujemaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Barthelemy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Molino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Birnbaum</surname>
          </string-name>
          , E. Mouysset, and
          <string-name>
            <given-names>M.</given-names>
            <surname>Picard</surname>
          </string-name>
          .
          <article-title>The clef 2011 plant image classi cation task</article-title>
          .
          <source>In CLEF 2011 working notes</source>
          , Amsterdam, The Netherlands,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>M.</given-names>
            <surname>Hall</surname>
          </string-name>
          , E. Frank,
          <string-name>
            <given-names>G.</given-names>
            <surname>Holmes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pfahringer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Reutemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and I. H.</given-names>
            <surname>Witten</surname>
          </string-name>
          .
          <article-title>The weka data mining software: An update</article-title>
          .
          <source>SIGKDD Explorations Newsletter</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ),
          <year>June 2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>J.</given-names>
            <surname>Hossain</surname>
          </string-name>
          and
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Amin</surname>
          </string-name>
          .
          <article-title>Leaf shape identi cation based plant biometrics</article-title>
          .
          <source>In International Conference on Computer and Information Technology</source>
          , pages
          <volume>458</volume>
          {
          <fpage>463</fpage>
          , Dhaka, Bangladesh,
          <year>December 2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>H.</given-names>
            <surname>Kebapci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yanikoglu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Unal</surname>
          </string-name>
          .
          <article-title>Plant image retrieval using color, shape and texture features</article-title>
          .
          <source>The Computer Journal</source>
          ,
          <volume>53</volume>
          (
          <issue>1</issue>
          ):1{
          <fpage>16</fpage>
          ,
          <string-name>
            <surname>April</surname>
          </string-name>
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>F.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Q.</given-names>
            <surname>Man</surname>
          </string-name>
          .
          <article-title>Multiple classi cation of plant leaves based on gabor transform and lbp operator</article-title>
          .
          <source>In International Conference on Intelligent Computing</source>
          , pages
          <volume>432</volume>
          {
          <fpage>439</fpage>
          , Shanghai, China,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>F.</given-names>
            <surname>Mindru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tuytelaars</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Van Gool</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Moons</surname>
          </string-name>
          .
          <article-title>Moment invariants for recognition under changing viewpoint and illumination</article-title>
          .
          <source>Computer Vision</source>
          and Image Understanding,
          <volume>94</volume>
          (
          <issue>1-3</issue>
          ):3{
          <fpage>27</fpage>
          ,
          <string-name>
            <surname>April</surname>
            <given-names>-June</given-names>
          </string-name>
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Feng</surname>
          </string-name>
          .
          <article-title>Shape based leaf image retrieval</article-title>
          .
          <source>IEE Proceedings - Vision, Image and Signal Processing</source>
          ,
          <volume>150</volume>
          (
          <issue>1</issue>
          ):
          <volume>34</volume>
          {
          <fpage>43</fpage>
          ,
          <string-name>
            <surname>Feb</surname>
          </string-name>
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>D. M. Woebbecke</surname>
            , G. E. Meyer,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Von Bargen</surname>
            , and
            <given-names>D. A.</given-names>
          </string-name>
          <string-name>
            <surname>Mortensen</surname>
          </string-name>
          .
          <article-title>Plant species identi cation, size, and enumeration using machine vision techniques on near-binary images</article-title>
          .
          <source>In Optics in Agriculture and Forestry</source>
          , volume
          <volume>1836</volume>
          , pages
          <fpage>208</fpage>
          {
          <fpage>219</fpage>
          , Boston, USA,
          <year>1993</year>
          . SPIE.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Itheri</surname>
            <given-names>Yahiaoui</given-names>
          </string-name>
          , Nicolas Herve, and
          <string-name>
            <given-names>Nozha</given-names>
            <surname>Boujemaa</surname>
          </string-name>
          .
          <article-title>Shape-based image retrieval in botanical collections</article-title>
          .
          <source>In PCM</source>
          , pages
          <volume>357</volume>
          {
          <fpage>364</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>S.</given-names>
            <surname>Yonekawa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sakai</surname>
          </string-name>
          , and
          <string-name>
            <given-names>O.</given-names>
            <surname>Kitani</surname>
          </string-name>
          .
          <article-title>Identi cation of idealized leaf types using simple dimensionless shape factors by image analysis</article-title>
          .
          <source>Transactions of the ASAE</source>
          ,
          <volume>39</volume>
          :
          <fpage>1525</fpage>
          {
          <fpage>2533</fpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>