<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Glasgow, UK, April</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Identification of plant species on large botanical image datasets</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Naiara Aginako</string-name>
          <email>naginako@vicomtech.org</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Javier Lozano</string-name>
          <email>jlozano@vicomtech.org</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Quartulli</string-name>
          <email>mquartullli@vicomtech.org</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Basilio Sierra</string-name>
          <email>b.sierra@ehu.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Igor G. Olaizola</string-name>
          <email>iolaizola@vicomtech.org</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer Sciences and, Artificial Intelligence Department, University of the Basque Country</institution>
          ,
          <addr-line>+34943015102</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Vicomtech-IK4</institution>
          ,
          <addr-line>Paseo Mikeletegi 57, 20009 Donostia-San Sebastián, +34943309230</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Vicomtech-IK4</institution>
          ,
          <addr-line>Paseo Mikeletegi 57, 20009 Donostia-San Sebastián, +34943309230</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Vicomtech-IK4</institution>
          ,
          <addr-line>Paseo Mikeletegi 57, 20009 Donostia-San Sebastián, +34943309230</addr-line>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Vicomtech-IK4</institution>
          ,
          <addr-line>Paseo Mikeletegi 57, 20009 Donostia-San Sebastián, +34943309230</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2014</year>
      </pub-date>
      <volume>1</volume>
      <issue>2014</issue>
      <fpage>38</fpage>
      <lpage>44</lpage>
      <abstract>
        <p>The continuously growing amount of multimedia content has enabled the application of image content retrieval solutions to different domains. Botanical scientists are working on the classification of plant species in order to infer the relevant knowledge that permits them going forward in their environmental researches. The manual annotation of the existing and newly creation plants datasets is an outsized task that is becoming more and more tedious with the daily incorporation of new images. In this paper we present an automatic system for the identification of plants based on not only the content of images but also on the metadata associated to them. The classification has been defined as a classification plus fusion solution, where the images representing different parts of a plant have been considered independently. The promising results bring to light the chances of the application computer vision solutions to botanical domain.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>The digital age has brought the development of new
technologies that allow making deeper studies about our reality
and therefore, winning a more exhaustive knowledge. In addition,
the ever increasing use of digital cameras and sensors in several
fields, has led to an exponential growth in the amount of
multimedia content being generated every day in the world.
Nowadays, the whole society is involved in the generation of any
kind of content; it’s already a fact that digital technologies are
introduced in all aspects of our daily lives.</p>
      <p>Although the multimedia analysis techniques in their beginning
were focused on application sectors directly related with the
technology, their penetration in divergent sectors such as
medicine, meteorology, environment it’s a reality that is bringing
huge progress.</p>
      <p>Regarding environmental multimedia content, there is an
increasing need of techniques for analyzing, interpreting and
labelling of the content in order to enrich the actual knowledge.
This automatically extracted knowledge leads to the adoption of
new strategies that can improve the actual insight of the
environment to move forward in the deployment of new
directives to help in its protection and care.</p>
      <p>
        Initiatives such as Tela Botanica and projects such as Pl@ntNet
foster the development of this kind of technologies. Even more,
open competitions as ImageCLEF[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], and more precisely plant
identification task [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], where technological researchers focused on
multimedia content analysis take part, promote the approach of
these two worlds. Newborn mobile applications such as Plantifier,
LeafSnap or NatureGate are also examples of the natural synergy
tendencies.
      </p>
      <p>The image-based identification of different species of plants
that both botanical scientists and expert users have collected has
become a key study among plant biology science. On the one
hand, one of the peculiarities of plant image analysis is that such
images may belong to different plant parts such as leaf, stem or
flower. On the other hand, content is also time dependent, thereby
increasing the difficulty of the identification task. The latter can
be mitigated by using not only image content but also the linked
metadata. Thus, the analysis process is enriched and more
accurate results can be obtained. This metadata is not only the
data that users can add manually but information that nowadays
digital cameras impress automatically.</p>
      <p>One of the biggest handicaps of multimedia content analysis is
to determine the working domain so that afterwards, more domain
specific implementations are applied. In the case of ImageCLEF
dataset, there is a division of 6 subcategories that identify these
domains. Each image has an associated XML which specifies
what subcategory belongs to, permitting the abstraction from the
domain categorization issue.</p>
      <p>In our plant identification approach we used ImageCLEF
dataset. This competition was first turned up in 2003. Since then,
it has become a benchmark platform for the evaluation of image
annotation and retrieval algorithms in several domains such as
medical imagery, robot vision imagery or botanical collections.
This year, a new lab dedicated to life media LifeCLEF which
includes plant identification task has been released. In the past
edition, 2013, there were 33 submitted runs. Training data
resulted in 20985 images while testing data resulted in 5092.</p>
      <p>The rest of the paper is organized as follows: section 2
describes category dependent image analysis (section 2), divided
into two subsections that go in depth in the metadata analysis
(section 2.1) and in the image content analysis (section 2.2).
Section 3 is focused on the classification algorithms for the plant
identification purpose while in section 4 fusion and merging
methodologies are described. We conclude with a summarization
of the obtained results (section 4), pointing out the challenges
ahead for the use of content based retrieval technologies in
botanical domain.
2. CATEGORY DEPENDENT IMAGE
ANALYSIS</p>
      <p>As mentioned in the prior sections, the available dataset for
ImageCLEF2013 Plant Identification Task is segmented into 2
main categories, NaturalBackground and SheetAsBackground,
that are also divided into several sub-categories: Scan and
Scanlike for SheetAsBackground category, that are considered equally
in our system, and Leaf, Flower, Fruit, Stem and Entire for
NaturalBackground category. Both training and testing images
have an associated XML describing their metadata that permits
the system to separate the images into groups for the later
processing and classification.</p>
      <p>This subcategory based groups are the key units of the overall
plant identification process until the merging done taking into
account the Individual Plant Identification, a metadata parameter
that determines images that belong to the same plant. For each of
the subcategories or groups is necessary to extract all the relevant
knowledge. First, inferring this knowledge from the metadata
such as localization and date and second, describing the content
of images as in detail as possible and using discriminative factors.
Not all the implementations have been considered for all the
groups, taken decisions permit obtaining better results.</p>
      <p>In the next subsections, more detailed explanations are
presented regarding the metadata analysis and the deployed image
content description algorithms.
2.1 Image metadata analysis: georreference
and seasonal nature</p>
      <p>Considering the metadata information attached to each of the
images, we determine the inclusion of two metadata parameters:
GPS data and the date in order to extract knowledge that can
improve the plant identification process. These parameters are
included not only for the training dataset but also for the testing
dataset.</p>
      <p>The schema of categories and subcategories of the image
dataset delimits the use of these metadata parameters to the
Natural Background category. Images included within
SheetAsBasckground category don’t belong to natural
environments; consequently, their latitude, longitude and date
parameters don’t represent the plant ecosystem. Including these
data in the classification process can insert too much noise in the
system preventing good results.</p>
      <sec id="sec-1-1">
        <title>2.1.1 Georeferenced data</title>
        <p>Since ancient times, studies to determine the influence of
topography on species identification have been done. One of the
most important factors is the altitude at which each species grows.
Therefore, altitude has been considered one of the key indicators
for the classification process. Altitude values have been extracted
using the actual digital elevation model (DEM) for Europe as the
vast majority of the images belong to France. The inputs to these
models are the latitude and longitude data (GPS data).</p>
        <p>In this case, the classification process has been focused in the
analysis of the altitude parameter, not taking into account
longitude and latitude variables as we judge that it could increase
the noise level as all the images pertain to a specific country.
2.1.2 Seasonal nature classification</p>
        <p>The plants are species that change throughout the seasons.
Although not all plants undergo this change that doesn’t affect to
different parts of the plants in the same way, this seasonal concept
has been considered an important factor that can be determinant
in recognizing the plant. As a consequence, date metadata
parameter has been added to the classification attribute list.</p>
        <p>Even though, we didn’t consider it a very discriminative
parameter we also added the dominant colour of the segmented
object.</p>
        <p>Concerning Fruit subcategory, as the segmentation process was
not as accurate as in the previous case because the photos had
been taken in real scenarios, only dominant colour parameter was
extracted as it’s also a factor that can make the difference
between different types of plants.</p>
      </sec>
      <sec id="sec-1-2">
        <title>2.2.1 Segmentation</title>
        <p>
          Although usually image segmentation has a crucial
significance for content description, as mentioned before, our
system only uses it for the SheetAsBackground category and Fruit
subcategory. In the first case, an isolated leaf is represented in the
image with uneven illumination and possible shadows. We
implemented colour clustering techniques based on Local
Relative Entropy Method (LRE) [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] for the subtraction of the
background. As this background doesn’t represent a real scenario,
the results for the segmentation of this uniform area are promising
and therefore, valid for the implementation of an automatic
segmentation process.
        </p>
        <p>
          In the case of Fruit image segmentation, the assumption about
the importance of the flower object itself carries the necessity of
isolating it from the forest background. As well as in the previous
approximation, colour clustering techniques based on Joint
Relative Entropy method (JRE) [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] are used.
        </p>
        <p>Even more, we observe that Stem subcategory contains
predominantly images with tree trunks both in vertical and
horizontal that fill the majority of the image. Hence, in order to
minimize the effect of the insertion of noisy backgrounds to the
system, four fifths of the images are cropped in a fixed direction.
To determine this orientation of the trunk along the image, local
gradients are analyzed.
3. PLANT CLASSIFICATION</p>
        <p>
          All the image content retrieval solutions include a classification
stage where data mining algorithms are implemented. These
algorithms are necessary to infer knowledge from the extracted
features. Five different algorithms have been studied with the aim
of determining the best one for each of the subcategories:
Bayesian Network [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], Naive Bayes [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], SMO [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], SVM [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]
and Kstar [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. For the comparison between classification
algorithms, training dataset is split into two subsets, one for the
training and the other one for the validation of the
implementation. KNIME [16] is an appropriate framework to
carry out this learning approach and for the experimentation with
a range of algorithms and parameterization of them. It permits
working with several feature spaces at a time, therefore it a very
suitable framework to undertake the evaluation of the algorithms
with the best performance.
        </p>
        <p>As starting point, we considered the classification as totally
independent problem for each of the subcategories. The
interdependency between some of the images has not been taking
into account till the merging of the results. Most suitable features
(see section 5) are extracted from all the images belonging to the
same subcategory and they are gathered into five groups when all
present. Each of the group is also considered an independent
classification approach; therefore, the overall classification
process is atomized as a subcategory classifications solution
based on feature associations.</p>
        <p>For the learning of the classification algorithms the training
subset of images has been used and we validate the performance
of the five implemented classification algorithms using the
validation subset. As a result, we got at most five classification
modules per category for each feature group. These modules
output is a ClassID probability list that represents the probability
of each image to belong to a plant species.</p>
        <sec id="sec-1-2-1">
          <title>Training dataset</title>
          <p>Color feature</p>
          <p>Globalfeature</p>
          <p>Metadata
feature
PrincipalObject
features</p>
          <p>Texture
features
Testing dataset
s
m
h
t
i
r
o
g
l
a
n
o
i
t
a
c
iif
s
s
a
l
C</p>
          <p>ClassID probability ranking
list per feature-group
n
o
i
s
u
f
t
s
il
g
n
i
k
n
a
r
y
t
iil
b
a
b
o
r
P</p>
        </sec>
        <sec id="sec-1-2-2">
          <title>Retrieved ClassID probability ranking list</title>
          <p>4. FUSION AND MERGING OF
CLASSIFICATION RESULTS</p>
          <p>
            We grouped the extracted features in five different groups to
analyze their relevance in the identification task results. In
general, most of the Content Based Image Retrieval (CBIR)
systems employ a unique probability output to determine the
belonging class of a new query image. Multiple feature fusion is a
classical technique used in CBIR and pattern recognition to
improve the efficiency and robustness of results but this fusion is
usually done to feature level. As an alternative to this, we propose
an approach that computes the fusion of the classification results
at feature space level. Probability scores lists for each of the
feature group are fused using a Leave Out algorithm (LO) [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ].
Despite the algorithm was defined for its application using
similarity scores, the adoption to probability lists is direct.
          </p>
          <p>For the plant identification of a new query image, its features
are extracted taking into account the aforementioned five feature
spaces: colour, principal object, texture, global (DITEC) and
metadata (see Figure 1). Classification modules have been already
trained at feature space level so each feature group vector is
classified by the corresponding classifier. As the output of this
classification stage, we get a ClassID probability ranking list that
denotes the probability for that image to belong to each of the
plant classes regarding a concrete feature space.</p>
          <p>In order to get a unique output, these probability lists are fused
by setting the probability of an image belonging to a class to the
maximum of the probabilities in each list. The resulting
probability list represents the ranking for the plant identification
ClassID.</p>
          <p>= Prob. img
∈ ID
where; ID
= sort Prob img ∈ ID⃗</p>
          <p>ID⃑ = {
}
matrix is composed of cells representing a tupla that
contains the ClassId and the probability value of pertaining to that
class. Each of the columns represents the probability ranking list
for each of the feature spaces.</p>
          <p>⃗
=
⃗
= 1, …
⃗ vector represent the retrieved ClassID
ranking list (see Figure 1).
probability</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>5. RESULTS</title>
      <p>Same
Individual ClassID Prob.</p>
      <p>PlantID Rank List</p>
      <p>I1
I2
…
In
…</p>
      <p>Merging
prob.
scores</p>
      <p>Final ClassID
Prob. ranking list
But there is another fact that must be taken into consideration
when estimating classification results: ImageCLEF dataset
includes a metadata that must be considered during the plant
identification; it is the IndividualPlantID which represents an
exclusive number identifying images taken from the same plant.
Therefore, there is a need of merging results coming from the
same plant (see Figure 2). The ClassID probability lists belonging
to the same plant are merged by means of empirically obtained
weights for each of the subcategories.</p>
      <p>⃗ =
⃗
{
}
where {TSC } is the group of images selected for the validation
of the classification modules and the definition of the weights.</p>
      <p>First, retrieved ClassID probability lists ( ⃗ ) with the same
IndividualPlantID are gathered. Taking into consideration the
subcategory that the images belong to, probabilities are multiplied
by a factor that has been deduced from the performance of the
system for each of the subcategories ( ⃗ ). More precisely, the
weight represents the mean accuracy value of the two best
classification methods for each of the subcategories. In order to
infer this value training dataset has been split into two sets, one
for the training and the other for the validation of the
classification system. SheetAsBackground and Flower
subcategories are the ones with the highest weight while Stem and
Entire have rather lower values.</p>
      <p>Second, weighted probability lists are merged by means of the
highest probability score that will determine the ClassID of the
images with the same IndividualPlantID.</p>
      <p>⃗ =
(
∙
∈
{ }
;</p>
      <p>= ∀
) =
( ⃗
∩
{
})</p>
      <p>In order to validate the influence of each of the extracted
features in the overall process of plant identification we
considered to analyze the results of the classification process for
each of the subcategories. The results presented in this section are
the rate of correct predictions for each of the subcategories. These
prediction results have been computed using only the training
dataset, splitting this dataset into two sets, 90% of the images for
the training of the classification and fusion modules and the other
10% for the validation.</p>
      <p>As summarized in Table 2, not all the features have been
contemplated for all the subcategories, as an example
aforementioned associated metadata has not been included in the
classification of images pertaining to SheetAsBackground
category. In addition, all the extracted attributes concerning the
identification of the principal object of the image such as the
solidity, eccentricity or area-perimeter relationship has only been
rated for the SheetAsBackground category. By contrast, principal
object dominant colour attribute is extracted from both Flower
and SheetAsBackground categories.</p>
      <p>Concerning Leaf and Stem subcategories, metadata, textural
and DITEC attributes have been included as the most
representative features. As there is no a clear principal object in
the image and the colour is not something characteristic other
attributes were not considered.</p>
      <p>In the case of Entire subcategory, images contain the entire
natural scene where the plants grow, so the elements of the image
are very diverse. This fact introduces lots of noise in the system
and the classification of this subcategory is considered the most
ambitious. In this case, metadata features and DITEC have been
selected for the description.</p>
      <p>Fruit and Flower are the subcategories where image colouring
is a leading figure. Hence, for both subcategories metadata and
colour attributes are extracted. Even at first it was considered to
add the dominant colour attribute to both cases, due to the weak
results of the segmentation algorithms for Flower images we
dismiss that possibility and it was only included for Fruit. The
opposite of textural features, that are more descriptive in the case
of Flower subcategory.
0,600
0,500
n0,400
o
i
is0,300
c
rP0,200
e
0,100
0,000
0,600
0,500
n0,400
o
i
is0,300
c
rP0,200
e
0,100
0,000
0,09
0,08
0,07
n0,06
io0,05
s
ice0,04
r
P0,03
0,02
0,01
0</p>
      <p>In Figure 3 we resume the results obtained for the classification
process visualized separately for each subcategory. For Flower,
Fruit and Leaf categories metadata attributes are the ones with the
best precision rates. The results for the SaB, Flower, Fruit and
Leaf categories are quite promising while Stem and Entire
classification doesn’t give very good results. In the case of Entire
category, the inclusion of very diverse elements in the images can
distort the general perception of the plant itself and therefore
identification task becomes quite difficult. However, if we
consider the Stem category, we conclude that the extracted
features are not feasible for the identification of this type of
images.</p>
      <p>In general, fusion algorithms increase precision results so a
deeper analysis of the consequences of the utilization of these
approaches is recommended for plant identification solutions.
5.1 Comparison with ImageCLEF official
results</p>
      <p>In this subsection some comparative indicators about the
results obtained with the method presented in this paper and the
overall results of ImageCLEF participants are presented.
ImageCLEF results are divided into two different blocks: one of
them including only image from SheetAsBackground category
and the other one for the rest of the dataset images considered as
NaturalBackground category. All the values for the final
validation have been computed only for the testing dataset.</p>
      <p>SheetAsBackground
0,7
0,6
0,5
0,4
0,3
0,2
0,1
0




0,45
0,4
0,35
0,3
0,25
0,2
0,15
0,1
0,05</p>
      <p>0
As shown in Figure 4 the metric is a score related to the rank of
the correct species in the list of retrieved species, where,
U : number of users (who have at least one image in
the test data)
Pu : number of individual plants observed by the u-th
user
Nu,p : number of pictures taken from the p-th plant
observed by the u-th user
Su,p,n : score between 1 and 0 equals to the inverse of
the rank of the correct species (for the n-th picture
taken from the p-th plant observed by the u-th user)
In the following figures, highlighted in the graphics, the results of
the described method for both categories compared with the
results of all the participants of ImageCLEF 2013.</p>
      <p>NaturalBackground
As appreciated in the figures, the results obtained with the
described method are among the first half of the participants. In
the case of SheetAsBackground category, more emphasis must be
done in the segmentation process in order to have a better defined
content for the analysis.</p>
      <p>The bad results obtained for Stem and Entire subcategories
have a direct influence in the scores of the NaturalBackground
category so better approaches for the classification of these two
subcategories are going to be implemented in the near future.</p>
    </sec>
    <sec id="sec-3">
      <title>6. CONCLUSION</title>
      <p>This paper presents a system for the identification of several
plant species based on the analysis of metadata associated to an
image and the content of the image. The inclusion of metadata
parameters reveals an opportunity to refine the results of the
image content analysis. Even the described system has been
proved for ImageCLEF dataset, the approaches defined in this
paper are applicable to collections that contain plant images, only
the categorization of plant parts should be keep in mind.</p>
      <p>Concerning technical aspect of the system, remark the need of
the inclusion of new algorithms that overcome the actual results
especially for Entire and Stem categories. Additionally, merging
strategies should consider the insertion of unique image instance
identifiers previously in the classification process.</p>
      <p>The growing botanical collections ease the inclusion of image
retrieval solutions which are considered as very promising by
experimented scientists. Competitions such as ImageCLEF are
key factors on the approach between image analysis research
groups and botanists which permits faster scientific discovery.
Having an accurate knowledge about the identity of plant species
is essential for our biodiversity conservation.
7. ACKNOWLEDGMENTS</p>
      <p>Our thanks to ImageCLEF organizers and all members of
Pl@ntNet project and Tela Botanica initiative who brought us the
possibility of researching on the application of multimedia
analysis techniques applied to environmental data.</p>
    </sec>
    <sec id="sec-4">
      <title>8. REFERENCES</title>
      <p>[16] http://www.knime.org/</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Caputo</surname>
          </string-name>
          ,
          <article-title>Barbara and Müller, Henning and Thomee, Bart and Villegas, Mauricio and Paredes, Roberto and Zellhofer, David and Gobeau, Herve and Joly, Alexis and Bonnet, Pierre and Martínez-Gómez, Jesus and García-Varea, Ismael and Cazorla, Miguel</article-title>
          .
          <source>ImageCLEF</source>
          <year>2013</year>
          :
          <article-title>The Vision, the Data and the Open Challenges</article-title>
          . Springer, Lecture Notes in Computer Science, p.
          <fpage>250</fpage>
          -
          <lpage>268</lpage>
          ,
          <year>2013</year>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Gobeau</surname>
          </string-name>
          , Herve and Joly, Alexis and Bonnet, Pierre and Bakic, Vera and Bartheemy, Daniel and Boujemaa, Nozha and Molino, Jean-Francois.
          <article-title>The imageCLEF 2013 Plant Identification Task</article-title>
          ,
          <article-title>CLEF 2013 Evaluation Labs</article-title>
          and Workshop, Online Working Note,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Olaizola</surname>
            ,
            <given-names>I.G</given-names>
          </string-name>
          ; Quartulli,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Florez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ;
            <surname>Sierra</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          <article-title>Trace Transform Based Method for Color Image Domain Identification</article-title>
          .
          <year>2014</year>
          . Multimedia, IEEE Transactions on
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Haralick</surname>
            ,
            <given-names>R.M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Shanmugam</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Dinstein</surname>
          </string-name>
          ,
          <string-name>
            <surname>Its'Hak</surname>
          </string-name>
          ,
          <year>1973</year>
          .
          <article-title>Textural Features for Image Classification</article-title>
          .
          <source>Systems, Man and Cybernetics</source>
          , IEEE Transactions on, SMC-
          <volume>3</volume>
          <fpage>610</fpage>
          -
          <lpage>621</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ming-Kuei</surname>
          </string-name>
          .
          <article-title>Visual pattern recognition by moment invariants</article-title>
          .
          <source>Information Theory, IRE Transactions on, 8.2</source>
          (
          <year>1962</year>
          ):
          <fpage>179</fpage>
          -
          <lpage>187</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Teague</surname>
            ,
            <given-names>Michael</given-names>
          </string-name>
          <string-name>
            <surname>Reed</surname>
          </string-name>
          .
          <article-title>Image analysis via the general theory of moments*</article-title>
          .
          <source>JOSA 70.8</source>
          (
          <year>1980</year>
          ):
          <fpage>920</fpage>
          -
          <lpage>930</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>DC.</given-names>
            <surname>He</surname>
          </string-name>
          and L.
          <string-name>
            <surname>Wang</surname>
          </string-name>
          (
          <year>1990</year>
          ), Texture Unit,
          <string-name>
            <given-names>Texture</given-names>
            <surname>Spectrum</surname>
          </string-name>
          ,
          <article-title>And Texture Analysis, Geoscience and Remote Sensing, IEEE Transactions on</article-title>
          , vol.
          <volume>28</volume>
          , pp.
          <fpage>509</fpage>
          -
          <lpage>512</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Jović</surname>
          </string-name>
          , Mladen and Hatakeyama, Yutaka and Dong, Fangyan and Hirota, Kaoru.
          <source>Image Retrieval Based on Similarity Score Fusion from Feature Similarity Ranking Lists</source>
          , Springer Berlin Heidelberg,
          <volume>4223</volume>
          ,
          <year>2006</year>
          , pp.
          <fpage>461</fpage>
          -
          <lpage>470</lpage>
          , DOI: http://dx.doi.org/10.1007/11881599_54}
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Chein-I Chang</surname>
          </string-name>
          , Kebo Chen, Jianwei Wang and
          <string-name>
            <surname>Mark L. G. Althouse</surname>
          </string-name>
          ,
          <article-title>A relative entropy-based approach to image thresholding, 1994, Pergamon Pattern Recognition</article-title>
          . Vol.
          <volume>27</volume>
          , No. 9. pp.
          <fpage>1275</fpage>
          <lpage>1289</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Eliza Yingzi</surname>
          </string-name>
          Du and
          <string-name>
            <surname>Chein-I Chang</surname>
          </string-name>
          and Thouin, P.D.
          <article-title>Relative entropy-based methods for image thresholding</article-title>
          ,
          <source>Circuits and Systems</source>
          ,
          <year>2002</year>
          .
          <article-title>ISCAS 2002</article-title>
          . IEEE International Symposium on, II-265
          <string-name>
            <surname>-</surname>
          </string-name>
          II-268 vol.
          <volume>2</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Pearl</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>1985</year>
          ),
          <article-title>Bayesian Networks: A Model of SelfActivated Memory for Evidential Reasoning</article-title>
          ,
          <source>Proceedings of the 7th Conference of the Cognitive Science Society</source>
          , University of California, Irvine, CA. pp.
          <fpage>329</fpage>
          -
          <lpage>334</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>George</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>John</surname>
          </string-name>
          , Pat Langley.,
          <source>Estimating Continuous Distributions in Bayesian Classifiers, Eleventh Conference on Uncertainty in Artificial Intelligence</source>
          , San Mateo,
          <fpage>338</fpage>
          -
          <lpage>345</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Platt</surname>
          </string-name>
          ,
          <article-title>Fast Training of Support Vector Machines using Sequential Minimal Optimization</article-title>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schoelkopf</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Burges</surname>
          </string-name>
          and
          <string-name>
            <surname>A</surname>
          </string-name>
          . Smola, editors,
          <source>Advances in Kernel Methods - Support Vector Learning</source>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Cristianini</surname>
          </string-name>
          , Nello; Shawe-Taylor,
          <article-title>John; An Introduction to Support Vector Machines and other kernel-based learning methods</article-title>
          , Cambridge University Press,
          <year>2000</year>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>John</surname>
            <given-names>G. Cleary</given-names>
          </string-name>
          , Leonard E. Trigg: K*:
          <article-title>An Instance-based Learner Using an Entropic Distance Measure</article-title>
          .
          <source>12th International Conference on Machine Learning</source>
          ,
          <fpage>108</fpage>
          -
          <lpage>114</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>