<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Plant Identi cation on Amazonian and Guiana Shield Flora: NEUON submission to LifeCLEF 2019 Plant</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sophia Chulif</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kiat Jing Heng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Teck Wei Chan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abdullah Al Monnaf</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yang Loong Chang</string-name>
          <email>yangloongg@neuon.ai</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Arti cial Intelligence, NEUON AI</institution>
          ,
          <addr-line>94300 Sarawak</addr-line>
          ,
          <country country="MY">Malaysia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper will look into the use of ne-tuned Inception-v4 and Inception-ResNet-v2 models to automate the classi cation of 10,000 plant species. Prior to training the networks, the training dataset was pre-processed to remove the noisy data. The team submitted three runs which achieved comparable performances to human experts on the test dataset comprising 745 observations for all the evaluation metrics. For the trained systems to generalise better, the systems were trained for multi-task classi cation and is able to classify plant images based on their species, with support of their genus and family labels. In particular, an ensemble of Inception-v4 and Inception-ResNet-v2 networks achieved a Top-1 accuracy of 0.316 and 0.246 for the test set identi ed by experts and the whole test set respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>Plant identi cation</kwd>
        <kwd>computer vision</kwd>
        <kwd>convolutional neural networks</kwd>
        <kwd>data cleaning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The plant identi cation challenge of the Conference and Labs of the Evaluation
Forum (CLEF) is an annual challenge focusing on the automation of plant
identi cation. In recent years, automated identi cation of plants has become a high
interest for botanical specialists and computer experts alike [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Transitioning
from the conventional computer vision through feature descriptor methods [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ],
deep learning based methods have been able to signi cantly improve accuracy of
the automation of plant identi cation [
        <xref ref-type="bibr" rid="ref11 ref7 ref8">7,8,11</xref>
        ]. This application is essential in the
utilisation, management and conservation of ora of any kind. When the plant
identi cation challenge rst commenced in 2011, the main focus was to be able
to identify over 70 di erent tree species given 3996 leaf images as training data
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The total number of species and training images in the challenge has then
notably increased each year. With the training dataset initially covering only
leaf images, it has now expanded to di erent plant organs and multiple views of
the plant. Since 2017, the total number of plant species has reached 10,000 along
with the presence of noisy images (images that are unrelated to the plant of
interest) in the training dataset - this posed a great challenge to the participants
in evaluating the extend of a noisy dataset competing with a trusted dataset
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Furthermore, the plants in the training dataset shared a low intra-class and
high inter-class similarity. They may belong to the same class but look clearly
di erent from one another [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Besides the variance, some of the distinct plant
organs and views may not be captured in the given training images, resulting
in the lack of training data thus di culty in predicting the species. This depicts
the realistic challenge of the real-world applications in plant identi cation.
      </p>
      <p>
        For PlantCLEF 2019, the dataset provided was focused on the Guiana shield
and the Amazon rainforest which is known to be the largest collection of living
plants and species in the world [
        <xref ref-type="bibr" rid="ref4 ref6">6,4</xref>
        ]. The task was to predict 10,000 plant species
and the participants were provided with 448,071 images of training data. Along
with a test set containing 2,974 images and 745 observations, the main evaluation
method of the competition was the Top 1 accuracy of the submitted predictions
based on the 745 observations.
      </p>
      <p>The primary challenge of PlantCLEF 2019 was the decreased average number
of images per species in the training dataset. Many species contained only a few
images and some even contained only one image. Moreover, the training dataset
consisted of noisy data that constituted from non-plant images and duplicate
images that may come with incorrect labels.</p>
      <p>This paper presents the preparation made prior to the training of the system.
The system was ne-tuned from pretrained Inception-v4 and
Inception-ResNetv2 models to automate the classi cation of the given 10,000 plant species.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Data Preparation</title>
      <sec id="sec-2-1">
        <title>Data Analysis</title>
        <p>The dataset provided consisted of 10,000 species which mainly focused on the
Guiana shield and the Amazon rainforest. The reported training dataset size is
448,071 (based on the number of rows in PlantCLEF2019MasterTraining.csv)
however, the downloaded training dataset size was 433,810. Therefore, the
actual downloaded dataset included the training dataset of 433,810 images which
consisted of 10,000 classes (species), and a test dataset of size 2,974 from 745
di erent observations.</p>
        <p>As mentioned in the challenges data collection description, among these
10,000 species many contain only a few images and some of them contain only
one image. Moreover, it was observed that there were many variations in the
training dataset that could a ect the performance of the plant identi cation.</p>
        <p>Other than real-world plant images i.e. fruit, branch, trunk, stem, root,
ower, leaf etc., the training dataset initially contained many images that do not
look like real actual plants i.e. sketches, paintings, illustrations, plant
herbariums, and small regions of the interested plants. In addition, the dataset also
contained non-plant images i.e. animals, logos, book covers, graphs, table, medicine
bottles, humans, chemical-bond diagrams, presentation slides, maps etc. Besides
this, the dataset consisted of images with duplicate names in di erent classes
(folders). Furthermore, the dataset consisted of duplicate images with di erent
names within the same folder as well as in di erent folders. Having di erent
labels for the same images could cause confusion to the machine.</p>
        <p>These characteristics observed could a ect the performance of the plant
identi cation therefore the approach was to pre-processed the dataset.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Data Cleaning</title>
        <p>From the analysis in subsection 2.1, the dataset was then pre-processed (cleaned)
to allow better execution of plant identi cation. First, the images with duplicate
names were removed as those images are actually the same. The dataset given
consisted of 433,810 images and 10,000 classes of plants. After eliminating the
duplicate names, the dataset was reduced to 279,183 images and 8,468 classes.</p>
        <p>Then, the duplicate images (without having the same name) were further
removed. After removing them, 263,987 images were left with 8,419 classes. Finally,
the non-plant images were eliminated, resulting in a total of 250,646 images and
8,263 of classes for training1.</p>
        <p>In the approach of eliminating images with duplicate names in di erent
folders, since there is no method to decide which class they actually belong to (no
experts to verify), all images with the same name were removed.</p>
        <p>On the other hand, in order to remove duplicate images within the same folder
as well as di erent folders, inception-v4's feature extractor layer (Mixed 7d) was
used to compare images so that those with 0.99 cosine similarities in features
can be eliminated. If the duplicate images only exist in the same class (folder),
only one of the images will be retained. Then, the di erence hash algorithm
was used to detect more duplicates within the dataset. The hash value for every
image was calculated and then compared to get those with very little di erence.
Images with little di erence hash values mean they are identical. Likewise, if the
identical images only exist in the same class (folder), only one image is retained,
the rest are eliminated.</p>
        <p>To detect non-plant images, a discrimination network for identifying plant
and non-plant images was trained. This process consisted of 3 phases.</p>
        <p>In phase 1, the positive samples (plant) were taken from PlantClef 2016 while
the negative samples (non-plant) were taken from ImageNet2012 (excluding the
plant classes). The training dataset size was 4,000, with each class having 2,000
samples. Meanwhile the validation dataset size was 2,000, with each class having
1,000 samples. An Inception-V4 plant and non-plant classi er was then trained
using these datasets.
1 The cleaned list can be found at Github via https://github.com/changyangloong/
plantclef2019_challenge</p>
        <p>In phase 2, 5,000 samples were randomly selected from PlantClef 2019 and
predicted using the trained classi er. The performance on this sample was
evaluated manually. After evaluating its performance, the training set was re ned
by manually correcting the prediction and adding them back to the training
set. Then, the Inception-V4 plant and non-plant classi er was retrained using
the new training set. Phase 2 was repeated until the performance satis ed the
accuracy of over 90%. In this case it was repeated 4 times.</p>
        <p>In phase 3, the Inception-V4 plant and non-plant classi er was applied to the
whole dataset to remove the non-plant images. Only the softmax probability of
0.98 for the non-plant class was regarded as non-plant images.</p>
        <p>The entire process is visualised in Fig. 1.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>This section will describe the motivations for using Inception-v4 and
InceptionResNet-v2 for the plant classi cation challenge, the training setup, network
hyperparameters and validation results obtained from the trained networks.
3.1</p>
      <sec id="sec-3-1">
        <title>Network Architecture</title>
        <p>
          The backbone networks used for classifying PlantCLEF 2019 images were
Inception-v4 and Inception-ResNet-v2 models. These two models were adopted in this
plant classi cation task as they come with the lowest Top-1 Error and Top-5
Error for single-crop - single-model experimental results, when being proposed
in the original paper [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. Besides, the use of inception modules allows lters with
multiple sizes to perform convolution on the same level to cater to the variation
in the location of the information on the image [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], which is applicable on the
training dataset.
        </p>
        <p>
          Both of the networks implement Convolutional Neural Network (CNN)
architecture, which typically consists of convolutional layers, pooling layers, dropout
layers and fully-connected layers. The models were ne-tuned from pre-trained
weights on the ImageNet dataset [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
        </p>
        <p>In fact, Inception-v4 and Inception-Resnet-v2 networks share an identical
stem as the input part of these networks. However, these two models demonstrate
di erences in architecture after the stem.</p>
        <p>For Inception-v4 model, there are three main inception modules following the
stem, namely Inception-A, Inception-B and Inception-C modules. The output
will then be concatenated and fed to the next module. Meanwhile, there is a
reduction module after Inception-A and Inception-B modules to alter the width
and height of the grid. These modules work together to serve as the model's
feature extractor.</p>
        <p>On the other hand, Inception-ResNet-v2 model makes use of residual
connections in addition to inception modules after the implementation of stem.
This allows Inception-ResNet-v2 to achieve higher accuracies within a shorter
time frame, while being similarly computationally expensive as the Inception-v4
model. For residual connections to work, the pooling layers from pure Inception
modules are replaced with residual connections. Similarly, there is a respective
reduction module following Inception-ResNet-A and Inception-ResNet-B
modules.</p>
        <p>Both Inception-v4 and Inception-ResNet-v2 model have the same structures
for the classi cation part, which consists of an average pooling layer, a dropout
layer, and a fully-connected layer which return the softmax probabilities over
predicted output classes.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Training Setup</title>
        <p>During the network training, two types of classi cation methods were
investigated, namely single label classi cation and multi-task classi cation.
Single Label Classi cation The rst classi cation method is the conventional
classi cation method which is based on single label prediction model. The labels
are the plant species, therefore there are 10,000 classes.</p>
        <p>
          Multi-Task Classi cation Another method used in this plant identi cation
task is the multi-task classi cation whereby the labels \Species",\Genus" and
\Family" of the samples were used in training. This allowed the network to
regularise and generalise better on images from a large number of classes. The
total number of family class is 248, while the genus class is 1,780 2.
2 The multi-task label can be found at https://github.com/changyangloong/
plantclef2019_challenge
Library Used The networks were implemented using TensorFlow Slim library,
with the weights being pre-trained on ImageNet datset [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. The networks were
then ne-tuned accordingly using the hyperparameters described in Sub-section
3.4. Since the adopted models pre-trained on ImageNet have only 1,000 classes,
the fully-connected layer of the adopted models was adapted to 10,000 classes
during transfer learning for PlantCLEF 2019. For multi-task classi cation, there
were two additional fully-connected layers which were catered to 248 \Family"
classes and 1,780 \Genus" classes.
        </p>
        <p>Training Data The input image size of the training was 299 299 3. By
separating 20,000 samples from the training samples as a validation set, the
total remaining training set comprised 230,646 images. Although the total
number of classes is reduced from 10,000 classes to 8,263 classes, the network was
still trained to classify 10,000 classes through class mapping. This allows model
update when missing classes are to be found.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Data Augmentation</title>
        <p>
          Data augmentation was performed on the training images to increase the
training sample size. With this, the CNN network can learn features that are invariant
to their locations in the images and various transforms, which then e ectively
reduces the chance of over tting [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. Random cropping, horizontal ipping and
colour distortion (brightness, saturation, hue, and contrast) of images were
applied to the training dataset to increase the possibility of classifying the correct
plants regardless of their di erent environments or orientations.
3.4
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>Network Hyperparameters</title>
        <p>The learning dropout rate was set to 0.2 (keeping 80 % of the neurons before
the fully-connected layer) while the optimizer used was Adam Optimizer. The
optimizer took on an initial learning rate of 0.0001. Meanwhile, the gradient was
clipped at 1.25 to prevent the occurrence of exploding gradients. Softmax cross
entropy loss was used to compute the error between the predicted labels and true
labels; while the L2 regularization loss was added with weight decay of 0.00004.
The network hyperparameters can be summarised in Table 1.
3.5</p>
      </sec>
      <sec id="sec-3-5">
        <title>Validation</title>
        <p>Prior to training, 20,000 samples were randomly separated from the cleaned
training dataset of 250,646 images as validation set. It was ensured that each of
the remaining 8,263 classes in the training set has at least one sample left for
training. A validation sample was not added if there is only one sample left for
training in each class. In other words, a class will not have a validation sample
if it has only one sample.</p>
        <p>There were 4 approaches in testing the validation set on 4 di erent kinds
of network. The approaches include \Top 1 Centre Crop", \Top 1 Centre Crop
+ Corner Crop", \Top 5 Centre Crop", and \Top 5 Centre Crop + Corner
Crop". Meanwhile the 4 di erent networks included \Inception-v4",
\Inceptionv4 Multi-task", \Inception-ResNet-v2 Multi-task", and \Inception-v4 Multi-task
+ Inception-ResNet-v2 Multi-task" with di erent dataset sizes. Note that all the
trained networks were validated upon the same 20,000 validation images.</p>
        <p>The \Centre Crop" approach considers the centre region of the sample. The
centre region was cropped and resized then passed into the network for testing.
The \Corner Crop" approach on the other hand focused on the centre, top
left, top right, lower left, and lower right region of the image. Each region was
cropped and resized then passed into the network for testing. The \Top 1" and
\Top 5" approaches represent the Top 1 and Top 5 predictions based on the
testing results.
3.6</p>
      </sec>
      <sec id="sec-3-6">
        <title>Inference Procedures</title>
        <p>The following procedures were adopted for inferencing the predictions of the
test dataset. There were three models used for inference, namely \Multi-Task
Inception-v4", \Multi-Task Inception-ResNet-v2" and \Multi-Task Inception-v4
+ Multi-Task Inception-ResNet-v2".
1. The test images with the same observation ID were grouped together.
2. A total of ve center crop and corner crop images were then produced for a
single test image.
3. The test images grouped under the same observation ID were fed into the
CNN models for label predictions, as shown in Fig. 2. This would mean a
total of ve predictions for a single test image.
4. The probabilities of each prediction were averaged over the total number of
test image crops grouped under the same observation ID.
5. The top 100 probabilities with value greater than 0.001 were collected for
result tabulations.
6. Step 2 to 5 were repeated for every di erent observation ID.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results and Discussions</title>
      <sec id="sec-4-1">
        <title>Validation Results</title>
        <p>Inception-v4 and Inception-ResNet-v2 were the 2 networks used for training and
classi cation in these 4 approaches. The Multi-task description represents the
classi cation on multiple labels, i.e. Species, Family and Genus.</p>
        <p>The validation results are shown in Table 2. The results have shown that
while eliminating the duplicate image could increase the performance of the
single-task network, multi-task classi cation method was able to achieve a even
higher accuracy on the validation dataset.</p>
        <p>Interestingly enough, removing non-plant images after ltering duplicate
images out leads to a slight drop in the single-task Inception-v4 network's
performance. Since no Inception-v4 network is trained without the non-plant images
(i.e. keeping potentially duplicate images but of plants only) as benchmark, there
is no way to deduce whether the presence of duplicate images or non-plant images
is the main factor to the performance drop in our current experiments.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Submitted Runs</title>
        <p>The team submitted a total of three runs based on three di erent trained models
which achieved the top-3 validation accuracy. The models are described in the
followings:
Holmes Run 1 utilises ne-tuned Inception-v4 model catered to multi-task
classi cation (which are Species, Genus and Family).
Holmes Run 2 is an ensemble of Inception-v4 and Inception-ResNet-v2, and
is also catered to multi-task classi cation.</p>
        <p>Holmes Run 3 summarises prediction results from ne-tuned
Inception-ResNetv2 catered to multi-task classi cation.</p>
        <p>The run les were formatted as ObservationID; ClassID; Probability; Rank.
4.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Sample of Predictions</title>
        <p>The team submitted three runs with Holmes Run2 being the best performance
among our submission in terms of Top-1 accuracy,Top-3 accuracy and Top-5
accuracy. Holmes Run 2 achieved Top 1 accuracy of 31.6% for the test set
identi ed by experts, despite the occurrence of missing classes after data cleaning.
This proves that using an ensemble of two di erent state-of-the-art CNN models
can increase the robustness of the system and return better predictions. Table 3
summaries the results achieved by the team.</p>
        <p>Mean Reciprocal Rank (MRR) was also used to evaluate the performance of
the submitted runs. The MRR is the average of the reciprocal ranks of the whole
test set, with the reciprocal rank of a query response being the multiplicative
inverse of the rank of the rst correct answer. It can be mathematically
represented as shown, with j Q j being the frequency of plant occurrences in the test
set.</p>
        <p>Holmes Run 2, being our best performing machine, has achieved a MRR
of 0.362 and 0.298 for the test set identi ed by experts and the whole test set
respectively, as shown in Fig. 6.</p>
        <p>At the same time, all the three runs submitted by the team outperform 1 out
of 5 human experts of the Amazonian French Guina ora according to Top-1 (Fig.
7) and Top-3 accuracies. As for Top-5 accuracy, Holmes Run 2 outperformed 2
human experts, as captured in Fig. 8.</p>
        <p>
          Fig. 4. Top-1 accuracy achieved by all the submitted runs.
The automated identi cation of plants for PlantCLEF 2019 has shown a notable
decrease in performance compared to the previous editions. This may be due
to the decrease of per class samples as mentioned in the challenge's description.
In such real-world condition, the dataset requires pre-processing before training
is done. Although it has been seen that training with noisy data can be more
pro table [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], the noisy data in this challenge worsens the performance of the
automatic plant identi cation. At this stage of understanding, the team believe
the portion of noisy images should not be occupying a large margin (duplicates
le names are 55.13% out of the whole training set) of training set to act as good
regularisation agent.
        </p>
        <p>Therefore, the dataset was cleaned before training. This prevents the network
from getting trained by large amount of noisy data such as the non-plant images
and duplicate images/names (which may come with incorrect labels). Moreover,
by adding multi-task classi cation in our system, it has helped in regularising
the network and improving the performance. Modelling a relationship between
Species, Genus and Family is essential for better plant recognition.</p>
        <p>Since the cleaned dataset is reduced to 8,263 classes, there was no
capability to distinguish the missing classes. Hence, the label for the missing classes
was unable to be predicted if it is present in the test set resulting in a lower
performance. Additionally, 20,000 images had been sidelined for validation
purposes which are believed to be helpful in increasing the performance if they were
added.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>The task of the PlantCLEF 2019 challenge was to return the most likely matching
species for each observation (a set of images of the same plant) of the test set.
In this paper, the team has presented the overview and results of our approach
in doing so. With regards to the diversity of the dataset, the team has found
that the cleaning of the dataset and multi-task classi cation of Species, Genus
and Family improved the prediction results.</p>
      <p>According to the competition results released, our trained model was better
than one of the experts in Top 1 and Top 3 accuracy. Additionally, it has a close
performance with Expert 4. Our model even outperformed 2 experts for Top 5
accuracy. Overall, 3 of our submitted runs obtained the good results in every
machine category.</p>
      <p>
        The identi cation of plants based o images is indeed a di cult task even
for some of the botanical experts. Relying on plant images alone is usually
insu cient to determine the correct species as they may contain only partial
information of the plant [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Based on the competition results of PlantCLEF 2018
and 2019, it can be considered that with su cient data for training, machines
are able to perform nearly equal or better than a human.
      </p>
      <p>For future work, the plant images from the missing classes can be retrieved
if the ground-truth labels are known and will be used for training the networks
for better predictions.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgment</title>
      <p>The resources of this project is supported by NEUON AI SDN. BHD., Malaysia.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Darwin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The di erent forms of owers on plants of the same species</article-title>
          .
          <source>John Murray</source>
          (
          <year>1877</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Goeau, H.,
          <string-name>
            <surname>Bonnet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Plant identi cation based on noisy web data: the amazing performance of deep learning (lifeclef 2017)</article-title>
          . In: CLEF 2017-
          <article-title>Conference and Labs of the Evaluation Forum</article-title>
          . pp.
          <volume>1</volume>
          {
          <issue>13</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. Goeau, H.,
          <string-name>
            <surname>Bonnet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of expertlifeclef 2018: how far automated identi cation systems are from the best experts?</article-title>
          <source>In: CLEF</source>
          <year>2018</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4. Goeau, H.,
          <string-name>
            <surname>Bonnet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of lifeclef plant identi cation task 2019: diving into data de cient tropical countries</article-title>
          .
          <source>In: CLEF working notes 2019</source>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5. Goeau, H.,
          <string-name>
            <surname>Bonnet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boujemaa</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barthelemy</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molino</surname>
            ,
            <given-names>J.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Birnbaum</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mouysset</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Picard</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The imageclef 2011 plant images classi cation task</article-title>
          .
          <source>In: ImageCLEF 2011</source>
          . pp.
          <volume>0</volume>
          {
          <issue>0</issue>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Goeau, H.,
          <string-name>
            <surname>Botella</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kah</surname>
          </string-name>
          , l.S.,
          <string-name>
            <surname>Servajean</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glotin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bonnet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vellinga</surname>
            ,
            <given-names>W.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Planque</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fabian-Robert</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Muller, H.:
          <article-title>Overview of lifeclef 2019: Identi cation of amazonian plants, south &amp; north american birds, and niche prediction</article-title>
          .
          <source>In: Proceedings of CLEF</source>
          <year>2019</year>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>S.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chan</surname>
            ,
            <given-names>C.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilkin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Remagnino</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Deep-plant: Plant identi cation with convolutional neural networks</article-title>
          .
          <source>In: 2015 IEEE International Conference on Image Processing (ICIP)</source>
          . pp.
          <volume>452</volume>
          {
          <fpage>456</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>S.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>Y.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chan</surname>
            ,
            <given-names>C.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Remagnino</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Hgo-cnn: Hybrid generic-organ convolutional neural network for multi-organ plant classi cation</article-title>
          .
          <source>In: 2017 IEEE International Conference on Image Processing (ICIP)</source>
          . pp.
          <volume>4462</volume>
          {
          <fpage>4466</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Mikolajczyk</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grochowski</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Data augmentation for improving deep learning in image classi cation problem</article-title>
          .
          <source>In: 2018 international interdisciplinary PhD workshop (IIPhDW)</source>
          . pp.
          <volume>117</volume>
          {
          <fpage>122</fpage>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Russakovsky</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krause</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Satheesh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Ma,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            ,
            <surname>Karpathy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Khosla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Bernstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Berg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.C.</given-names>
            ,
            <surname>Fei-Fei</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          :
          <article-title>Imagenet large scale visual recognition challenge</article-title>
          .
          <source>International journal of computer vision 115(3)</source>
          ,
          <volume>211</volume>
          {
          <fpage>252</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Sulc</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Picek</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Matas</surname>
          </string-name>
          , J.:
          <article-title>Plant recognition by inception networks with testtime class prior estimation</article-title>
          .
          <source>Working Notes of CLEF</source>
          <year>2018</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , Io e, S.,
          <string-name>
            <surname>Vanhoucke</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alemi</surname>
            ,
            <given-names>A.A.</given-names>
          </string-name>
          :
          <article-title>Inception-v4, inception-resnet and the impact of residual connections on learning</article-title>
          .
          <source>In: Thirty-First AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sermanet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reed</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anguelov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Erhan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanhoucke</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rabinovich</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Going deeper with convolutions</article-title>
          .
          <source>In: Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          . pp.
          <volume>1</volume>
          {
          <issue>9</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Walsh</surname>
            , J.,
            <given-names>O</given-names>
          </string-name>
          ' Mahony, N.,
          <string-name>
            <surname>Campbell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carvalho</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krpalkova</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>VelascoHernandez</surname>
          </string-name>
          , G.,
          <string-name>
            <surname>Harapanahalli</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riordan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Deep learning vs</article-title>
          .
          <source>traditional computer vision</source>
          (04
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>